| /linux/net/core/ |
| H A D | netdev_config.c | 10 struct netdev_queue_config *qcfg, in netdev_nop_validate_qcfg() argument 17 struct netdev_queue_config *qcfg, in __netdev_queue_config() argument 22 struct netdev_queue_config *qcfg, in __netdev_queue_config() 31 memset(qcfg, 0, sizeof(*qcfg)); in __netdev_queue_config() 35 dev->queue_mgmt_ops->ndo_default_qcfg(dev, qcfg); in __netdev_queue_config() 36 err = validate_cb(dev, qcfg, extack); in __netdev_queue_config() 43 qcfg->rx_page_size = mpp->rx_page_size; in __netdev_queue_config() 44 err = validate_cb(dev, qcfg, extack); in __netdev_queue_config() 67 struct netdev_queue_config *qcfg) in netdev_queue_config() argument 69 __netdev_queue_config(dev, rxq_idx, qcfg, NULL, false); in netdev_queue_config() [all …]
|
| H A D | netdev_rx_queue.c | 104 struct netdev_queue_config qcfg; in netdev_rx_queue_restart() local 106 netdev_queue_config(dev, rxq_idx, &qcfg); in netdev_rx_queue_restart() 107 return netdev_rx_queue_reconfig(dev, rxq_idx, &qcfg, &qcfg); in netdev_rx_queue_restart() 116 struct netdev_queue_config qcfg[2]; in __net_mp_open_rxq() local 158 netdev_queue_config(dev, rxq_idx, &qcfg[0]); in __net_mp_open_rxq() 160 ret = netdev_queue_config_validate(dev, rxq_idx, &qcfg[1], extack); in __net_mp_open_rxq() 164 ret = netdev_rx_queue_reconfig(dev, rxq_idx, &qcfg[0], &qcfg[1]); in __net_mp_open_rxq() 189 struct netdev_queue_config qcfg[2]; in __net_mp_close_rxq() local 209 netdev_queue_config(dev, ifq_idx, &qcfg[0]); in __net_mp_close_rxq() 211 netdev_queue_config(dev, ifq_idx, &qcfg[1]); in __net_mp_close_rxq() [all …]
|
| H A D | dev.h | 96 struct netdev_queue_config *qcfg,
|
| /linux/include/net/ |
| H A D | netdev_queues.h | 162 struct netdev_queue_config *qcfg, 168 struct netdev_queue_config *qcfg, 175 struct netdev_queue_config *qcfg); 177 struct netdev_queue_config *qcfg, 186 struct netdev_queue_config *qcfg);
|
| H A D | netdev_rx_queue.h | 31 struct netdev_queue_config qcfg; member
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_tx.c | 341 total_queues = cfg->qcfg->num_queues + cfg->num_xdp_rings; in gve_tx_alloc_rings_gqi() 342 if (total_queues > cfg->qcfg->max_queues) { in gve_tx_alloc_rings_gqi() 348 tx = kvzalloc_objs(struct gve_tx_ring, cfg->qcfg->max_queues); in gve_tx_alloc_rings_gqi() 381 for (i = 0; i < cfg->qcfg->num_queues + cfg->qcfg->num_xdp_queues; i++) in gve_tx_free_rings_gqi()
|
| H A D | gve_tx_dqo.c | 362 if (idx >= cfg->qcfg->num_queues && cfg->num_xdp_rings) { in gve_tx_alloc_ring_dqo() 421 total_queues = cfg->qcfg->num_queues + cfg->num_xdp_rings; in gve_tx_alloc_rings_dqo() 422 if (total_queues > cfg->qcfg->max_queues) { in gve_tx_alloc_rings_dqo() 428 tx = kvzalloc_objs(struct gve_tx_ring, cfg->qcfg->max_queues); in gve_tx_alloc_rings_dqo() 461 for (i = 0; i < cfg->qcfg->num_queues + cfg->qcfg->num_xdp_queues; i++) in gve_tx_free_rings_dqo()
|
| H A D | gve_main.c | 966 cfg->qcfg = &priv->tx_cfg; in gve_tx_get_curr_alloc_cfg() 969 cfg->num_xdp_rings = cfg->qcfg->num_xdp_queues; in gve_tx_get_curr_alloc_cfg() 1370 priv->tx_cfg = *tx_alloc_cfg->qcfg; in gve_queues_start() 1862 tx_alloc_cfg.qcfg = &new_tx_config; in gve_adjust_queues() 2621 struct netdev_queue_config *qcfg, in gve_rx_queue_mem_alloc() argument 2644 struct netdev_queue_config *qcfg, in gve_rx_queue_start() argument
|
| H A D | gve.h | 713 struct gve_tx_queue_config *qcfg; member
|
| /linux/drivers/net/netdevsim/ |
| H A D | netdev.c | 765 struct netdev_queue_config *qcfg, in nsim_queue_mem_alloc() argument 815 nsim_queue_start(struct net_device *dev, struct netdev_queue_config *qcfg, in nsim_queue_start() argument
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_tx_rx.c | 1264 #define BNA_GET_RXQS(qcfg) (((qcfg)->rxp_type == BNA_RXP_SINGLE) ? \ argument 1265 (qcfg)->num_paths : ((qcfg)->num_paths * 2))
|
| /linux/drivers/net/ethernet/broadcom/bnxt/ |
| H A D | bnxt.c | 4329 struct netdev_queue_config qcfg; in bnxt_init_ring_struct() local 4352 netdev_queue_config(bp->dev, i, &qcfg); in bnxt_init_ring_struct() 4353 rxr->rx_page_size = qcfg.rx_page_size; in bnxt_init_ring_struct() 15998 struct netdev_queue_config *qcfg) in bnxt_queue_default_qcfg() argument 16000 qcfg->rx_page_size = BNXT_RX_PAGE_SIZE; in bnxt_queue_default_qcfg() 16004 struct netdev_queue_config *qcfg, in bnxt_validate_qcfg() argument 16011 qcfg->rx_page_size != BNXT_RX_PAGE_SIZE) in bnxt_validate_qcfg() 16014 if (!is_power_of_2(qcfg->rx_page_size)) in bnxt_validate_qcfg() 16017 if (qcfg->rx_page_size < BNXT_RX_PAGE_SIZE || in bnxt_validate_qcfg() 16018 qcfg->rx_page_size > BNXT_MAX_RX_PAGE_SIZE) in bnxt_validate_qcfg() [all …]
|
| /linux/drivers/net/ethernet/meta/fbnic/ |
| H A D | fbnic_txrx.c | 2832 struct netdev_queue_config *qcfg, in fbnic_queue_mem_alloc() argument 2887 struct netdev_queue_config *qcfg, in fbnic_queue_start() argument
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_main.c | 5556 struct netdev_queue_config *qcfg, in mlx5e_queue_mem_alloc() argument 5619 struct netdev_queue_config *qcfg, in mlx5e_queue_start() argument
|