/linux/net/sched/ |
H A D | sch_mqprio_lib.c | 38 if (qopt->offset[i] >= dev->real_num_tx_queues || in mqprio_validate_queue_counts() 39 last > dev->real_num_tx_queues) { in mqprio_validate_queue_counts() 43 i, dev->real_num_tx_queues); in mqprio_validate_queue_counts()
|
H A D | sch_mq.c | 119 if (ntx < dev->real_num_tx_queues) in mq_attach()
|
H A D | sch_multiq.c | 186 qopt->bands = qdisc_dev(sch)->real_num_tx_queues; in multiq_tune()
|
H A D | act_skbedit.c | 74 skb->dev->real_num_tx_queues > params->queue_mapping) { in tcf_skbedit_act()
|
/linux/drivers/net/ethernet/cavium/liquidio/ |
H A D | octeon_network.h | 563 for (i = 0; i < netdev->real_num_tx_queues; i++) in stop_txqs() 576 for (i = 0; i < netdev->real_num_tx_queues; i++) { in wake_txqs() 597 for (i = 0; i < netdev->real_num_tx_queues; i++) in start_txqs()
|
/linux/drivers/net/ |
H A D | ifb.c | 219 for (i = 0; i < dev->real_num_tx_queues; i++) in ifb_get_strings() 233 dev->real_num_tx_queues); in ifb_get_sset_count() 270 for (i = 0; i < dev->real_num_tx_queues; i++) { in ifb_get_ethtool_stats()
|
H A D | xen-netfront.c | 364 unsigned int num_queues = dev->real_num_tx_queues; in xennet_open() 589 unsigned int num_queues = dev->real_num_tx_queues; in xennet_select_queue() 647 unsigned int num_queues = dev->real_num_tx_queues; in xennet_xdp_xmit() 720 unsigned int num_queues = dev->real_num_tx_queues; in xennet_start_xmit() 866 unsigned int num_queues = np->queues ? dev->real_num_tx_queues : 0; in xennet_close() 884 for (i = 0; i < info->netdev->real_num_tx_queues; i++) { in xennet_destroy_queues() 1595 unsigned int num_queues = dev->real_num_tx_queues; in xennet_poll_controller() 1656 bpf_prog_add(prog, dev->real_num_tx_queues); in xennet_xdp_set() 1658 for (i = 0; i < dev->real_num_tx_queues; ++i) in xennet_xdp_set() 1662 for (i = 0; i < dev->real_num_tx_queues; ++i) in xennet_xdp_set() [all …]
|
H A D | veth.c | 158 for (i = 0; i < dev->real_num_tx_queues; i++) in veth_get_strings() 174 VETH_TQ_STATS_LEN * dev->real_num_tx_queues + in veth_get_sset_count() 233 tx_idx += (i % dev->real_num_tx_queues) * VETH_TQ_STATS_LEN; in veth_get_ethtool_stats() 242 pp_idx = idx + dev->real_num_tx_queues * VETH_TQ_STATS_LEN; in veth_get_ethtool_stats() 251 channels->tx_count = dev->real_num_tx_queues; in veth_get_channels() 1281 if (peer && peer->real_num_tx_queues <= dev->real_num_rx_queues) { in veth_set_xdp_features() 1312 if (priv->_xdp_prog && peer && ch->rx_count < peer->real_num_tx_queues) in veth_set_channels() 1593 if (dev->real_num_rx_queues < peer->real_num_tx_queues) { in veth_xdp_set()
|
/linux/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_ethtool.c | 518 for (i = 0; i < netdev->real_num_tx_queues; i++) in fun_set_coalesce() 531 chan->tx_count = netdev->real_num_tx_queues; in fun_get_channels() 540 if (chan->tx_count == netdev->real_num_tx_queues && in fun_set_channels() 622 n = (dev->real_num_tx_queues + 1) * ARRAY_SIZE(txq_stat_names) + in fun_get_sset_count() 652 for (i = 0; i < netdev->real_num_tx_queues; i++) { in fun_get_strings() 770 totals = data + netdev->real_num_tx_queues * ARRAY_SIZE(txq_stat_names); in fun_get_ethtool_stats() 772 for (i = 0; i < netdev->real_num_tx_queues; i++) { in fun_get_ethtool_stats()
|
H A D | funeth_main.c | 455 qset->ntxqs = netdev->real_num_tx_queues; in fun_free_rings() 839 .ntxqs = netdev->real_num_tx_queues, in funeth_open() 890 for (i = 0; i < netdev->real_num_tx_queues; i++) { in fun_get_stats64() 1592 newqs->ntxqs = dev->real_num_tx_queues; in fun_replace_queues() 1625 unsigned int keep_tx = min(dev->real_num_tx_queues, ntx); in fun_change_num_queues() 1632 .ntxqs = dev->real_num_tx_queues, in fun_change_num_queues() 1673 if (ntx < dev->real_num_tx_queues) in fun_change_num_queues() 1680 if (ntx > dev->real_num_tx_queues) in fun_change_num_queues()
|
/linux/drivers/s390/net/ |
H A D | qeth_ethtool.c | 209 channels->tx_count = dev->real_num_tx_queues; in qeth_get_channels() 238 channels->tx_count < dev->real_num_tx_queues) in qeth_set_channels()
|
/linux/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_bpf.c | 62 q_idx = smp_processor_id() % ndev->real_num_tx_queues; in mana_xdp_xmit()
|
/linux/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_qos.c | 481 ndev->num_tc, ndev->real_num_tx_queues); in sparx5_tc_mqprio_add() 491 ndev->num_tc, ndev->real_num_tx_queues); in sparx5_tc_mqprio_del()
|
/linux/net/core/ |
H A D | netpoll.c | 108 if (unlikely(q_index >= dev->real_num_tx_queues)) { in queue_process() 109 q_index = q_index % dev->real_num_tx_queues; in queue_process()
|
H A D | net-sysfs.c | 2110 real_tx = dev->real_num_tx_queues; in register_queue_kobjects() 2145 real_tx = ndev->real_num_tx_queues; in queue_change_owner() 2165 real_tx = dev->real_num_tx_queues; in remove_queue_kobjects() 2172 dev->real_num_tx_queues = 0; in remove_queue_kobjects()
|
/linux/Documentation/networking/net_cachelines/ |
H A D | net_device.rst | 118 unsigned_int real_num_tx_queues read_mostly …
|
/linux/drivers/net/xen-netback/ |
H A D | interface.c | 193 dev->real_num_tx_queues; in xenvif_select_queue() 198 return skb_get_hash_raw(skb) % dev->real_num_tx_queues; in xenvif_select_queue()
|
/linux/net/shaper/ |
H A D | shaper.c | 479 shaper->handle.id >= binding->netdev->real_num_tx_queues) { in net_shaper_validate_caps() 483 binding->netdev->real_num_tx_queues); in net_shaper_validate_caps() 1404 for (i = txq; i < dev->real_num_tx_queues; ++i) { in net_shaper_set_real_num_tx_queues()
|
/linux/net/xdp/ |
H A D | xsk.c | 104 if (queue_id < dev->real_num_tx_queues) in xsk_get_pool_from_qid() 128 dev->real_num_tx_queues)) in xsk_reg_pool_at_qid() 133 if (queue_id < dev->real_num_tx_queues) in xsk_reg_pool_at_qid() 800 if (xs->queue_id >= xs->dev->real_num_tx_queues) in __xsk_generic_xmit()
|
/linux/drivers/net/ethernet/myricom/myri10ge/ |
H A D | myri10ge.c | 977 if (mgp->dev->real_num_tx_queues > 1) in myri10ge_reset() 985 if (status != 0 && mgp->dev->real_num_tx_queues > 1) { in myri10ge_reset() 1426 if ((ss->mgp->dev->real_num_tx_queues > 1) && in myri10ge_tx_done() 1551 if ((mgp->dev->real_num_tx_queues == 1) && (ss != mgp->ss)) { in myri10ge_intr() 2239 if (slice == 0 || (mgp->dev->real_num_tx_queues > 1)) { in myri10ge_get_txrx() 2315 if (mgp->dev->real_num_tx_queues > 1) in myri10ge_open() 2404 if (slice == 0 || mgp->dev->real_num_tx_queues > 1) in myri10ge_open() 2856 if ((mgp->dev->real_num_tx_queues > 1) && tx->queue_active == 0) { in myri10ge_xmit()
|
/linux/drivers/net/ethernet/ibm/ |
H A D | ibmveth.c | 668 for (i = 0; i < netdev->real_num_tx_queues; i++) { in ibmveth_open() 815 for (i = 0; i < netdev->real_num_tx_queues; i++) in ibmveth_close() 1111 channels->tx_count = netdev->real_num_tx_queues; in ibmveth_get_channels() 1121 unsigned int old = netdev->real_num_tx_queues, in ibmveth_set_channels()
|
/linux/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_stats.c | 171 #define MAX_Q(lif) ((lif)->netdev->real_num_tx_queues)
|
/linux/drivers/net/netdevsim/ |
H A D | netdev.c | 56 if (dev->real_num_tx_queues != peer_dev->num_rx_queues) in nsim_start_peer_tx_queue() 76 if (rx_dev->real_num_tx_queues != tx_dev->num_rx_queues) in nsim_stop_tx_queue()
|
/linux/drivers/net/hyperv/ |
H A D | netvsc_drv.c | 283 if (q_idx < 0 || skb->ooo_okay || q_idx >= ndev->real_num_tx_queues) { in netvsc_pick_tx() 323 while (txq >= ndev->real_num_tx_queues) in netvsc_select_queue() 324 txq -= ndev->real_num_tx_queues; in netvsc_select_queue() 520 if (hash != 0 && net->real_num_tx_queues > 1) { in netvsc_xmit()
|
/linux/drivers/net/ethernet/intel/igb/ |
H A D | igb_xsk.c | 110 qid >= netdev->real_num_tx_queues) in igb_xsk_pool_enable()
|