| /linux/net/shaper/ |
| H A D | shaper.c | 136 const struct net_shaper *shaper, in net_shaper_fill_one() argument 146 net_shaper_fill_handle(msg, &shaper->parent, in net_shaper_fill_one() 148 net_shaper_fill_handle(msg, &shaper->handle, in net_shaper_fill_one() 150 ((shaper->bw_min || shaper->bw_max || shaper->burst) && in net_shaper_fill_one() 151 nla_put_u32(msg, NET_SHAPER_A_METRIC, shaper->metric)) || in net_shaper_fill_one() 152 (shaper->bw_min && in net_shaper_fill_one() 153 nla_put_uint(msg, NET_SHAPER_A_BW_MIN, shaper->bw_min)) || in net_shaper_fill_one() 154 (shaper->bw_max && in net_shaper_fill_one() 155 nla_put_uint(msg, NET_SHAPER_A_BW_MAX, shaper->bw_max)) || in net_shaper_fill_one() 156 (shaper->burst && in net_shaper_fill_one() [all …]
|
| H A D | Makefile | 8 obj-y += shaper.o shaper_nl_gen.o
|
| /linux/tools/testing/selftests/drivers/net/ |
| H A D | shaper.py | 165 shaper = nl_shaper.get({'ifindex': cfg.ifindex, 167 ksft_eq(shaper, {'ifindex': cfg.ifindex, 219 shaper = nl_shaper.get({'ifindex': cfg.ifindex, 221 ksft_eq(shaper, {'ifindex': cfg.ifindex, 225 shaper = nl_shaper.get({'ifindex': cfg.ifindex, 227 ksft_eq(shaper, {'ifindex': cfg.ifindex, 257 shaper = nl_shaper.get({'ifindex': cfg.ifindex, 259 ksft_eq(shaper, {'ifindex': cfg.ifindex,
|
| H A D | Makefile | 23 shaper.py \
|
| /linux/net/sched/ |
| H A D | sch_mqprio.c | 26 u16 shaper; member 48 if (priv->shaper != TC_MQPRIO_SHAPER_DCB) in mqprio_enable_offload() 56 mqprio.shaper = priv->shaper; in mqprio_enable_offload() 274 priv->shaper = nla_get_u16(tb[TCA_MQPRIO_SHAPER]); in mqprio_parse_nlattr() 278 if (priv->shaper != TC_MQPRIO_SHAPER_BW_RATE) { in mqprio_parse_nlattr() 307 if (priv->shaper != TC_MQPRIO_SHAPER_BW_RATE) { in mqprio_parse_nlattr() 594 nla_put_u16(skb, TCA_MQPRIO_SHAPER, priv->shaper)) in mqprio_dump()
|
| /linux/drivers/gpu/drm/amd/display/dc/resource/dcn30/ |
| H A D | dcn30_resource.h | 88 struct dc_transfer_func **shaper); 94 struct dc_transfer_func **shaper);
|
| H A D | dcn30_resource.c | 1466 struct dc_transfer_func **shaper) in dcn30_acquire_post_bldn_3dlut() argument 1472 ASSERT(*lut == NULL && *shaper == NULL); in dcn30_acquire_post_bldn_3dlut() 1474 *shaper = NULL; in dcn30_acquire_post_bldn_3dlut() 1479 *shaper = pool->mpc_shaper[i]; in dcn30_acquire_post_bldn_3dlut() 1501 struct dc_transfer_func **shaper) in dcn30_release_post_bldn_3dlut() argument 1507 if (pool->mpc_lut[i] == *lut && pool->mpc_shaper[i] == *shaper) { in dcn30_release_post_bldn_3dlut() 1511 *shaper = NULL; in dcn30_release_post_bldn_3dlut()
|
| /linux/Documentation/networking/ |
| H A D | netdevices.rst | 214 lock if the driver implements queue management or shaper API. 219 lock if the driver implements queue management or shaper API. 232 lock if the driver implements queue management or shaper API. 240 lock if the driver implements queue management or shaper API. 248 lock if the driver implements queue management or shaper API. 256 lock if the driver implements queue management or shaper API. 297 implements queue management or shaper API. 301 the driver implements queue management or shaper API. 369 All net shaper callbacks are invoked while holding the netdev instance
|
| /linux/drivers/net/ethernet/intel/ice/virt/ |
| H A D | queues.c | 622 if (qbw->cfg[i].shaper.peak != 0 && vf->max_tx_rate != 0 && in ice_vc_cfg_q_bw() 623 qbw->cfg[i].shaper.peak > vf->max_tx_rate) { in ice_vc_cfg_q_bw() 630 if (qbw->cfg[i].shaper.committed != 0 && vf->min_tx_rate != 0 && in ice_vc_cfg_q_bw() 631 qbw->cfg[i].shaper.committed < vf->min_tx_rate) { in ice_vc_cfg_q_bw() 654 vf->qs_bw[i].peak = qbw->cfg[i].shaper.peak; in ice_vc_cfg_q_bw() 655 vf->qs_bw[i].committed = qbw->cfg[i].shaper.committed; in ice_vc_cfg_q_bw()
|
| H A D | virtchnl.c | 459 cfg->shaper.committed = vsi_ctx->sched.bw_t_info[i].cir_bw.bw; in ice_vc_get_qos_caps() 460 cfg->shaper.peak = vsi_ctx->sched.bw_t_info[i].eir_bw.bw; in ice_vc_get_qos_caps()
|
| /linux/include/net/ |
| H A D | net_shaper.h | 97 const struct net_shaper *shaper,
|
| H A D | pkt_sched.h | 169 u16 shaper; member
|
| /linux/net/ |
| H A D | Makefile | 82 obj-$(CONFIG_NET_SHAPER) += shaper/
|
| /linux/drivers/gpu/drm/amd/display/dc/inc/ |
| H A D | core_types.h | 197 struct dc_transfer_func **shaper); 203 struct dc_transfer_func **shaper);
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_main.c | 5064 const struct net_shaper *shaper, in iavf_verify_shaper() argument 5070 if (shaper->handle.scope == NET_SHAPER_SCOPE_QUEUE) { in iavf_verify_shaper() 5071 vf_max = adapter->qos_caps->cap[0].shaper.peak; in iavf_verify_shaper() 5072 if (vf_max && shaper->bw_max > vf_max) { in iavf_verify_shaper() 5074 shaper->bw_max, shaper->handle.id, in iavf_verify_shaper() 5084 const struct net_shaper *shaper, in iavf_shaper_set() argument 5088 const struct net_shaper_handle *handle = &shaper->handle; in iavf_shaper_set() 5097 ret = iavf_verify_shaper(binding, shaper, extack); in iavf_shaper_set() 5103 tx_ring->q_shaper.bw_min = div_u64(shaper->bw_min, 1000); in iavf_shaper_set() 5104 tx_ring->q_shaper.bw_max = div_u64(shaper->bw_max, 1000); in iavf_shaper_set()
|
| H A D | iavf_virtchnl.c | 1795 qs_bw_cfg->cfg[inx].shaper.peak = q_shaper->bw_max; in iavf_cfg_queues_bw() 1796 qs_bw_cfg->cfg[inx].shaper.committed = q_shaper->bw_min; in iavf_cfg_queues_bw()
|
| /linux/Documentation/networking/device_drivers/ethernet/intel/ |
| H A D | iavf.rst | 156 The shaper bw_rlimit parameter is optional. 164 queues 16@0 16@16 hw 1 mode channel shaper bw_rlimit min_rate 1Gbit 2Gbit 178 shaper bw_rlimit: for each tc, sets minimum and maximum bandwidth rates.
|
| H A D | i40e.rst | 677 The shaper bw_rlimit parameter is optional. 685 queues 16@0 16@16 hw 1 mode channel shaper bw_rlimit min_rate 1Gbit 2Gbit 699 shaper bw_rlimit: for each tc, sets minimum and maximum bandwidth rates.
|
| /linux/drivers/gpu/drm/amd/display/dc/resource/dcn32/ |
| H A D | dcn32_resource.c | 1635 struct dc_transfer_func **shaper) in dcn32_acquire_post_bldn_3dlut() argument 1639 ASSERT(*lut == NULL && *shaper == NULL); in dcn32_acquire_post_bldn_3dlut() 1641 *shaper = NULL; in dcn32_acquire_post_bldn_3dlut() 1645 *shaper = pool->mpc_shaper[mpcc_id]; in dcn32_acquire_post_bldn_3dlut() 1656 struct dc_transfer_func **shaper) in dcn32_release_post_bldn_3dlut() argument 1662 if (pool->mpc_lut[i] == *lut && pool->mpc_shaper[i] == *shaper) { in dcn32_release_post_bldn_3dlut() 1666 *shaper = NULL; in dcn32_release_post_bldn_3dlut()
|
| /linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
| H A D | amdgpu_dm_color.c | 1108 const struct drm_color_lut *shaper = NULL, *lut3d = NULL; in amdgpu_dm_verify_lut3d_size() local 1114 shaper = __extract_blob_lut(dm_plane_state->shaper_lut, &size); in amdgpu_dm_verify_lut3d_size() 1116 if (shaper && size != exp_size) { in amdgpu_dm_verify_lut3d_size()
|
| /linux/include/linux/avf/ |
| H A D | virtchnl.h | 1616 struct virtchnl_shaper_bw shaper; member 1637 struct virtchnl_shaper_bw shaper; member
|
| /linux/drivers/gpu/drm/amd/display/dc/ |
| H A D | dc_types.h | 1396 const struct dc_transfer_func *shaper; member
|
| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | cxgb4_tc_mqprio.c | 28 } else if (mqprio->shaper != TC_MQPRIO_SHAPER_BW_RATE) { in cxgb4_mqprio_validate()
|
| /linux/drivers/gpu/drm/amd/display/dc/hwss/dcn401/ |
| H A D | dcn401_hwseq.c | 458 if (mcm_luts.shaper && mcm_luts.lut3d_data.mpc_3dlut_enable) { in dcn401_populate_mcm_luts() 460 if (mcm_luts.shaper->type == TF_TYPE_HWPWL) in dcn401_populate_mcm_luts() 461 m_lut_params.pwl = &mcm_luts.shaper->pwl; in dcn401_populate_mcm_luts() 462 else if (mcm_luts.shaper->type == TF_TYPE_DISTRIBUTED_POINTS) { in dcn401_populate_mcm_luts() 465 mcm_luts.shaper, in dcn401_populate_mcm_luts()
|
| /linux/Documentation/networking/device_drivers/ethernet/ti/ |
| H A D | cpsw.rst | 77 - S - shaper configured
|