/linux/drivers/thunderbolt/ |
H A D | path.c | 163 path->hops = kcalloc(num_hops, sizeof(*path->hops), GFP_KERNEL); in tb_path_discover() 164 if (!path->hops) { in tb_path_discover() 198 path->hops[i].in_port = p; in tb_path_discover() 199 path->hops[i].in_hop_index = h; in tb_path_discover() 200 path->hops[i].in_counter_index = -1; in tb_path_discover() 201 path->hops[i].out_port = out_port; in tb_path_discover() 202 path->hops[i].next_hop_index = next_hop; in tb_path_discover() 204 tb_dump_hop(&path->hops[i], &hop); in tb_path_discover() 270 path->hops in tb_path_alloc() [all...] |
H A D | test.c | 913 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane0() 914 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane0() 975 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane1() 976 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane1() 1055 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane1_chain() 1056 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane1_chain() 1135 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane1_chain_reverse() 1136 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane1_chain_reverse() 1227 in_port = path->hops[i].in_port; in tb_test_path_mixed_chain() 1228 out_port = path->hops[ in tb_test_path_mixed_chain() [all...] |
H A D | tunnel.c | 1120 paths[TB_DP_VIDEO_PATH_OUT]->hops[0].in_hop_index, in tb_dp_activate() 1121 paths[TB_DP_AUX_PATH_OUT]->hops[0].in_hop_index, in tb_dp_activate() 1122 paths[TB_DP_AUX_PATH_IN]->hops[last].next_hop_index); in tb_dp_activate() 1125 paths[TB_DP_VIDEO_PATH_OUT]->hops[last].next_hop_index, in tb_dp_activate() 1126 paths[TB_DP_AUX_PATH_IN]->hops[0].in_hop_index, in tb_dp_activate() 1127 paths[TB_DP_AUX_PATH_OUT]->hops[last].next_hop_index); in tb_dp_activate() 1781 hop = &path->hops[0]; in tb_dma_init_rx_path() 1789 ret = tb_dma_reserve_credits(&path->hops[i], credits); in tb_dma_init_rx_path() 1958 if (tb_port_is_nhi(path->hops[0].in_port)) in tb_tunnel_match_dma() 1960 else if (tb_port_is_nhi(path->hops[pat in tb_tunnel_match_dma() [all...] |
H A D | tb.h | 396 * @TB_PATH_INTERNAL: Activate on the intermediate hops (not the first/last) 398 * @TB_PATH_ALL: Activate on all hops on the path 422 * @hops: Path hops 423 * @path_length: How many hops the path uses 426 * A path consists of a number of hops (see &struct tb_path_hop). To 443 struct tb_path_hop *hops; member 1196 * @path: Path whose hops to iterate 1202 for ((hop) = &(path)->hops[0]; \ 1203 (hop) <= &(path)->hops[(pat [all...] |
/linux/drivers/accel/habanalabs/common/mmu/ |
H A D | mmu_v2.c | 19 * page tables hops related to this context. 37 * - Free DRAM default page mapping hops 245 static int hl_mmu_v2_get_tlb_info(struct hl_ctx *ctx, u64 virt_addr, struct hl_mmu_hop_info *hops) in hl_mmu_v2_get_tlb_info() argument 262 hops->range_type = HL_VA_RANGE_TYPE_DRAM; in hl_mmu_v2_get_tlb_info() 264 hops->scrambled_vaddr = hdev->asic_funcs->scramble_addr(hdev, virt_addr); in hl_mmu_v2_get_tlb_info() 266 hops->hop_info[0].hop_addr = hl_mmu_dr_get_phys_hop0_addr(ctx); in hl_mmu_v2_get_tlb_info() 267 hops->hop_info[0].hop_pte_addr = hl_mmu_get_hop_pte_phys_addr(ctx, mmu_prop, 0, in hl_mmu_v2_get_tlb_info() 268 hops->hop_info[0].hop_addr, in hl_mmu_v2_get_tlb_info() 269 hops->scrambled_vaddr); in hl_mmu_v2_get_tlb_info() 270 if (hops in hl_mmu_v2_get_tlb_info() [all...] |
H A D | mmu.c | 103 * page tables hops related to this context. 143 * - Free DRAM default page mapping hops 477 struct hl_mmu_hop_info *hops, in hl_mmu_pa_page_with_offset() argument 485 if (hops->unscrambled_paddr) in hl_mmu_pa_page_with_offset() 486 tmp_phys_addr = hops->unscrambled_paddr; in hl_mmu_pa_page_with_offset() 488 tmp_phys_addr = hops->hop_info[hops->used_hops - 1].hop_pte_val; in hl_mmu_pa_page_with_offset() 490 if (hops->range_type == HL_VA_RANGE_TYPE_HOST_HUGE) in hl_mmu_pa_page_with_offset() 492 else if (hops->range_type == HL_VA_RANGE_TYPE_HOST) in hl_mmu_pa_page_with_offset() 497 if ((hops in hl_mmu_pa_page_with_offset() 534 struct hl_mmu_hop_info hops; hl_mmu_va_to_pa() local 549 hl_mmu_get_tlb_info(struct hl_ctx * ctx,u64 virt_addr,struct hl_mmu_hop_info * hops) hl_mmu_get_tlb_info() argument 1155 hl_mmu_hr_get_tlb_info(struct hl_ctx * ctx,u64 virt_addr,struct hl_mmu_hop_info * hops,struct hl_hr_mmu_funcs * hr_func) hl_mmu_hr_get_tlb_info() argument [all...] |
H A D | mmu_v1.c | 80 /* need only pte 0 in hops 0 and 1 */ in dram_default_mapping_init() 179 * page tables hops related to this context. 196 * - Free DRAM default page mapping hops 335 * there are only 3 hops rather than 4. Currently the DRAM allocation in hl_mmu_v1_map() 383 dev_err(hdev->dev, "DRAM mapping should not allocate more hops\n"); in hl_mmu_v1_map() 454 struct hl_mmu_hop_info *hops) in hl_mmu_v1_get_tlb_info() argument 487 /* huge pages use lesser hops */ in hl_mmu_v1_get_tlb_info() 491 hops->hop_info[0].hop_addr = hl_mmu_dr_get_phys_hop0_addr(ctx); in hl_mmu_v1_get_tlb_info() 492 hops->hop_info[0].hop_pte_addr = in hl_mmu_v1_get_tlb_info() 494 hops in hl_mmu_v1_get_tlb_info() [all...] |
H A D | mmu_v2_hr.c | 76 * page tables hops related to this context. 93 * - Free DRAM default page mapping hops 220 * there are only 4 hops rather than 5. Currently the DRAM allocation in _hl_mmu_v2_hr_map() 335 struct hl_mmu_hop_info *hops, in hl_mmu_v2_hr_get_tlb_mapping_params() argument 354 hops->range_type = HL_VA_RANGE_TYPE_DRAM; in hl_mmu_v2_hr_get_tlb_mapping_params() 358 hops->range_type = HL_VA_RANGE_TYPE_HOST; in hl_mmu_v2_hr_get_tlb_mapping_params() 362 hops->range_type = HL_VA_RANGE_TYPE_HOST_HUGE; in hl_mmu_v2_hr_get_tlb_mapping_params() 371 struct hl_mmu_hop_info *hops) in hl_mmu_v2_hr_get_tlb_info() argument 373 return hl_mmu_hr_get_tlb_info(ctx, virt_addr, hops, in hl_mmu_v2_hr_get_tlb_info()
|
/linux/Documentation/networking/ |
H A D | nexthop-group-resilient.rst | 9 weights of constituent next hops. 48 constituent next hops: a hash table. The selection algorithm uses SKB hash 55 the individual next hops is arbitrary. Therefore when a next hop is deleted 56 the buckets that held it are simply reassigned to other next hops:: 69 When weights of next hops in a group are altered, it may be possible to 83 cause bucket allocation change, the wants counts for individual next hops 86 Next hops that have fewer buckets than their wants count, are called 88 overweight (and therefore no underweight) next hops in the group, it is 97 After assigning wants counts to next hops, an "upkeep" algorithm runs. For 105 underweight next hops [all...] |
/linux/block/ |
H A D | bdev.c | 533 * @hops: holder ops 541 const struct blk_holder_ops *hops) in bd_may_claim() argument 552 if (WARN_ON_ONCE(bdev->bd_holder_ops != hops)) in bd_may_claim() 573 * @hops: holder ops. 583 const struct blk_holder_ops *hops) in bd_prepare_to_claim() argument 592 if (!bd_may_claim(bdev, holder, hops)) { in bd_prepare_to_claim() 629 * @hops: block device holder operations 635 const struct blk_holder_ops *hops) in bd_finish_claiming() argument 640 BUG_ON(!bd_may_claim(bdev, holder, hops)); in bd_finish_claiming() 650 bdev->bd_holder_ops = hops; in bd_finish_claiming() 927 bdev_open(struct block_device * bdev,blk_mode_t mode,void * holder,const struct blk_holder_ops * hops,struct file * bdev_file) bdev_open() argument 1035 bdev_file_open_by_dev(dev_t dev,blk_mode_t mode,void * holder,const struct blk_holder_ops * hops) bdev_file_open_by_dev() argument 1072 bdev_file_open_by_path(const char * path,blk_mode_t mode,void * holder,const struct blk_holder_ops * hops) bdev_file_open_by_path() argument [all...] |
/linux/drivers/firmware/arm_scmi/ |
H A D | sensors.c | 253 if (!ph->hops->protocol_msg_check(ph, in scmi_sensor_attributes_get() 257 if (!ph->hops->protocol_msg_check(ph, in scmi_sensor_attributes_get() 368 iter = ph->hops->iter_response_init(ph, &ops, s->intervals.count, in scmi_sensor_update_intervals() 375 return ph->hops->iter_response_run(iter); in scmi_sensor_update_intervals() 506 iter = ph->hops->iter_response_init(ph, &ops, s->num_axis, in scmi_sensor_axis_extended_names_get() 517 ret = ph->hops->iter_response_run(iter); in scmi_sensor_axis_extended_names_get() 547 iter = ph->hops->iter_response_init(ph, &ops, s->num_axis, in scmi_sensor_axis_description() 554 ret = ph->hops->iter_response_run(iter); in scmi_sensor_axis_description() 664 ph->hops->extended_name_get(ph, SENSOR_NAME_GET, s->id, in iter_sens_descr_process_response() 703 iter = ph->hops in scmi_sensor_description_get() [all...] |
H A D | perf.c | 230 if (!ph->hops->protocol_msg_check(ph, PERF_NOTIFY_LEVEL, NULL)) in scmi_perf_attributes_get() 233 if (!ph->hops->protocol_msg_check(ph, PERF_NOTIFY_LIMITS, NULL)) in scmi_perf_attributes_get() 328 ph->hops->extended_name_get(ph, PERF_DOMAIN_NAME_GET, in scmi_perf_domain_attributes_get() 487 iter = ph->hops->iter_response_init(ph, &ops, MAX_OPPS, in scmi_perf_describe_levels_get() 494 ret = ph->hops->iter_response_run(iter); in scmi_perf_describe_levels_get() 569 ph->hops->fastchannel_db_ring(fci->set_db); in __scmi_perf_limits_set() 723 ph->hops->fastchannel_db_ring(fci->set_db); in __scmi_perf_level_set() 849 ph->hops->fastchannel_init(ph, PERF_DESCRIBE_FASTCHANNEL, in scmi_perf_domain_init_fc() 854 ph->hops->fastchannel_init(ph, PERF_DESCRIBE_FASTCHANNEL, in scmi_perf_domain_init_fc() 860 ph->hops in scmi_perf_domain_init_fc() [all...] |
H A D | powercap.c | 164 if (!ph->hops->protocol_msg_check(ph, in scmi_powercap_attributes_get() 168 if (!ph->hops->protocol_msg_check(ph, in scmi_powercap_attributes_get() 291 ph->hops->extended_name_get(ph, POWERCAP_DOMAIN_NAME_GET, in scmi_powercap_domain_attributes_get() 427 ph->hops->fastchannel_db_ring(fci->set_db); in __scmi_powercap_cap_set() 547 ph->hops->fastchannel_db_ring(fci->set_db); in scmi_powercap_pai_set() 719 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc() 725 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc() 730 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc() 736 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc()
|
H A D | pinctrl.c | 237 ret = ph->hops->extended_name_get(ph, PINCTRL_NAME_GET, in scmi_pinctrl_attributes() 311 iter = ph->hops->iter_response_init(ph, &ops, size, in scmi_pinctrl_list_associations() 318 return ph->hops->iter_response_run(iter); in scmi_pinctrl_list_associations() 426 iter = ph->hops->iter_response_init(ph, &ops, max_configs, in scmi_pinctrl_settings_get() 433 return ph->hops->iter_response_run(iter); in scmi_pinctrl_settings_get() 476 int max_msg_size = ph->hops->get_max_msg_size(ph); in scmi_pinctrl_settings_conf()
|
H A D | clock.c | 215 if (!ph->hops->protocol_msg_check(ph, CLOCK_RATE_NOTIFY, NULL)) in scmi_clock_protocol_attributes_get() 218 if (!ph->hops->protocol_msg_check(ph, in scmi_clock_protocol_attributes_get() 307 iter = ph->hops->iter_response_init(ph, &ops, 0, in scmi_clock_possible_parents() 314 ret = ph->hops->iter_response_run(iter); in scmi_clock_possible_parents() 386 ph->hops->extended_name_get(ph, CLOCK_NAME_GET, clk_id, in scmi_clock_attributes_get() 529 iter = ph->hops->iter_response_init(ph, &ops, SCMI_MAX_NUM_RATES, in scmi_clock_describe_rates_get() 536 ret = ph->hops->iter_response_run(iter); in scmi_clock_describe_rates_get()
|
H A D | power.c | 103 if (!ph->hops->protocol_msg_check(ph, POWER_STATE_NOTIFY, NULL)) in scmi_power_attributes_get() 146 ph->hops->extended_name_get(ph, POWER_DOMAIN_NAME_GET, in scmi_power_domain_attributes_get()
|
H A D | system.c | 163 if (!ph->hops->protocol_msg_check(ph, SYSTEM_POWER_STATE_NOTIFY, NULL)) in scmi_system_protocol_init()
|
H A D | protocols.h | 181 const struct scmi_proto_helpers_ops *hops; member
|
/linux/net/ipv6/ |
H A D | exthdrs.c | 1082 int hops; in ipv6_push_rthdr0() local 1089 hops = ihdr->rt_hdr.hdrlen >> 1; in ipv6_push_rthdr0() 1091 if (hops > 1) in ipv6_push_rthdr0() 1093 (hops - 1) * sizeof(struct in6_addr)); in ipv6_push_rthdr0() 1095 phdr->addr[hops - 1] = **addr_p; in ipv6_push_rthdr0() 1107 int plen, hops; in ipv6_push_rthdr4() local 1115 hops = sr_ihdr->first_segment + 1; in ipv6_push_rthdr4() 1117 (hops - 1) * sizeof(struct in6_addr)); in ipv6_push_rthdr4() 1122 if (sr_ihdr->hdrlen > hops * 2) { in ipv6_push_rthdr4() 1125 tlvs_offset = (1 + hops * in ipv6_push_rthdr4() [all...] |
/linux/arch/s390/kernel/ |
H A D | dis.c | 509 int start, end, opsize, hops, i; in show_code() local 529 for (i = 0, hops = 0; start + i < 32 && hops < 3; hops++) { in show_code() 542 hops = 0; in show_code() 543 while (start < end && hops < 8) { in show_code() 565 hops++; in show_code()
|
/linux/include/linux/ |
H A D | topology.h | 71 * between NUMA nodes 2-hops apart. 270 extern const struct cpumask *sched_numa_hop_mask(unsigned int node, unsigned int hops); 278 sched_numa_hop_mask(unsigned int node, unsigned int hops) in sched_numa_hop_mask() argument
|
H A D | blkdev.h | 1773 const struct blk_holder_ops *hops); 1775 void *holder, const struct blk_holder_ops *hops); 1777 const struct blk_holder_ops *hops);
|
/linux/include/dt-bindings/usb/ |
H A D | pd.h | 367 * <7:6> :: USB2 hub hops consumed 404 #define VDO_ACABLE2(mtemp, stemp, u3p, trans, phy, ele, u4, hops, u2, u32, lane, iso, gen) \ argument 407 | ((hops) & 0x3) << 6 | (u2) << 5 | (u32) << 4 | (lane) << 3 \
|
/linux/drivers/firmware/arm_scmi/vendors/imx/ |
H A D | imx-sm-misc.c | 208 int max_msg_size = ph->hops->get_max_msg_size(ph); in scmi_imx_misc_ctrl_get() 247 int max_msg_size = ph->hops->get_max_msg_size(ph); in scmi_imx_misc_ctrl_set()
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-bus-rapidio | 156 (RO) number of hops on the path to the switch
|