/linux/rust/kernel/ |
H A D | devres.rs | 22 /// [`Devres`] inner data accessed from [`Devres::callback`]. 106 /// [`Self::inner`] is guaranteed to be initialized and is always accessed read-only. 121 inner: Opaque<Inner<T>>, field 142 // INVARIANT: `inner` is properly initialized. in new() 143 inner <- { in new() 145 let inner = unsafe { &raw mut (*this.as_ptr()).inner }; in new() localVariable 149 // - `inner` is guaranteed to be a valid for the duration of the lifetime of `Self`. in new() 154 bindings::devm_add_action(dev.as_raw(), Some(callback), inner.cast()) in new() 166 fn inner( method 180 let inner = unsafe { &*ptr.cast::<Inner<T>>() }; devres_callback() localVariable 183 let inner = ScopeGuard::new_with_data(inner, |inner| inner.devm.complete_all()); devres_callback() localVariable [all...] |
H A D | regulator.rs | 239 /// - `inner` is a non-null wrapper over a pointer to a `struct 247 inner: NonNull<bindings::regulator>, field 259 self.inner.as_ptr(), in set_voltage() 269 let voltage = unsafe { bindings::regulator_get_voltage(self.inner.as_ptr()) }; in get_voltage() 280 let inner = from_err_ptr(unsafe { bindings::regulator_get(dev.as_raw(), name.as_ptr()) })?; in get_internal() localVariable 282 // SAFETY: We can safely trust `inner` to be a pointer to a valid in get_internal() 284 let inner = unsafe { NonNull::new_unchecked(inner) }; in get_internal() localVariable 287 inner, in get_internal() 294 to_result(unsafe { bindings::regulator_enable(self.inner in enable_internal() [all...] |
H A D | time.rs | 165 /// The `inner` value is in the range from 0 to `KTIME_MAX`. 169 inner: bindings::ktime_t, field 188 inner: C::ktime_get(), in now() 201 self.inner in as_nanos() 212 nanos: self.inner - other.inner, in sub()
|
H A D | sync.rs | 34 inner: Opaque<bindings::lock_class_key>, field 75 inner <- Opaque::ffi_init(|slot| unsafe { bindings::lockdep_register_key(slot) }) in new_dynamic() 80 self.inner.get() in as_ptr()
|
/linux/include/net/ |
H A D | inet_ecn.h | 46 static inline __u8 INET_ECN_encapsulate(__u8 outer, __u8 inner) in INET_ECN_encapsulate() argument 49 outer |= !INET_ECN_is_ce(inner) ? (inner & INET_ECN_MASK) : in INET_ECN_encapsulate() 120 static inline void ipv4_copy_dscp(unsigned int dscp, struct iphdr *inner) in ipv4_copy_dscp() argument 123 ipv4_change_dsfield(inner, INET_ECN_MASK, dscp); in ipv4_copy_dscp() 166 static inline void ipv6_copy_dscp(unsigned int dscp, struct ipv6hdr *inner) in ipv6_copy_dscp() argument 169 ipv6_change_dsfield(inner, INET_ECN_MASK, dscp); in ipv6_copy_dscp() 229 * To decapsulate the inner header at the tunnel egress, a compliant 231 * intersection of the appropriate arriving inner header (row) and outer 251 static inline int __INET_ECN_decapsulate(__u8 outer, __u8 inner, boo argument 270 INET_ECN_decapsulate(struct sk_buff * skb,__u8 outer,__u8 inner) INET_ECN_decapsulate() argument 289 __u8 inner; IP_ECN_decapsulate() local 308 __u8 inner; IP6_ECN_decapsulate() local [all...] |
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
H A D | dr_ste.c | 740 dr_ste_build_pre_check_spec(dmn, &mask->inner)) in mlx5dr_ste_build_pre_check() 1068 dr_ste_copy_mask_spec(buff, &set_param->inner, clr); in mlx5dr_ste_copy_param() 1130 bool inner, bool rx) in mlx5dr_ste_build_eth_l2_src_dst() argument 1133 sb->inner = inner; in mlx5dr_ste_build_eth_l2_src_dst() 1140 bool inner, bool rx) in mlx5dr_ste_build_eth_l3_ipv6_dst() argument 1143 sb->inner = inner; in mlx5dr_ste_build_eth_l3_ipv6_dst() 1150 bool inner, bool rx) in mlx5dr_ste_build_eth_l3_ipv6_src() argument 1153 sb->inner in mlx5dr_ste_build_eth_l3_ipv6_src() 1160 mlx5dr_ste_build_eth_l3_ipv4_5_tuple(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_eth_l3_ipv4_5_tuple() argument 1170 mlx5dr_ste_build_eth_l2_src(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_eth_l2_src() argument 1180 mlx5dr_ste_build_eth_l2_dst(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_eth_l2_dst() argument 1189 mlx5dr_ste_build_eth_l2_tnl(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_eth_l2_tnl() argument 1199 mlx5dr_ste_build_eth_l3_ipv4_misc(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_eth_l3_ipv4_misc() argument 1209 mlx5dr_ste_build_eth_ipv6_l3_l4(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_eth_ipv6_l3_l4() argument 1234 mlx5dr_ste_build_mpls(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_mpls() argument 1244 mlx5dr_ste_build_tnl_gre(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_tnl_gre() argument 1255 mlx5dr_ste_build_tnl_mpls_over_gre(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_cmd_caps * caps,bool inner,bool rx) mlx5dr_ste_build_tnl_mpls_over_gre() argument 1267 mlx5dr_ste_build_tnl_mpls_over_udp(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_cmd_caps * caps,bool inner,bool rx) mlx5dr_ste_build_tnl_mpls_over_udp() argument 1279 mlx5dr_ste_build_icmp(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_cmd_caps * caps,bool inner,bool rx) mlx5dr_ste_build_icmp() argument 1290 mlx5dr_ste_build_general_purpose(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_general_purpose() argument 1300 mlx5dr_ste_build_eth_l4_misc(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_eth_l4_misc() argument 1310 mlx5dr_ste_build_tnl_vxlan_gpe(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_tnl_vxlan_gpe() argument 1320 mlx5dr_ste_build_tnl_geneve(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_tnl_geneve() argument 1331 mlx5dr_ste_build_tnl_geneve_tlv_opt(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_cmd_caps * caps,bool inner,bool rx) mlx5dr_ste_build_tnl_geneve_tlv_opt() argument 1343 mlx5dr_ste_build_tnl_geneve_tlv_opt_exist(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_cmd_caps * caps,bool inner,bool rx) mlx5dr_ste_build_tnl_geneve_tlv_opt_exist() argument 1357 mlx5dr_ste_build_tnl_gtpu(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_tnl_gtpu() argument 1368 mlx5dr_ste_build_tnl_gtpu_flex_parser_0(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_cmd_caps * caps,bool inner,bool rx) mlx5dr_ste_build_tnl_gtpu_flex_parser_0() argument 1380 mlx5dr_ste_build_tnl_gtpu_flex_parser_1(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_cmd_caps * caps,bool inner,bool rx) mlx5dr_ste_build_tnl_gtpu_flex_parser_1() argument 1391 mlx5dr_ste_build_register_0(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_register_0() argument 1401 mlx5dr_ste_build_register_1(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_register_1() argument 1412 mlx5dr_ste_build_src_gvmi_qpn(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,struct mlx5dr_domain * dmn,bool inner,bool rx) mlx5dr_ste_build_src_gvmi_qpn() argument 1426 mlx5dr_ste_build_flex_parser_0(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_flex_parser_0() argument 1436 mlx5dr_ste_build_flex_parser_1(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_flex_parser_1() argument 1446 mlx5dr_ste_build_tnl_header_0_1(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask,bool inner,bool rx) mlx5dr_ste_build_tnl_header_0_1() argument [all...] |
H A D | dr_matcher.c | 413 bool inner, rx; in dr_matcher_set_ste_builders() local 428 mask.inner = matcher->mask.inner; in dr_matcher_set_ste_builders() 463 inner = false; in dr_matcher_set_ste_builders() 467 &mask, inner, rx); in dr_matcher_set_ste_builders() 471 &mask, inner, rx); in dr_matcher_set_ste_builders() 475 &mask, inner, rx); in dr_matcher_set_ste_builders() 481 &mask, dmn, inner, rx); in dr_matcher_set_ste_builders() 487 &mask, inner, rx); in dr_matcher_set_ste_builders() 492 &mask, inner, r in dr_matcher_set_ste_builders() [all...] |
H A D | dr_ste_v0.c | 37 #define DR_STE_CALC_LU_TYPE(lookup_type, rx, inner) \ argument 38 ((inner) ? DR_STE_V0_LU_TYPE_##lookup_type##_I : \ 707 bool inner, u8 *bit_mask) in dr_ste_v0_build_eth_l2_src_dst_bit_mask() argument 709 struct mlx5dr_match_spec *mask = inner ? &value->inner : &value->outer; in dr_ste_v0_build_eth_l2_src_dst_bit_mask() 742 struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer; in dr_ste_v0_build_eth_l2_src_dst_tag() 786 dr_ste_v0_build_eth_l2_src_dst_bit_mask(mask, sb->inner, sb->bit_mask); in dr_ste_v0_build_eth_l2_src_dst_init() 788 sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_SRC_DST, sb->rx, sb->inner); in dr_ste_v0_build_eth_l2_src_dst_init() 798 struct mlx5dr_match_spec *spec = sb->inner in dr_ste_v0_build_eth_l3_ipv6_dst_tag() 884 dr_ste_v0_build_eth_l2_src_or_dst_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask) dr_ste_v0_build_eth_l2_src_or_dst_bit_mask() argument 935 dr_ste_v0_build_eth_l2_src_or_dst_tag(struct mlx5dr_match_param * value,bool inner,u8 * tag) dr_ste_v0_build_eth_l2_src_or_dst_tag() argument 996 dr_ste_v0_build_eth_l2_src_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask) dr_ste_v0_build_eth_l2_src_bit_mask() argument 1068 dr_ste_v0_build_eth_l2_tnl_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask) dr_ste_v0_build_eth_l2_tnl_bit_mask() argument [all...] |
H A D | dr_ste_v1.c | 966 bool inner, u8 *bit_mask) in dr_ste_v1_build_eth_l2_src_dst_bit_mask() argument 968 struct mlx5dr_match_spec *mask = inner ? &value->inner : &value->outer; in dr_ste_v1_build_eth_l2_src_dst_bit_mask() 994 struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer; in dr_ste_v1_build_eth_l2_src_dst_tag() 1029 dr_ste_v1_build_eth_l2_src_dst_bit_mask(mask, sb->inner, sb->bit_mask); in dr_ste_v1_build_eth_l2_src_dst_init() 1031 sb->lu_type = DR_STE_CALC_DFNR_TYPE(ETHL2_SRC_DST, sb->inner); in dr_ste_v1_build_eth_l2_src_dst_init() 1040 struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer; in dr_ste_v1_build_eth_l3_ipv6_dst_tag() 1055 sb->lu_type = DR_STE_CALC_DFNR_TYPE(IPV6_DES, sb->inner); in dr_ste_v1_build_eth_l3_ipv6_dst_init() 1120 dr_ste_v1_build_eth_l2_src_or_dst_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask) dr_ste_v1_build_eth_l2_src_or_dst_bit_mask() argument 1170 dr_ste_v1_build_eth_l2_src_or_dst_tag(struct mlx5dr_match_param * value,bool inner,u8 * tag) dr_ste_v1_build_eth_l2_src_or_dst_tag() argument 1228 dr_ste_v1_build_eth_l2_src_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask) dr_ste_v1_build_eth_l2_src_bit_mask() argument 1261 dr_ste_v1_build_eth_l2_dst_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask) dr_ste_v1_build_eth_l2_dst_bit_mask() argument 1294 dr_ste_v1_build_eth_l2_tnl_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask) dr_ste_v1_build_eth_l2_tnl_bit_mask() argument [all...] |
H A D | dr_types.h | 215 u8 inner:1; member 394 bool inner, bool rx); 398 bool inner, bool rx); 402 bool inner, bool rx); 406 bool inner, bool rx); 410 bool inner, bool rx); 414 bool inner, bool rx); 418 bool inner, bool rx); 422 bool inner, bool rx); 426 bool inner, boo 814 struct mlx5dr_match_spec inner; global() member [all...] |
/linux/rust/kernel/sync/ |
H A D | arc.rs | 237 let inner = KBox::new(value, flags)?; in new() localVariable 238 let inner = KBox::leak(inner).into(); in new() localVariable 240 // SAFETY: We just created `inner` with a reference count of 1, which is owned by the new in new() 242 Ok(unsafe { Self::from_inner(inner) }) in new() 251 /// The caller must ensure that `inner` points to a valid location and has a non-zero reference 253 unsafe fn from_inner(inner: NonNull<ArcInner<T>>) -> Self { in from_inner() 256 ptr: inner, in from_inner() 368 inner: ManuallyDrop::into_inner(me), in into_unique_or_drop() 391 let inner in from_foreign() localVariable 402 let inner = unsafe { NonNull::new_unchecked(ptr.cast::<ArcInner<T>>()) }; borrow() localVariable 565 inner: NonNull<ArcInner<T>>, global() field 719 inner: Arc<T>, global() field 776 let inner = KBox::try_init::<AllocError>( new_uninit() localVariable 807 let inner = ManuallyDrop::new(self).inner.ptr; assume_init() localVariable [all...] |
H A D | poll.rs | 71 inner: CondVar, field 78 inner <- CondVar::new(name, key), in new() 88 &self.inner in deref() 99 unsafe { bindings::__wake_up_pollfree(self.inner.wait_queue_head.get()) }; in drop()
|
H A D | completion.rs | 69 inner: Opaque<bindings::completion>, field 82 inner <- Opaque::ffi_init(|slot: *mut bindings::completion| { in new() 90 self.inner.get() in as_raw()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | rss.c | 174 bool inner) in rss_get_tirp() argument 176 return inner ? &rss->inner_tir[tt] : &rss->tir[tt]; in rss_get_tirp() 180 bool inner) in rss_get_tir() argument 182 return *rss_get_tirp(rss, tt, inner); in rss_get_tir() 198 bool inner) in mlx5e_rss_create_tir() argument 207 if (inner && !rss->inner_ft_support) { in mlx5e_rss_create_tir() 209 "Cannot create inner indirect TIR[%d], RSS inner FT is not supported.\n", in mlx5e_rss_create_tir() 214 tir_p = rss_get_tirp(rss, tt, inner); in mlx5e_rss_create_tir() 233 mlx5e_tir_builder_build_rss(builder, &rss->hash, &rss_tt, inner); in mlx5e_rss_create_tir() 252 mlx5e_rss_destroy_tir(struct mlx5e_rss * rss,enum mlx5_traffic_types tt,bool inner) mlx5e_rss_destroy_tir() argument 269 mlx5e_rss_create_tirs(struct mlx5e_rss * rss,const struct mlx5e_packet_merge_param * init_pkt_merge_param,bool inner) mlx5e_rss_create_tirs() argument 289 mlx5e_rss_destroy_tirs(struct mlx5e_rss * rss,bool inner) mlx5e_rss_destroy_tirs() argument 298 mlx5e_rss_update_tir(struct mlx5e_rss * rss,enum mlx5_traffic_types tt,bool inner) mlx5e_rss_update_tir() argument 447 mlx5e_rss_get_tirn(struct mlx5e_rss * rss,enum mlx5_traffic_types tt,bool inner) mlx5e_rss_get_tirn() argument 463 mlx5e_rss_valid_tir(struct mlx5e_rss * rss,enum mlx5_traffic_types tt,bool inner) mlx5e_rss_valid_tir() argument 474 mlx5e_rss_obtain_tirn(struct mlx5e_rss * rss,enum mlx5_traffic_types tt,const struct mlx5e_packet_merge_param * init_pkt_merge_param,bool inner,u32 * tirn) mlx5e_rss_obtain_tirn() argument [all...] |
/linux/tools/testing/selftests/net/ |
H A D | l2_tos_ttl_inherit.sh | 67 local inner="$3" 102 "$type" "$outer" "$inner" "$tos" "$ttl" "$vlan" 182 # Assign inner IPv4/IPv6 addresses 183 if [ "$inner" = "4" ] || [ "$inner" = "other" ]; then 186 elif [ "$inner" = "6" ]; then 196 local inner="$2" 203 if [ "$inner" = "4" ]; then 205 elif [ "$inner" = "6" ]; then 207 elif [ "$inner" [all...] |
/linux/rust/kernel/sync/lock/ |
H A D | global.rs | 37 inner: Lock<B::Item, B::Backend>, field 49 inner: Lock { in new() 72 self.inner.state.get(), in init() 82 inner: self.inner.lock(), in lock() 89 inner: self.inner.try_lock()?, in try_lock() 98 inner: Guard<'static, B::Item, B::Backend>, field 105 &self.inner in deref() 111 &mut self.inner in deref_mut() [all...] |
/linux/rust/kernel/io/ |
H A D | resource.rs | 128 let inner = self.0.get(); in size() localVariable 130 unsafe { bindings::resource_size(inner) } in size() 135 let inner = self.0.get(); in start() localVariable 137 unsafe { (*inner).start } in start() 142 let inner = self.0.get(); in name() localVariable 145 let name = unsafe { (*inner).name }; in name() 160 let inner = self.0.get(); in flags() localVariable 162 let flags = unsafe { (*inner).flags }; in flags()
|
/linux/rust/kernel/time/hrtimer/ |
H A D | tbox.rs | 18 /// - `self.inner` comes from a `Box::into_raw` call. 24 pub(crate) inner: NonNull<T>, field 36 // SAFETY: As we obtained `self.inner` from a valid reference when we in cancel() 38 let timer_ptr = unsafe { <T as HasHrTimer<T>>::raw_get_timer(self.inner.as_ptr()) }; in cancel() 53 // SAFETY: By type invariant, `self.inner` came from a `Box::into_raw` in drop() 55 drop(unsafe { Box::<T, A>::from_raw(self.inner.as_ptr()) }) in drop() 78 let inner = in start() localVariable 85 unsafe { T::start(inner.as_ptr(), expires) }; in start() 87 // INVARIANT: `inner` came from `Box::into_raw` above. in start() 89 inner, in start() [all...] |
/linux/drivers/gpu/drm/nova/ |
H A D | uapi.rs | 8 ($name:ident <= $inner:ty) => { 10 pub struct $name(::kernel::types::Opaque<$inner>); 12 impl ::core::convert::From<&::kernel::types::Opaque<$inner>> for &$name { 13 fn from(value: &::kernel::types::Opaque<$inner>) -> Self { 14 // SAFETY: `Self` is a transparent wrapper of `$inner`.
|
/linux/rust/kernel/block/mq/ |
H A D | tag_set.rs | 25 /// - `inner` is initialized and valid. 30 inner: Opaque<bindings::blk_mq_tag_set>, field 64 inner <- tag_set.pin_chain(|tag_set| { in new() 76 self.inner.get() in raw_tag_set() 83 // SAFETY: By type invariant `inner` is valid and has been properly in drop() 85 unsafe { bindings::blk_mq_free_tag_set(self.inner.get()) }; in drop()
|
/linux/tools/testing/selftests/openat2/ |
H A D | openat2_test.c | 33 struct open_how inner; member 58 .arg.inner.flags = O_RDONLY, in test_openat2_struct() 62 .arg.inner.flags = O_RDONLY, in test_openat2_struct() 69 .arg.inner.flags = O_RDONLY, .size = 0, .err = -EINVAL }, in test_openat2_struct() 71 .arg.inner.flags = O_RDONLY, in test_openat2_struct() 76 .arg.inner.flags = O_RDONLY, .arg.extra1 = 0xdeadbeef, in test_openat2_struct() 79 .arg.inner.flags = O_RDONLY, .arg.extra2 = 0xfeedcafe, in test_openat2_struct() 82 .arg.inner.flags = O_RDONLY, .arg.extra3 = 0xabad1dea, in test_openat2_struct()
|
/linux/Documentation/bpf/ |
H A D | map_of_maps.rst | 14 an outer map contains instances of a single type of inner map, for example 17 When creating an outer map, an inner map instance is used to initialize the 18 metadata that the outer map holds about its inner maps. This inner map has a 28 - Any BPF map type can be used as an inner map, except for 56 helper returns a pointer to the inner map, or ``NULL`` if no entry was found. 111 This snippet shows how to add an inner map to an outer map:
|
/linux/samples/rust/ |
H A D | rust_misc_device.rs | 152 inner: Mutex<Inner>, field 168 inner <- new_mutex!( Inner{ value: 0_i32 } ), in open() 207 let mut guard = self.inner.lock(); in set_value() 220 let guard = self.inner.lock(); in get_value()
|
/linux/rust/kernel/device/ |
H A D | property.rs | 150 inner: res, in property_read_array_vec() 195 inner: T::read_from_fwnode_property(self, name), in property_read() 582 inner: Result<T>, field 596 if self.inner.is_err() { in required_by() 604 self.inner in required_by() 612 self.inner.ok() in optional() 620 self.inner.unwrap_or(default) in or() 629 self.inner.unwrap_or_default() in or_default()
|
/linux/rust/pin-init/examples/ |
H A D | mutex.rs | 29 inner: AtomicBool, field 36 .inner in acquire() 41 while self.inner.load(Ordering::Relaxed) { in acquire() 52 inner: AtomicBool::new(false), in new() 62 self.0.inner.store(false, Ordering::Release); in drop()
|