| /linux/rust/syn/gen/ |
| H A D | clone.rs | 10 fn clone(&self) -> Self { in clone() method 12 extern_token: self.extern_token.clone(), in clone() 13 name: self.name.clone(), in clone() 20 fn clone(&self) -> Self { in clone() method 22 colon2_token: self.colon2_token.clone(), in clone() 23 lt_token: self.lt_token.clone(), in clone() 24 args: self.args.clone(), in clone() 25 gt_token: self.gt_token.clone(), in clone() 32 fn clone(&self) -> Self { in clone() method 34 attrs: self.attrs.clone(), in clone() [all …]
|
| /linux/drivers/md/ |
| H A D | dm-clone-target.c | 69 struct clone { struct 152 static enum clone_metadata_mode get_clone_mode(struct clone *clone) in get_clone_mode() argument 154 return READ_ONCE(clone->mode); in get_clone_mode() 157 static const char *clone_device_name(struct clone *clone) in clone_device_name() argument 159 return dm_table_device_name(clone->ti->table); in clone_device_name() 162 static void __set_clone_mode(struct clone *clone, enum clone_metadata_mode new_mode) in __set_clone_mode() argument 170 enum clone_metadata_mode old_mode = get_clone_mode(clone); in __set_clone_mode() 179 dm_clone_metadata_set_read_only(clone->cmd); in __set_clone_mode() 183 dm_clone_metadata_set_read_write(clone->cmd); in __set_clone_mode() 187 WRITE_ONCE(clone->mode, new_mode); in __set_clone_mode() [all …]
|
| H A D | dm-rq.c | 21 struct request *orig, *clone; member 78 static void end_clone_bio(struct bio *clone) in end_clone_bio() argument 81 container_of(clone, struct dm_rq_clone_bio_info, clone); in end_clone_bio() 84 blk_status_t error = clone->bi_status; in end_clone_bio() 85 bool is_last = !clone->bi_next; in end_clone_bio() 87 bio_put(clone); in end_clone_bio() 119 tio->clone->bio = NULL; in end_clone_bio() 164 static void dm_end_request(struct request *clone, blk_status_t error) in dm_end_request() argument 166 struct dm_rq_target_io *tio = clone->end_io_data; in dm_end_request() 170 blk_rq_unprep_clone(clone); in dm_end_request() [all …]
|
| H A D | dm-flakey.c | 407 static void clone_free(struct bio *clone) in clone_free() argument 411 if (clone->bi_vcnt > 0) { /* bio_for_each_folio_all crashes with an empty bio */ in clone_free() 412 bio_for_each_folio_all(fi, clone) in clone_free() 416 bio_uninit(clone); in clone_free() 417 kfree(clone); in clone_free() 420 static void clone_endio(struct bio *clone) in clone_endio() argument 422 struct bio *bio = clone->bi_private; in clone_endio() 423 bio->bi_status = clone->bi_status; in clone_endio() 424 clone_free(clone); in clone_endio() 430 struct bio *clone; in clone_bio() local [all …]
|
| H A D | dm.c | 93 static inline struct dm_target_io *clone_to_tio(struct bio *clone) in clone_to_tio() argument 95 return container_of(clone, struct dm_target_io, clone); in clone_to_tio() 119 return container_of(bio, struct dm_target_io, clone)->target_bio_nr; in dm_bio_get_target_bio_nr() 540 static void dm_start_io_acct(struct dm_io *io, struct bio *clone) in dm_start_io_acct() argument 549 if (!clone || likely(dm_tio_is_normal(clone_to_tio(clone)))) { in dm_start_io_acct() 575 struct bio *clone; in alloc_io() local 577 clone = bio_alloc_clone(NULL, bio, gfp_mask, &md->mempools->io_bs); in alloc_io() 578 if (unlikely(!clone)) in alloc_io() 580 tio = clone_to_tio(clone); in alloc_io() 610 bio_put(&io->tio.clone); in free_io() [all …]
|
| H A D | dm-crypt.c | 285 static void crypt_endio(struct bio *clone); 1643 static void crypt_free_buffer_pages(struct crypt_config *cc, struct bio *clone); 1668 struct bio *clone; in crypt_alloc_buffer() local 1678 clone = bio_alloc_bioset(cc->dev->bdev, nr_iovecs, io->base_bio->bi_opf, in crypt_alloc_buffer() 1680 clone->bi_private = io; in crypt_alloc_buffer() 1681 clone->bi_end_io = crypt_endio; in crypt_alloc_buffer() 1682 clone->bi_ioprio = io->base_bio->bi_ioprio; in crypt_alloc_buffer() 1683 clone->bi_iter.bi_sector = cc->start + io->sector; in crypt_alloc_buffer() 1710 crypt_free_buffer_pages(cc, clone); in crypt_alloc_buffer() 1711 bio_put(clone); in crypt_alloc_buffer() [all …]
|
| H A D | dm-zoned-target.c | 103 static void dmz_clone_endio(struct bio *clone) in dmz_clone_endio() argument 105 struct dmz_bioctx *bioctx = clone->bi_private; in dmz_clone_endio() 106 blk_status_t status = clone->bi_status; in dmz_clone_endio() 108 bio_put(clone); in dmz_clone_endio() 123 struct bio *clone; in dmz_submit_bio() local 128 clone = bio_alloc_clone(dev->bdev, bio, GFP_NOIO, &dmz->bio_set); in dmz_submit_bio() 129 if (!clone) in dmz_submit_bio() 133 clone->bi_iter.bi_sector = in dmz_submit_bio() 135 clone->bi_iter.bi_size = dmz_blk2sect(nr_blocks) << SECTOR_SHIFT; in dmz_submit_bio() 136 clone->bi_end_io = dmz_clone_endio; in dmz_submit_bio() [all …]
|
| /linux/rust/pin-init/internal/src/ |
| H A D | helpers.rs | |
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | dm-clone.rst | 4 dm-clone 10 dm-clone is a device mapper target which produces a one-to-one copy of an 15 The main use case of dm-clone is to clone a potentially remote, high-latency, 26 When the cloning completes, the dm-clone table can be removed altogether and be 29 The dm-clone target reuses the metadata library used by the thin-provisioning 56 clone of the source device. 68 dm-clone divides the source and destination devices in fixed sized regions. 72 The region size is configurable when you first create the dm-clone device. The 92 dm-clone interprets a discard request to a range that hasn't been hydrated yet 97 If the destination device supports discards, then by default dm-clone will pass [all …]
|
| /linux/drivers/net/ethernet/meta/fbnic/ |
| H A D | fbnic_ethtool.c | 156 struct fbnic_net *clone; in fbnic_clone_create() local 158 clone = kmemdup(orig, sizeof(*orig), GFP_KERNEL); in fbnic_clone_create() 159 if (!clone) in fbnic_clone_create() 162 memset(clone->tx, 0, sizeof(clone->tx)); in fbnic_clone_create() 163 memset(clone->rx, 0, sizeof(clone->rx)); in fbnic_clone_create() 164 memset(clone->napi, 0, sizeof(clone->napi)); in fbnic_clone_create() 165 return clone; in fbnic_clone_create() 169 struct fbnic_net *clone) in fbnic_clone_swap_cfg() argument 171 swap(clone->rcq_size, orig->rcq_size); in fbnic_clone_swap_cfg() 172 swap(clone->hpq_size, orig->hpq_size); in fbnic_clone_swap_cfg() [all …]
|
| /linux/fs/ |
| H A D | posix_acl.c | 220 struct posix_acl *clone = NULL; in posix_acl_clone() local 223 clone = kmemdup(acl, struct_size(acl, a_entries, acl->a_count), in posix_acl_clone() 225 if (clone) in posix_acl_clone() 226 refcount_set(&clone->a_refcount, 1); in posix_acl_clone() 228 return clone; in posix_acl_clone() 557 struct posix_acl *clone = posix_acl_clone(*acl, gfp); in __posix_acl_create() local 559 if (clone) { in __posix_acl_create() 560 err = posix_acl_create_masq(clone, mode_p); in __posix_acl_create() 562 posix_acl_release(clone); in __posix_acl_create() 563 clone = NULL; in __posix_acl_create() [all …]
|
| /linux/fs/f2fs/ |
| H A D | acl.c | 296 struct posix_acl *clone = NULL; in f2fs_acl_clone() local 299 clone = kmemdup(acl, struct_size(acl, a_entries, acl->a_count), in f2fs_acl_clone() 301 if (clone) in f2fs_acl_clone() 302 refcount_set(&clone->a_refcount, 1); in f2fs_acl_clone() 304 return clone; in f2fs_acl_clone() 366 struct posix_acl *clone; in f2fs_acl_create() local 383 clone = f2fs_acl_clone(p, GFP_NOFS); in f2fs_acl_create() 384 if (!clone) { in f2fs_acl_create() 389 ret = f2fs_acl_create_masq(clone, mode); in f2fs_acl_create() 394 posix_acl_release(clone); in f2fs_acl_create() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | dynptr_fail.c | 1715 struct bpf_dynptr clone; in clone_invalid2() local 1719 bpf_ringbuf_reserve_dynptr(&ringbuf, 64, 0, &clone); in clone_invalid2() 1722 bpf_dynptr_clone(&ptr1, &clone); in clone_invalid2() 1724 bpf_ringbuf_submit_dynptr(&clone, 0); in clone_invalid2() 1734 struct bpf_dynptr clone; in clone_invalidate1() local 1740 bpf_dynptr_clone(&ptr, &clone); in clone_invalidate1() 1745 bpf_dynptr_read(read_data, sizeof(read_data), &clone, 0, 0); in clone_invalidate1() 1756 struct bpf_dynptr clone; in clone_invalidate2() local 1761 bpf_dynptr_clone(&ptr, &clone); in clone_invalidate2() 1763 bpf_ringbuf_submit_dynptr(&clone, in clone_invalidate2() 1803 struct bpf_dynptr clone; clone_invalidate4() local 1829 struct bpf_dynptr clone; clone_invalidate5() local 1883 struct bpf_dynptr clone; clone_skb_packet_data() local 1908 struct bpf_dynptr clone, meta; clone_skb_packet_meta() local 1932 struct bpf_dynptr clone; clone_xdp_packet_data() local [all...] |
| /linux/rust/syn/ |
| H A D | punctuated.rs | 392 fn clone(&self) -> Self { in clone() function 394 inner: self.inner.clone(), in clone() 395 last: self.last.clone(), in clone() 599 fn clone(&self) -> Self { in clone() method 601 inner: self.inner.clone(), in clone() 602 last: self.last.clone(), in clone() 692 fn clone(&self) -> Self { in clone() function 694 inner: self.inner.clone(), in clone() 695 last: self.last.clone(), in clone() 737 fn clone(&self) -> Self { in clone() function [all …]
|
| H A D | buffer.rs | 194 Entry::Ident(ident) => Some((ident.clone(), unsafe { self.bump_ignore_group() })), in ident() 205 Some((punct.clone(), unsafe { self.bump_ignore_group() })) in punct() 216 Entry::Literal(literal) => Some((literal.clone(), unsafe { self.bump_ignore_group() })), in literal() 281 return Some((group.clone(), after_group)); in any_group_token() 308 Entry::Group(group, end_offset) => (group.clone().into(), *end_offset), in token_tree() 309 Entry::Literal(literal) => (literal.clone().into(), 1), in token_tree() 310 Entry::Ident(ident) => (ident.clone().into(), 1), in token_tree() 311 Entry::Punct(punct) => (punct.clone().into(), 1), in token_tree() 387 fn clone(&self) -> Self { in clone() method
|
| H A D | error.rs | 381 fn clone(&self) -> Self { in clone() method 383 messages: self.messages.clone(), in clone() 389 fn clone(&self) -> Self { in clone() method 392 message: self.message.clone(), in clone() 398 fn clone(&self) -> Self { in clone() method 458 messages: vec![self.messages.next()?.clone()], in next()
|
| /linux/drivers/android/binder/ |
| H A D | transaction.rs | 75 let to = node_ref.node.owner.clone(); in new() 77 to.clone(), in new() 97 alloc.set_info_oneway_node(node_ref.node.clone()); in new() 102 let target_node = node_ref.node.clone(); in new() 111 from: from.clone(), in new() 134 let mut alloc = match from.copy_transaction_data(to.clone(), tr, debug_id, allow_fds, None) in new_reply() 151 from: from.clone(), in new_reply() 197 Some(self.from_parent.as_ref()?.clone()) in clone_next() 207 return Some(transaction.from.clone()); in find_target_thread() 256 let process = self.to.clone(); in submit() [all …]
|
| H A D | context.rs | 32 ctxs.push(ctx.clone(), GFP_KERNEL)?; in get_all_contexts() 68 CONTEXTS.lock().contexts.push(ctx.clone(), GFP_KERNEL)?; in new() 131 .clone(strong) in get_manager_node() 149 procs.push(Arc::clone(proc), GFP_KERNEL)?; in get_all_procs() 159 matching_procs.push(Arc::clone(proc), GFP_KERNEL)?; in get_procs_with_pid()
|
| /linux/rust/quote/ |
| H A D | to_tokens.rs | 237 tokens.append(self.clone()); in to_tokens() 243 tokens.append(self.clone()); in to_tokens() 249 tokens.append(self.clone()); in to_tokens() 255 tokens.append(self.clone()); in to_tokens() 261 tokens.append(self.clone()); in to_tokens() 267 tokens.extend(iter::once(self.clone())); in to_tokens()
|
| /linux/net/ipv4/ |
| H A D | inet_fragment.c | 562 struct sk_buff *clone; in inet_frag_reasm_prepare() local 565 clone = alloc_skb(0, GFP_ATOMIC); in inet_frag_reasm_prepare() 566 if (!clone) in inet_frag_reasm_prepare() 568 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in inet_frag_reasm_prepare() 572 clone->data_len = head->data_len - plen; in inet_frag_reasm_prepare() 573 clone->len = clone->data_len; in inet_frag_reasm_prepare() 574 head->truesize += clone->truesize; in inet_frag_reasm_prepare() 575 clone->csum = 0; in inet_frag_reasm_prepare() 576 clone->ip_summed = head->ip_summed; in inet_frag_reasm_prepare() 577 add_frag_mem_limit(q->fqdir, clone->truesize); in inet_frag_reasm_prepare() [all …]
|
| /linux/net/core/ |
| H A D | timestamping.c | 28 struct sk_buff *clone; in skb_clone_tx_timestamp() local 59 clone = skb_clone_sk(skb); in skb_clone_tx_timestamp() 60 if (!clone) in skb_clone_tx_timestamp() 62 mii_ts->txtstamp(mii_ts, clone, type); in skb_clone_tx_timestamp()
|
| /linux/include/linux/dsa/ |
| H A D | ocelot.h | 16 struct sk_buff *clone; member 265 struct sk_buff *clone = OCELOT_SKB_CB(skb)->clone; in ocelot_ptp_rew_op() local 269 if (ptp_cmd == IFH_REW_OP_TWO_STEP_PTP && clone) { in ocelot_ptp_rew_op() 271 rew_op |= OCELOT_SKB_CB(clone)->ts_id << 3; in ocelot_ptp_rew_op()
|
| /linux/net/rds/ |
| H A D | tcp_recv.c | 162 struct sk_buff *clone; in rds_tcp_data_recv() local 219 clone = pskb_extract(skb, offset, to_copy, arg->gfp); in rds_tcp_data_recv() 220 if (!clone) { in rds_tcp_data_recv() 225 skb_queue_tail(&tinc->ti_skb_list, clone); in rds_tcp_data_recv() 230 clone, clone->data, clone->len); in rds_tcp_data_recv()
|
| /linux/fs/btrfs/ |
| H A D | fiemap.c | 308 struct extent_buffer *clone = path->nodes[0]; in fiemap_next_leaf_item() local 322 ASSERT(test_bit(EXTENT_BUFFER_UNMAPPED, &clone->bflags)); in fiemap_next_leaf_item() 323 refcount_inc(&clone->refs); in fiemap_next_leaf_item() 349 clone->start = path->nodes[0]->start; in fiemap_next_leaf_item() 351 copy_extent_buffer_full(clone, path->nodes[0]); in fiemap_next_leaf_item() 355 path->nodes[0] = clone; in fiemap_next_leaf_item() 359 free_extent_buffer(clone); in fiemap_next_leaf_item() 374 struct extent_buffer *clone; in fiemap_search_slot() local 419 clone = btrfs_clone_extent_buffer(path->nodes[0]); in fiemap_search_slot() 420 if (!clone) in fiemap_search_slot() [all …]
|
| /linux/net/netfilter/ |
| H A D | nft_set_pipapo.c | 1213 static int pipapo_realloc_scratch(struct nft_pipapo_match *clone, argument 1235 pipapo_free_scratch(clone, i); 1237 *per_cpu_ptr(clone->scratch, i) = scratch; 1267 if (priv->clone) 1268 return priv->clone; 1272 priv->clone = pipapo_clone(m); 1274 return priv->clone; 1837 if (!priv->clone) 1841 pipapo_gc_scan(set, priv->clone); 1843 old = rcu_replace_pointer(priv->match, priv->clone, [all …]
|