/linux/fs/btrfs/ |
H A D | ordered-data.c | 124 * look find the first ordered struct that has this offset, otherwise 162 * ordered extent by calling btrfs_qgroup_release_data() and tracking in alloc_ordered_extent() 163 * the qgroup reserved amount in the ordered extent, so that later after in alloc_ordered_extent() 164 * completing the ordered extent, when running the data delayed ref it in alloc_ordered_extent() 245 "inconsistency in ordered tree at offset %llu", in insert_ordered_extent() 263 * Add an ordered extent to the per-inode tree. 276 * tree is given a single reference on the ordered extent that was inserted, and 279 * Return: the new ordered extent or error pointer. 295 * For PREALLOC, we do not use ordered extent members, but in btrfs_alloc_ordered_extent() 298 * So here we always pass 0 as offset for NOCOW/PREALLOC ordered extent in btrfs_alloc_ordered_extent() 336 btrfs_mark_ordered_extent_error(struct btrfs_ordered_extent * ordered) btrfs_mark_ordered_extent_error() argument 350 can_finish_ordered_extent(struct btrfs_ordered_extent * ordered,struct folio * folio,u64 file_offset,u64 len,bool uptodate) can_finish_ordered_extent() argument 404 btrfs_queue_ordered_fn(struct btrfs_ordered_extent * ordered) btrfs_queue_ordered_fn() argument 415 btrfs_finish_ordered_extent(struct btrfs_ordered_extent * ordered,struct folio * folio,u64 file_offset,u64 len,bool uptodate) btrfs_finish_ordered_extent() argument 740 struct btrfs_ordered_extent *ordered; btrfs_run_ordered_extent_work() local 758 struct btrfs_ordered_extent *ordered, *next; btrfs_wait_ordered_extents() local 909 struct btrfs_ordered_extent *ordered; btrfs_wait_ordered_range() local 1046 struct btrfs_ordered_extent *ordered; btrfs_get_ordered_extents_for_logging() local 1179 struct btrfs_ordered_extent *ordered; btrfs_lock_and_flush_ordered_range() local 1216 struct btrfs_ordered_extent *ordered; btrfs_try_lock_ordered_range() local 1233 btrfs_split_ordered_extent(struct btrfs_ordered_extent * ordered,u64 len) btrfs_split_ordered_extent() argument [all...] |
H A D | direct-io.c | 17 struct btrfs_ordered_extent *ordered; member 40 struct btrfs_ordered_extent *ordered; in lock_extent_direct() local 63 * doing DIO to, so we need to make sure there's no ordered in lock_extent_direct() 66 ordered = btrfs_lookup_ordered_range(BTRFS_I(inode), lockstart, in lock_extent_direct() 76 if (!ordered && in lock_extent_direct() 83 if (ordered) { in lock_extent_direct() 85 btrfs_put_ordered_extent(ordered); in lock_extent_direct() 90 * If we are doing a DIO read and the ordered extent we in lock_extent_direct() 96 * created an ordered extent for a previous extent map in lock_extent_direct() 105 test_bit(BTRFS_ORDERED_DIRECT, &ordered in lock_extent_direct() 145 struct btrfs_ordered_extent *ordered; btrfs_create_dio_extent() local 671 btrfs_extract_ordered_extent(struct btrfs_bio * bbio,struct btrfs_ordered_extent * ordered) btrfs_extract_ordered_extent() argument [all...] |
H A D | ordered-data.h | 51 * Different types for ordered extents, one and only one of the 4 types 52 * need to be set when creating ordered extent. 70 /* Extra status bits for ordered extents */ 82 /* We have already logged all the csums of the ordered extent */ 117 * this ordered extent so that we do not expose stale data. 148 /* a per root list of all the pending ordered extents */ 166 void btrfs_finish_ordered_extent(struct btrfs_ordered_extent *ordered, 223 struct btrfs_ordered_extent *ordered, u64 len); 224 void btrfs_mark_ordered_extent_error(struct btrfs_ordered_extent *ordered);
|
H A D | inode.c | 42 #include "ordered-data.h" 386 * Cleanup all submitted ordered extents in specified range to handle errors 392 * to be released, which we want to happen only when finishing the ordered 413 * the ordered extent accounting for the range. in btrfs_cleanup_ordered_extents() 837 * This is done inside an ordered work queue, and the compression is spread 838 * across many cpus. The actual IO submission is step two, and the ordered work 1092 struct btrfs_ordered_extent *ordered; in submit_one_async_extent() local 1160 ordered = btrfs_alloc_ordered_extent(inode, start, &file_extent, in submit_one_async_extent() 1162 if (IS_ERR(ordered)) { in submit_one_async_extent() 1164 ret = PTR_ERR(ordered); in submit_one_async_extent() 1343 struct btrfs_ordered_extent *ordered; cow_file_range() local 1968 struct btrfs_ordered_extent *ordered; nocow_one_range() local 2737 struct btrfs_ordered_extent *ordered; btrfs_writepage_fixup_worker() local 3321 btrfs_finish_ordered_io(struct btrfs_ordered_extent * ordered) btrfs_finish_ordered_io() argument 4848 struct btrfs_ordered_extent *ordered; btrfs_truncate_block() local 7480 struct btrfs_ordered_extent *ordered; btrfs_invalidate_folio() local 7891 struct btrfs_ordered_extent *ordered; btrfs_destroy_inode() local 9517 struct btrfs_ordered_extent *ordered; btrfs_encoded_read() local 9540 struct btrfs_ordered_extent *ordered; btrfs_encoded_read() local 9660 struct btrfs_ordered_extent *ordered; btrfs_do_encoded_write() local 9781 struct btrfs_ordered_extent *ordered; btrfs_do_encoded_write() local 10404 struct btrfs_ordered_extent *ordered; btrfs_assert_inode_range_clean() local [all...] |
H A D | compression.c | 30 #include "ordered-data.h" 319 btrfs_finish_ordered_extent(cb->bbio.ordered, NULL, cb->start, cb->len, in btrfs_finish_compressed_write_work() 371 void btrfs_submit_compressed_write(struct btrfs_ordered_extent *ordered, in btrfs_submit_compressed_write() argument 377 struct btrfs_inode *inode = ordered->inode; in btrfs_submit_compressed_write() 381 ASSERT(IS_ALIGNED(ordered->file_offset, fs_info->sectorsize)); in btrfs_submit_compressed_write() 382 ASSERT(IS_ALIGNED(ordered->num_bytes, fs_info->sectorsize)); in btrfs_submit_compressed_write() 384 cb = alloc_compressed_bio(inode, ordered->file_offset, in btrfs_submit_compressed_write() 387 cb->start = ordered->file_offset; in btrfs_submit_compressed_write() 388 cb->len = ordered->num_bytes; in btrfs_submit_compressed_write() 390 cb->compressed_len = ordered in btrfs_submit_compressed_write() [all...] |
H A D | zoned.c | 1849 static void btrfs_rewrite_logical_zoned(struct btrfs_ordered_extent *ordered, in btrfs_rewrite_logical_zoned() argument 1852 struct extent_map_tree *em_tree = &ordered->inode->extent_tree; in btrfs_rewrite_logical_zoned() 1855 ordered->disk_bytenr = logical; in btrfs_rewrite_logical_zoned() 1858 em = btrfs_search_extent_mapping(em_tree, ordered->file_offset, in btrfs_rewrite_logical_zoned() 1859 ordered->num_bytes); in btrfs_rewrite_logical_zoned() 1867 static bool btrfs_zoned_split_ordered(struct btrfs_ordered_extent *ordered, in btrfs_zoned_split_ordered() argument 1872 if (!test_bit(BTRFS_ORDERED_NOCOW, &ordered->flags) && in btrfs_zoned_split_ordered() 1873 btrfs_split_extent_map(ordered->inode, ordered->file_offset, in btrfs_zoned_split_ordered() 1874 ordered in btrfs_zoned_split_ordered() 1885 btrfs_finish_ordered_zoned(struct btrfs_ordered_extent * ordered) btrfs_finish_ordered_zoned() argument [all...] |
H A D | extent_io.c | 95 * how many bytes are there before stripe/ordered extent boundary. 469 btrfs_finish_ordered_extent(bbio->ordered, folio, start, len, in end_bbio_data_write() 675 /* Limit data write bios to the ordered boundary. */ in alloc_new_bio() 677 struct btrfs_ordered_extent *ordered; in alloc_new_bio() local 679 ordered = btrfs_lookup_ordered_extent(inode, file_offset); in alloc_new_bio() 680 if (ordered) { in alloc_new_bio() 682 ordered->file_offset + in alloc_new_bio() 683 ordered->disk_num_bytes - file_offset); in alloc_new_bio() 684 bbio->ordered = ordered; in alloc_new_bio() 1072 can_skip_one_ordered_range(struct btrfs_inode * inode,struct btrfs_ordered_extent * ordered,u64 * fileoff) can_skip_one_ordered_range() argument 1168 can_skip_ordered_extent(struct btrfs_inode * inode,struct btrfs_ordered_extent * ordered,u64 start,u64 end) can_skip_ordered_extent() argument 1207 struct btrfs_ordered_extent *ordered; lock_extents_for_read() local [all...] |
H A D | bio.c | 93 refcount_inc(&orig_bbio->ordered->refs); in btrfs_split_bio() 94 bbio->ordered = orig_bbio->ordered; in btrfs_split_bio() 108 btrfs_put_ordered_extent(bbio->ordered); in btrfs_bio_end_io() 127 struct btrfs_ordered_extent *ordered = bbio->ordered; in btrfs_bio_end_io() local 130 btrfs_put_ordered_extent(ordered); in btrfs_bio_end_io() 531 * sums are attached onto the ordered extent record. 533 * At IO completion time the csums attached on the ordered extent record are 550 * sums are attached onto the ordered exten [all...] |
/linux/Documentation/devicetree/bindings/scsi/ |
H A D | hisilicon-sas.txt | 22 sources; the interrupts are ordered in 3 groups, as follows: 30 The phy interrupts are ordered into groups of 3 per phy 34 The interrupts are ordered in increasing order. 35 Fatal interrupts : the fatal interrupts are ordered as follows: 39 the interrupts are ordered in 3 groups, as follows: 47 interrupt. The interrupts are ordered in increasing 50 interrupt source. The interrupts are ordered in
|
/linux/include/trace/events/ |
H A D | btrfs.h | 501 const struct btrfs_ordered_extent *ordered), 503 TP_ARGS(inode, ordered), 521 __entry->file_offset = ordered->file_offset; 522 __entry->start = ordered->disk_bytenr; 523 __entry->len = ordered->num_bytes; 524 __entry->disk_len = ordered->disk_num_bytes; 525 __entry->bytes_left = ordered->bytes_left; 526 __entry->flags = ordered->flags; 527 __entry->compress_type = ordered->compress_type; 528 __entry->refs = refcount_read(&ordered [all...] |
/linux/tools/lib/subcmd/ |
H A D | parse-options.c | 811 struct option *opt, *ordered = NULL, *group; in options__order() local 823 group = realloc(ordered, len); in options__order() 826 ordered = group; in options__order() 827 memcpy(&ordered[nr_parent], p, sizeof(*o) * (nr_opts - nr_parent)); in options__order() 832 memcpy(&ordered[nr_opts], o, sizeof(*o)); in options__order() 835 for (opt = group = ordered; opt->type != OPTION_END; opt++) { in options__order() 847 return ordered; in options__order() 885 struct option *ordered; in usage_with_options_internal() local 910 ordered = options__order(opts); in usage_with_options_internal() 911 if (ordered) in usage_with_options_internal() [all...] |
/linux/Documentation/devicetree/bindings/sound/ |
H A D | sirf-audio-port.txt | 6 - dmas: List of DMA controller phandle and DMA request line ordered pairs. 8 These strings correspond 1:1 with the ordered pairs in dmas.
|
H A D | brcm,bcm2835-i2s.txt | 7 - dmas: List of DMA controller phandle and DMA request line ordered pairs. 9 These strings correspond 1:1 with the ordered pairs in dmas.
|
/linux/Documentation/core-api/ |
H A D | refcount-vs-atomic.rst | 67 then further stores are ordered against this operation. 135 * fully ordered --> control dependency on success for stores 151 * fully ordered --> ACQUIRE ordering on success 164 * fully ordered --> RELEASE ordering + ACQUIRE ordering on success 177 * fully ordered --> RELEASE ordering + control dependency 192 * fully ordered --> RELEASE ordering + control dependency + hold
|
/linux/Documentation/devicetree/bindings/ |
H A D | dts-coding-style.rst | 51 ordered by unit address in ascending order. 56 2. Nodes without unit addresses shall be ordered alpha-numerically by the node 57 name. For a few node types, they can be ordered by the main property, e.g. 58 pin configuration states ordered by value of "pins" property. 61 ordered either alpha-numerically or by keeping the order from DTSI, where
|
/linux/Documentation/ |
H A D | atomic_t.txt | 156 atomic variable) can be fully ordered and no intermediate state is lost or 169 - RMW operations that have a return value are fully ordered; 183 Fully ordered primitives are ordered against everything prior and everything 184 subsequent. Therefore a fully ordered primitive is like having an smp_mb() 198 ordered, so it is advisable to place the barrier right next to the RMW atomic 203 provide full ordered atomics and these barriers are no-ops. 205 NOTE: when the atomic RmW ops are fully ordered, they should also imply a
|
H A D | atomic_bitops.txt | 59 - RMW operations that have a return value are fully ordered. 61 - RMW operations that are conditional are fully ordered.
|
/linux/Documentation/litmus-tests/ |
H A D | README | 18 the RMW are ordered before the subsequential memory accesses. 24 cmpxchg-fail-ordered-1.litmus 28 cmpxchg-fail-ordered-2.litmus
|
/linux/virt/kvm/ |
H A D | Kconfig | 22 # Only strongly ordered architectures can select this, as it doesn't 30 # Weakly ordered architectures can only select this, advertising
|
/linux/Documentation/admin-guide/perf/ |
H A D | nvidia-pmu.rst | 32 strongly-ordered (SO) PCIE write traffic to local/remote memory. Please see 62 In this config, the PMU captures read and relaxed ordered (RO) writes from 155 to local memory. For PCIE traffic, this PMU captures read and relaxed ordered 281 PCIE1 traffic represents strongly ordered (SO) writes. 282 PCIE2 traffic represents reads and relaxed ordered (RO) writes. 332 PCIE1 traffic represents strongly ordered (SO) writes. 333 PCIE2 traffic represents reads and relaxed ordered (RO) writes.
|
/linux/tools/memory-model/litmus-tests/ |
H A D | CoWW+poonceonce.litmus | 7 * writes to the same variable are ordered.
|
H A D | CoRR+poonceonce+Once.litmus | 7 * reads from the same variable are ordered.
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-driver-input-cros-ec-keyb | 5 ordered by the physical positions of the keys, from left
|
/linux/tools/perf/Documentation/ |
H A D | perf.txt | 60 ordered-events - ordered events object debug messages
|
/linux/Documentation/litmus-tests/atomic/ |
H A D | cmpxchg-fail-ordered-2.litmus | 1 C cmpxchg-fail-ordered-2
|