/linux-6.8/drivers/media/dvb-core/ |
D | dvb_ringbuffer.c | 136 size_t split; in dvb_ringbuffer_read_user() local 138 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user() 139 if (split > 0) { in dvb_ringbuffer_read_user() 140 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user() 142 buf += split; in dvb_ringbuffer_read_user() 143 todo -= split; in dvb_ringbuffer_read_user() 162 size_t split; in dvb_ringbuffer_read() local 164 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read() 165 if (split > 0) { in dvb_ringbuffer_read() 166 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read() [all …]
|
/linux-6.8/lib/ |
D | sg_split.c | 81 struct sg_splitter *split; in sg_split_phys() local 83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys() 84 in_sg = split->in_sg0; in sg_split_phys() 85 out_sg = split->out_sg; in sg_split_phys() 86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys() 89 out_sg->offset += split->skip_sg0; in sg_split_phys() 90 out_sg->length -= split->skip_sg0; in sg_split_phys() 98 out_sg[-1].length = split->length_last_sg; in sg_split_phys() 107 struct sg_splitter *split; in sg_split_mapped() local 109 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_mapped() [all …]
|
D | maple_tree.c | 1805 static inline bool mab_middle_node(struct maple_big_node *b_node, int split, in mab_middle_node() argument 1813 if (!b_node->slot[split] && (size >= 2 * slot_count - 1)) in mab_middle_node() 1828 unsigned char split, unsigned char slot_count) in mab_no_null_split() argument 1830 if (!b_node->slot[split]) { in mab_no_null_split() 1835 if ((split < slot_count - 1) && in mab_no_null_split() 1836 (b_node->b_end - split) > (mt_min_slots[b_node->type])) in mab_no_null_split() 1837 split++; in mab_no_null_split() 1839 split--; in mab_no_null_split() 1841 return split; in mab_no_null_split() 1856 int split = b_end / 2; /* Assume equal split. */ in mab_calc_split() local [all …]
|
/linux-6.8/fs/btrfs/ |
D | extent_map.c | 680 struct extent_map *split; in btrfs_drop_extent_map_range() local 706 split = alloc_extent_map(); in btrfs_drop_extent_map_range() 756 if (!split) { in btrfs_drop_extent_map_range() 757 split = split2; in btrfs_drop_extent_map_range() 759 if (!split) in btrfs_drop_extent_map_range() 762 split->start = em->start; in btrfs_drop_extent_map_range() 763 split->len = start - em->start; in btrfs_drop_extent_map_range() 766 split->orig_start = em->orig_start; in btrfs_drop_extent_map_range() 767 split->block_start = em->block_start; in btrfs_drop_extent_map_range() 770 split->block_len = em->block_len; in btrfs_drop_extent_map_range() [all …]
|
/linux-6.8/tools/verification/dot2/ |
D | automata.py | 54 line = dot_lines[cursor].split() 64 while self.__dot_lines[cursor].split()[0] != "{node": 70 while self.__dot_lines[cursor].split()[0] != "{node": 72 while self.__dot_lines[cursor].split()[0] == "{node": 87 while self.__dot_lines[cursor].split()[0] == "{node": 88 line = self.__dot_lines[cursor].split() 127 if self.__dot_lines[cursor].split()[1] == "->": 128 line = self.__dot_lines[cursor].split() 136 for i in event.split(): 165 if self.__dot_lines[cursor].split()[1] == "->": [all …]
|
/linux-6.8/fs/jfs/ |
D | jfs_xtree.c | 96 uint split; member 109 struct xtsplit * split, struct btstack * btstack); 111 static int xtSplitPage(tid_t tid, struct inode *ip, struct xtsplit * split, 115 struct xtsplit * split, struct metapage ** rmpp); 524 struct xtsplit split; /* split information */ in xtInsert() local 586 split.mp = mp; in xtInsert() 587 split.index = index; in xtInsert() 588 split.flag = xflag; in xtInsert() 589 split.off = xoff; in xtInsert() 590 split.len = xlen; in xtInsert() [all …]
|
D | jfs_dtree.c | 140 struct dtsplit * split, struct btstack * btstack); 142 static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split, 146 struct dtsplit * split, struct btstack * btstack); 149 struct dtsplit * split, struct metapage ** rmpp); 822 struct dtsplit split; /* split information */ in dtInsert() local 863 split.mp = mp; in dtInsert() 864 split.index = index; in dtInsert() 865 split.nslot = n; in dtInsert() 866 split.key = name; in dtInsert() 867 split.data = &data; in dtInsert() [all …]
|
/linux-6.8/security/apparmor/ |
D | lib.c | 68 char *split = strchr(&name[1], ':'); in aa_split_fqname() local 70 if (split) { in aa_split_fqname() 72 *split++ = 0; in aa_split_fqname() 73 if (strncmp(split, "//", 2) == 0) in aa_split_fqname() 74 split += 2; in aa_split_fqname() 75 name = skip_spaces(split); in aa_split_fqname() 117 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local 121 if (split) { in aa_splitn_fqname() 122 *ns_len = split - *ns_name; in aa_splitn_fqname() 125 split++; in aa_splitn_fqname() [all …]
|
D | policy_ns.c | 177 const char *split; in __aa_lookupn_ns() local 179 for (split = strnstr(hname, "//", n); split; in __aa_lookupn_ns() 180 split = strnstr(hname, "//", n)) { in __aa_lookupn_ns() 181 ns = __aa_findn_ns(&ns->sub_ns, hname, split - hname); in __aa_lookupn_ns() 185 n -= split + 2 - hname; in __aa_lookupn_ns() 186 hname = split + 2; in __aa_lookupn_ns()
|
D | policy.c | 448 char *split; in __lookup_parent() local 452 for (split = strstr(hname, "//"); split;) { in __lookup_parent() 454 split - hname); in __lookup_parent() 458 hname = split + 2; in __lookup_parent() 459 split = strstr(hname, "//"); in __lookup_parent() 483 char *split; in __create_missing_ancestors() local 490 for (split = strstr(hname, "//"); split;) { in __create_missing_ancestors() 493 split - hname); in __create_missing_ancestors() 495 const char *name = kstrndup(hname, split - hname, in __create_missing_ancestors() 507 hname = split + 2; in __create_missing_ancestors() [all …]
|
/linux-6.8/drivers/virtio/ |
D | virtio_ring.c | 201 struct vring_virtqueue_split split; member 463 struct vring_desc_extra *extra = vq->split.desc_extra; in vring_unmap_one_split() 524 struct vring_desc_extra *extra = vring->split.desc_extra; in virtqueue_add_desc_split() 580 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split() 591 desc = vq->split.vring.desc; in virtqueue_add_split() 647 vq->split.desc_extra[prev & (vq->split.vring.num - 1)].flags &= in virtqueue_add_split() 662 virtqueue_add_desc_split(_vq, vq->split.vring.desc, in virtqueue_add_split() 674 vq->free_head = vq->split.desc_extra[head].next; in virtqueue_add_split() 679 vq->split.desc_state[head].data = data; in virtqueue_add_split() 681 vq->split.desc_state[head].indir_desc = desc; in virtqueue_add_split() [all …]
|
/linux-6.8/arch/x86/kernel/cpu/ |
D | cacheinfo.c | 152 } split; member 161 } split; member 168 } split; member 300 eax->split.is_self_initializing = 1; in amd_cpuid4() 301 eax->split.type = types[leaf]; in amd_cpuid4() 302 eax->split.level = levels[leaf]; in amd_cpuid4() 303 eax->split.num_threads_sharing = 0; in amd_cpuid4() 304 eax->split.num_cores_on_die = __this_cpu_read(cpu_info.x86_max_cores) - 1; in amd_cpuid4() 308 eax->split.is_fully_associative = 1; in amd_cpuid4() 309 ebx->split.coherency_line_size = line_size - 1; in amd_cpuid4() [all …]
|
/linux-6.8/Documentation/mm/ |
D | split_page_table_lock.rst | 8 scalability, split page table lock was introduced. 10 With split page table lock we have separate per-table lock to serialize 11 access to the table. At the moment we use split lock for PTE and PMD 38 If split lock is disabled, all tables are guarded by mm->page_table_lock. 43 Hugetlb and split page table lock 46 Hugetlb can support several page sizes. We use split lock only for PMD 52 takes pmd split lock for PMD_SIZE page, mm->page_table_lock 57 Support of split page table lock by an architecture 60 There's no need in special enabling of PTE split page table lock: everything 68 PMD split lock only makes sense if you have more than two page table [all …]
|
/linux-6.8/arch/x86/events/amd/ |
D | lbr.c | 46 } split; member 57 } split; member 178 if (!entry.to.split.valid && !entry.to.split.spec) in amd_pmu_lbr_read() 183 br[out].from = sign_ext_branch_ip(entry.from.split.ip); in amd_pmu_lbr_read() 184 br[out].to = sign_ext_branch_ip(entry.to.split.ip); in amd_pmu_lbr_read() 185 br[out].mispred = entry.from.split.mispredict; in amd_pmu_lbr_read() 204 idx = (entry.to.split.valid << 1) | entry.to.split.spec; in amd_pmu_lbr_read() 434 x86_pmu.lbr_nr = ebx.split.lbr_v2_stack_sz; in amd_pmu_lbr_init()
|
D | uncore.c | 69 } split; member 394 return info->split.cid; in amd_uncore_ctx_cid() 401 return info->split.gid; in amd_uncore_ctx_gid() 408 return info->split.num_pmcs; in amd_uncore_ctx_num_pmcs() 639 info.split.aux_data = 0; in amd_uncore_df_ctx_scan() 640 info.split.num_pmcs = NUM_COUNTERS_NB; in amd_uncore_df_ctx_scan() 641 info.split.gid = 0; in amd_uncore_df_ctx_scan() 642 info.split.cid = topology_die_id(cpu); in amd_uncore_df_ctx_scan() 646 info.split.num_pmcs = ebx.split.num_df_pmc; in amd_uncore_df_ctx_scan() 772 info.split.aux_data = 0; in amd_uncore_l3_ctx_scan() [all …]
|
/linux-6.8/lib/zstd/compress/ |
D | zstd_ldm.c | 291 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_fillHashTable() local 292 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_fillHashTable() 296 entry.offset = (U32)(split - base); in ZSTD_ldm_fillHashTable() 369 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_generateSequences_internal() local 370 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_generateSequences_internal() 373 candidates[n].split = split; in ZSTD_ldm_generateSequences_internal() 384 BYTE const* const split = candidates[n].split; in ZSTD_ldm_generateSequences_internal() local 392 newEntry.offset = (U32)(split - base); in ZSTD_ldm_generateSequences_internal() 398 if (split < anchor) { in ZSTD_ldm_generateSequences_internal() 418 ZSTD_count_2segments(split, pMatch, iend, matchEnd, lowPrefixPtr); in ZSTD_ldm_generateSequences_internal() [all …]
|
/linux-6.8/drivers/thermal/intel/ |
D | intel_hfi.c | 64 } split; member 74 } split; member 336 info->index = edx.split.index; in init_hfi_cpu_index() 545 if (!edx.split.capabilities.split.performance) { in hfi_parse_features() 554 edx.split.capabilities.split.__reserved = 0; in hfi_parse_features() 555 nr_capabilities = hweight8(edx.split.capabilities.bits); in hfi_parse_features() 558 hfi_features.nr_table_pages = edx.split.table_pages + 1; in hfi_parse_features()
|
/linux-6.8/Documentation/arch/x86/ |
D | buslock.rst | 16 A split lock is any atomic operation whose operand crosses two cache lines. 20 A bus lock is acquired through either split locked access to writeback (WB) 29 mechanisms to detect split locks and bus locks. 31 #AC exception for split lock detection 34 Beginning with the Tremont Atom CPU split lock operations may raise an 35 Alignment Check (#AC) exception when a split lock operation is attempted. 51 |split_lock_detect=|#AC for split lock |#DB for bus lock | 60 | |split lock in parallel. | | 89 generating split lock and bus lock to block the hard real time code to 102 Disable checking for split lock and bus lock. This option can be useful if
|
/linux-6.8/tools/hv/ |
D | vmbus_testing | 154 f_name = f_path.split("/")[-1] 169 .format(device.split("/")[5])) 172 .format(device.split("/")[5])) 209 .format(state_path.split("/")[5])) 213 .format(state_path.split("/")[5])) 238 interrupt.split("/")[5])) 241 message.split("/")[5])) 263 print("ALL testing now OFF for {}".format(device.split("/")[-1]))
|
/linux-6.8/arch/x86/include/asm/ |
D | perf_event.h | 155 } split; member 168 } split; member 179 } split; member 203 } split; member 215 } split; member 230 } split; member 248 } split; member
|
/linux-6.8/scripts/ |
D | jobserver-exec | 23 opts = [x for x in flags.split(" ") if x.startswith("--jobserver")] 28 fds = opts[-1].split("=", 1)[1] 38 reader, writer = [int(x) for x in fds.split(",", 1)]
|
/linux-6.8/security/apparmor/include/ |
D | lib.h | 161 char *split; in basename() local 164 for (split = strstr(hname, "//"); split; split = strstr(hname, "//")) in basename() 165 hname = split + 2; in basename()
|
/linux-6.8/arch/x86/kvm/vmx/ |
D | pmu_intel.c | 523 pmu->version = eax.split.version_id; in intel_pmu_refresh() 527 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh() 529 eax.split.bit_width = min_t(int, eax.split.bit_width, in intel_pmu_refresh() 531 pmu->counter_bitmask[KVM_PMC_GP] = ((u64)1 << eax.split.bit_width) - 1; in intel_pmu_refresh() 532 eax.split.mask_length = min_t(int, eax.split.mask_length, in intel_pmu_refresh() 535 ((1ull << eax.split.mask_length) - 1); in intel_pmu_refresh() 540 pmu->nr_arch_fixed_counters = min_t(int, edx.split.num_counters_fixed, in intel_pmu_refresh() 542 edx.split.bit_width_fixed = min_t(int, edx.split.bit_width_fixed, in intel_pmu_refresh() 545 ((u64)1 << edx.split.bit_width_fixed) - 1; in intel_pmu_refresh()
|
/linux-6.8/Documentation/sphinx/ |
D | maintainers_include.py | 122 field, details = line.split(':', 1) 148 for separated in output.split('\n'): 163 for separated in field_content.split('\n'): 183 (path, tail) = os.path.split(path)
|
/linux-6.8/block/ |
D | blk-merge.c | 288 goto split; in bio_split_rw() 298 goto split; in bio_split_rw() 307 split: in bio_split_rw() 355 struct bio *split; in __bio_split_to_limits() local 360 split = bio_split_discard(bio, lim, nr_segs, bs); in __bio_split_to_limits() 363 split = bio_split_write_zeroes(bio, lim, nr_segs, bs); in __bio_split_to_limits() 366 split = bio_split_rw(bio, lim, nr_segs, bs, in __bio_split_to_limits() 368 if (IS_ERR(split)) in __bio_split_to_limits() 373 if (split) { in __bio_split_to_limits() 375 split->bi_opf |= REQ_NOMERGE; in __bio_split_to_limits() [all …]
|