/linux/drivers/media/dvb-core/ |
H A D | dvb_ringbuffer.c | 136 size_t split; in dvb_ringbuffer_read_user() local 138 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user() 139 if (split > 0) { in dvb_ringbuffer_read_user() 140 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user() 142 buf += split; in dvb_ringbuffer_read_user() 143 todo -= split; in dvb_ringbuffer_read_user() 162 size_t split; in dvb_ringbuffer_read() local 164 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read() 165 if (split > 0) { in dvb_ringbuffer_read() 166 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read() [all …]
|
/linux/lib/ |
H A D | sg_split.c | 81 struct sg_splitter *split; in sg_split_phys() local 83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys() 84 in_sg = split->in_sg0; in sg_split_phys() 85 out_sg = split->out_sg; in sg_split_phys() 86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys() 89 out_sg->offset += split->skip_sg0; in sg_split_phys() 90 out_sg->length -= split->skip_sg0; in sg_split_phys() 96 out_sg[-1].length = split->length_last_sg; in sg_split_phys() 105 struct sg_splitter *split; in sg_split_mapped() local 107 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_mapped() [all …]
|
/linux/tools/verification/rvgen/rvgen/ |
H A D | automata.py | 55 line = dot_lines[cursor].split() 65 while self.__dot_lines[cursor].split()[0] != "{node": 71 while self.__dot_lines[cursor].split()[0] != "{node": 73 while self.__dot_lines[cursor].split()[0] == "{node": 88 while self.__dot_lines[cursor].split()[0] == "{node": 89 line = self.__dot_lines[cursor].split() 128 if self.__dot_lines[cursor].split()[1] == "->": 129 line = self.__dot_lines[cursor].split() 137 for i in event.split(): 166 if self.__dot_lines[cursor].split()[1] == "->": [all …]
|
/linux/fs/btrfs/ |
H A D | extent_map.c | 802 struct extent_map *split; in btrfs_drop_extent_map_range() local 828 split = btrfs_alloc_extent_map(); in btrfs_drop_extent_map_range() 876 if (!split) { in btrfs_drop_extent_map_range() 877 split = split2; in btrfs_drop_extent_map_range() 879 if (!split) in btrfs_drop_extent_map_range() 882 split->start = em->start; in btrfs_drop_extent_map_range() 883 split->len = start - em->start; in btrfs_drop_extent_map_range() 886 split->disk_bytenr = em->disk_bytenr; in btrfs_drop_extent_map_range() 887 split->disk_num_bytes = em->disk_num_bytes; in btrfs_drop_extent_map_range() 888 split->offset = em->offset; in btrfs_drop_extent_map_range() [all …]
|
/linux/drivers/virtio/ |
H A D | virtio_ring.c | 201 struct vring_virtqueue_split split; member 567 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split() 579 desc = vq->split.vring.desc; in virtqueue_add_split() 580 extra = vq->split.desc_extra; in virtqueue_add_split() 637 vq->split.desc_extra[prev & (vq->split.vring.num - 1)].flags &= in virtqueue_add_split() 648 virtqueue_add_desc_split(_vq, vq->split.vring.desc, in virtqueue_add_split() 649 vq->split.desc_extra, in virtqueue_add_split() 660 vq->free_head = vq->split.desc_extra[head].next; in virtqueue_add_split() 665 vq->split.desc_state[head].data = data; in virtqueue_add_split() 667 vq->split.desc_state[head].indir_desc = desc; in virtqueue_add_split() [all …]
|
/linux/tools/testing/selftests/vDSO/ |
H A D | vdso_test_chacha.c | 99 for (unsigned int split = 0; split < BLOCKS; ++split) { in main() local 102 if (split) in main() 103 __arch_chacha20_blocks_nostack(output2, key, counter2, split); in main() 104 __arch_chacha20_blocks_nostack(output2 + split * BLOCK_SIZE, key, counter2, BLOCKS - split); in main() 106 ksft_exit_fail_msg("Main loop outputs do not match on trial %u, split %u\n", trial, split); in main() 108 ksft_exit_fail_msg("Main loop counters do not match on trial %u, split %u\n", trial, split); in main()
|
/linux/Documentation/mm/ |
H A D | split_page_table_lock.rst | 8 scalability, split page table lock was introduced. 10 With split page table lock we have separate per-table lock to serialize 11 access to the table. At the moment we use split lock for PTE and PMD 42 If split lock is disabled, all tables are guarded by mm->page_table_lock. 47 Hugetlb and split page table lock 50 Hugetlb can support several page sizes. We use split lock only for PMD 56 takes pmd split lock for PMD_SIZE page, mm->page_table_lock 61 Support of split page table lock by an architecture 64 There's no need in special enabling of PTE split page table lock: everything 72 PMD split lock only makes sense if you have more than two page table [all …]
|
/linux/arch/x86/events/amd/ |
H A D | lbr.c | 47 } split; member 58 } split; member 180 if ((!entry.to.split.valid && !entry.to.split.spec) || in amd_pmu_lbr_read() 181 entry.to.split.reserved) in amd_pmu_lbr_read() 186 br[out].from = sign_ext_branch_ip(entry.from.split.ip); in amd_pmu_lbr_read() 187 br[out].to = sign_ext_branch_ip(entry.to.split.ip); in amd_pmu_lbr_read() 188 br[out].mispred = entry.from.split.mispredict; in amd_pmu_lbr_read() 207 idx = (entry.to.split.valid << 1) | entry.to.split.spec; in amd_pmu_lbr_read() 431 x86_pmu.lbr_nr = ebx.split.lbr_v2_stack_sz; in amd_pmu_lbr_init()
|
H A D | uncore.c | 73 } split; member 455 return info->split.cid; in amd_uncore_ctx_cid() 462 return info->split.gid; in amd_uncore_ctx_gid() 469 return info->split.num_pmcs; in amd_uncore_ctx_num_pmcs() 703 info.split.aux_data = 0; in amd_uncore_df_ctx_scan() 704 info.split.num_pmcs = NUM_COUNTERS_NB; in amd_uncore_df_ctx_scan() 705 info.split.gid = 0; in amd_uncore_df_ctx_scan() 706 info.split.cid = topology_logical_package_id(cpu); in amd_uncore_df_ctx_scan() 710 info.split.num_pmcs = ebx.split.num_df_pmc; in amd_uncore_df_ctx_scan() 839 info.split.aux_data = 0; in amd_uncore_l3_ctx_scan() [all …]
|
/linux/arch/x86/kernel/cpu/resctrl/ |
H A D | internal.h | 133 } split; member 141 } split; member 150 } split; member 158 } split; member
|
/linux/arch/x86/include/asm/ |
H A D | perf_event.h | 164 } split; member 177 } split; member 188 } split; member 210 } split; member 221 } split; member 237 } split; member 249 } split; member 264 } split; member 282 } split; member
|
/linux/lib/zstd/compress/ |
H A D | zstd_ldm.c | 311 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_fillHashTable() local 312 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_fillHashTable() 316 entry.offset = (U32)(split - base); in ZSTD_ldm_fillHashTable() 391 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_generateSequences_internal() local 392 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_generateSequences_internal() 395 candidates[n].split = split; in ZSTD_ldm_generateSequences_internal() 406 BYTE const* const split = candidates[n].split; in ZSTD_ldm_generateSequences_internal() local 414 newEntry.offset = (U32)(split - base); in ZSTD_ldm_generateSequences_internal() 420 if (split < anchor) { in ZSTD_ldm_generateSequences_internal() 440 ZSTD_count_2segments(split, pMatch, iend, matchEnd, lowPrefixPtr); in ZSTD_ldm_generateSequences_internal() [all …]
|
/linux/Documentation/arch/x86/ |
H A D | buslock.rst | 16 A split lock is any atomic operation whose operand crosses two cache lines. 20 A bus lock is acquired through either split locked access to writeback (WB) 29 mechanisms to detect split locks and bus locks. Some AMD processors also 32 #AC exception for split lock detection 35 Beginning with the Tremont Atom CPU split lock operations may raise an 36 Alignment Check (#AC) exception when a split lock operation is attempted. 52 |split_lock_detect=|#AC for split lock |#DB for bus lock | 61 | |split lock in parallel. | | 90 generating split lock and bus lock to block the hard real time code to 103 Disable checking for split lock and bus lock. This option can be useful if
|
/linux/tools/hv/ |
H A D | vmbus_testing | 154 f_name = f_path.split("/")[-1] 169 .format(device.split("/")[5])) 172 .format(device.split("/")[5])) 209 .format(state_path.split("/")[5])) 213 .format(state_path.split("/")[5])) 238 interrupt.split("/")[5])) 241 message.split("/")[5])) 263 print("ALL testing now OFF for {}".format(device.split("/")[-1]))
|
/linux/drivers/thermal/intel/ |
H A D | intel_hfi.c | 64 } split; member 74 } split; member 337 info->index = edx.split.index; in init_hfi_cpu_index() 549 if (!edx.split.capabilities.split.performance) { in hfi_parse_features() 558 edx.split.capabilities.split.__reserved = 0; in hfi_parse_features() 559 nr_capabilities = hweight8(edx.split.capabilities.bits); in hfi_parse_features() 562 hfi_features.nr_table_pages = edx.split.table_pages + 1; in hfi_parse_features()
|
/linux/security/apparmor/ |
H A D | lib.c | 193 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local 197 if (split) { in aa_splitn_fqname() 198 *ns_len = split - *ns_name; in aa_splitn_fqname() 201 split++; in aa_splitn_fqname() 202 if (end - split > 1 && strncmp(split, "//", 2) == 0) in aa_splitn_fqname() 203 split += 2; in aa_splitn_fqname() 204 name = skipn_spaces(split, end - split); in aa_splitn_fqname()
|
H A D | policy.c | 482 char *split; in __lookup_parent() local 486 for (split = strstr(hname, "//"); split;) { in __lookup_parent() 488 split - hname); in __lookup_parent() 492 hname = split + 2; in __lookup_parent() 493 split = strstr(hname, "//"); in __lookup_parent() 517 char *split; in __create_missing_ancestors() local 524 for (split = strstr(hname, "//"); split;) { in __create_missing_ancestors() 527 split - hname); in __create_missing_ancestors() 529 const char *name = kstrndup(hname, split - hname, in __create_missing_ancestors() 541 hname = split + 2; in __create_missing_ancestors() [all …]
|
/linux/drivers/md/dm-vdo/indexer/ |
H A D | volume-index.c | 319 struct split_config *split) in split_configuration() argument 325 split->hook_config = *config; in split_configuration() 326 split->hook_geometry = *config->geometry; in split_configuration() 327 split->hook_config.geometry = &split->hook_geometry; in split_configuration() 328 split->non_hook_config = *config; in split_configuration() 329 split->non_hook_geometry = *config->geometry; in split_configuration() 330 split->non_hook_config.geometry = &split->non_hook_geometry; in split_configuration() 338 split->hook_geometry.records_per_chapter = sample_records; in split_configuration() 339 split->non_hook_geometry.records_per_chapter -= sample_records; in split_configuration() 342 split->hook_geometry.sparse_chapters_per_volume = 0; in split_configuration() [all …]
|
/linux/scripts/ |
H A D | jobserver-exec | 23 opts = [x for x in flags.split(" ") if x.startswith("--jobserver")] 28 fds = opts[-1].split("=", 1)[1] 38 reader, writer = [int(x) for x in fds.split(",", 1)]
|
/linux/tools/testing/selftests/turbostat/ |
H A D | added_perf_counters.py | 157 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t') 176 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t')
|
/linux/arch/x86/kvm/vmx/ |
H A D | pmu_intel.c | 531 pmu->version = eax.split.version_id; in intel_pmu_refresh() 535 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh() 537 eax.split.bit_width = min_t(int, eax.split.bit_width, in intel_pmu_refresh() 539 pmu->counter_bitmask[KVM_PMC_GP] = ((u64)1 << eax.split.bit_width) - 1; in intel_pmu_refresh() 540 eax.split.mask_length = min_t(int, eax.split.mask_length, in intel_pmu_refresh() 543 ((1ull << eax.split.mask_length) - 1); in intel_pmu_refresh() 548 pmu->nr_arch_fixed_counters = min_t(int, edx.split.num_counters_fixed, in intel_pmu_refresh() 550 edx.split.bit_width_fixed = min_t(int, edx.split.bit_width_fixed, in intel_pmu_refresh() 553 ((u64)1 << edx.split.bit_width_fixed) - 1; in intel_pmu_refresh()
|
/linux/fs/iomap/ |
H A D | ioend.c | 384 struct bio *split; in iomap_split_ioend() local 407 split = bio_split(bio, sector_offset, GFP_NOFS, &iomap_ioend_bioset); in iomap_split_ioend() 408 if (IS_ERR(split)) in iomap_split_ioend() 409 return ERR_CAST(split); in iomap_split_ioend() 410 split->bi_private = bio->bi_private; in iomap_split_ioend() 411 split->bi_end_io = bio->bi_end_io; in iomap_split_ioend() 413 split_ioend = iomap_init_ioend(ioend->io_inode, split, ioend->io_offset, in iomap_split_ioend()
|
/linux/Documentation/sphinx/ |
H A D | maintainers_include.py | 122 field, details = line.split(':', 1) 148 for separated in output.split('\n'): 163 for separated in field_content.split('\n'): 183 (path, tail) = os.path.split(path)
|
/linux/security/apparmor/include/ |
H A D | lib.h | 180 char *split; in basename() local 183 for (split = strstr(hname, "//"); split; split = strstr(hname, "//")) in basename() 184 hname = split + 2; in basename()
|
/linux/fs/jfs/ |
H A D | jfs_dtree.c | 141 struct dtsplit * split, struct btstack * btstack); 143 static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split, 147 struct dtsplit * split, struct btstack * btstack); 150 struct dtsplit * split, struct metapage ** rmpp); 823 struct dtsplit split; /* split information */ in dtInsert() local 866 split.mp = mp; in dtInsert() 867 split.index = index; in dtInsert() 868 split.nslot = n; in dtInsert() 869 split.key = name; in dtInsert() 870 split.data = &data; in dtInsert() [all …]
|