/linux/drivers/media/dvb-core/ |
H A D | dvb_ringbuffer.c | 136 size_t split; in dvb_ringbuffer_read_user() local 138 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user() 139 if (split > 0) { in dvb_ringbuffer_read_user() 140 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user() 142 buf += split; in dvb_ringbuffer_read_user() 143 todo -= split; in dvb_ringbuffer_read_user() 162 size_t split; in dvb_ringbuffer_read() local 164 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read() 165 if (split > 0) { in dvb_ringbuffer_read() 166 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read() 185 size_t split; dvb_ringbuffer_write() local 212 size_t split; dvb_ringbuffer_write_user() local 256 size_t split; dvb_ringbuffer_pkt_read_user() local 284 size_t split; dvb_ringbuffer_pkt_read() local [all...] |
/linux/lib/ |
H A D | sg_split.c | 81 struct sg_splitter *split; in sg_split_phys() local 83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys() 84 in_sg = split->in_sg0; in sg_split_phys() 85 out_sg = split->out_sg; in sg_split_phys() 86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys() 89 out_sg->offset += split->skip_sg0; in sg_split_phys() 90 out_sg->length -= split->skip_sg0; in sg_split_phys() 96 out_sg[-1].length = split->length_last_sg; in sg_split_phys() 105 struct sg_splitter *split; in sg_split_mapped() local [all...] |
/linux/tools/verification/rvgen/rvgen/ |
H A D | automata.py | 55 line = dot_lines[cursor].split() 65 while self.__dot_lines[cursor].split()[0] != "{node": 71 while self.__dot_lines[cursor].split()[0] != "{node": 73 while self.__dot_lines[cursor].split()[0] == "{node": 88 while self.__dot_lines[cursor].split()[0] == "{node": 89 line = self.__dot_lines[cursor].split() 128 if self.__dot_lines[cursor].split()[1] == "->": 129 line = self.__dot_lines[cursor].split() 134 # so split them. 137 for i in event.split() [all...] |
/linux/fs/btrfs/ |
H A D | extent_map.c | 795 * are split. 802 struct extent_map *split; in btrfs_drop_extent_map_range() local 826 * because we need to split those two extent maps at the boundaries. in btrfs_drop_extent_map_range() 828 split = btrfs_alloc_extent_map(); in btrfs_drop_extent_map_range() 859 * In case we split the extent map, we want to preserve the in btrfs_drop_extent_map_range() 868 * split it, we can remove it directly. in btrfs_drop_extent_map_range() 876 if (!split) { in btrfs_drop_extent_map_range() 877 split = split2; in btrfs_drop_extent_map_range() 879 if (!split) in btrfs_drop_extent_map_range() 882 split in btrfs_drop_extent_map_range() [all...] |
/linux/drivers/virtio/ |
H A D | virtio_ring.c | 189 * for split ring, it just contains last used index 200 /* Available for split ring */ 201 struct vring_virtqueue_split split; member 567 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split() 579 desc = vq->split.vring.desc; in virtqueue_add_split() 580 extra = vq->split.desc_extra; in virtqueue_add_split() 637 vq->split.desc_extra[prev & (vq->split.vring.num - 1)].flags &= in virtqueue_add_split() 648 virtqueue_add_desc_split(_vq, vq->split.vring.desc, in virtqueue_add_split() 649 vq->split in virtqueue_add_split() [all...] |
/linux/tools/testing/selftests/vDSO/ |
H A D | vdso_test_chacha.c | 99 for (unsigned int split = 0; split < BLOCKS; ++split) { in main() local 102 if (split) in main() 103 __arch_chacha20_blocks_nostack(output2, key, counter2, split); in main() 104 __arch_chacha20_blocks_nostack(output2 + split * BLOCK_SIZE, key, counter2, BLOCKS - split); in main() 106 ksft_exit_fail_msg("Main loop outputs do not match on trial %u, split %u\n", trial, split); in main() 108 ksft_exit_fail_msg("Main loop counters do not match on trial %u, split in main() [all...] |
/linux/Documentation/mm/ |
H A D | split_page_table_lock.rst | 8 scalability, split page table lock was introduced. 10 With split page table lock we have separate per-table lock to serialize 11 access to the table. At the moment we use split lock for PTE and PMD 42 If split lock is disabled, all tables are guarded by mm->page_table_lock. 47 Hugetlb and split page table lock 50 Hugetlb can support several page sizes. We use split lock only for PMD 56 takes pmd split lock for PMD_SIZE page, mm->page_table_lock 61 Support of split page table lock by an architecture 64 There's no need in special enabling of PTE split page table lock: everything 72 PMD split loc [all...] |
/linux/arch/x86/events/amd/ |
H A D | lbr.c | 47 } split; member 58 } split; member 180 if ((!entry.to.split.valid && !entry.to.split.spec) || in amd_pmu_lbr_read() 181 entry.to.split.reserved) in amd_pmu_lbr_read() 186 br[out].from = sign_ext_branch_ip(entry.from.split.ip); in amd_pmu_lbr_read() 187 br[out].to = sign_ext_branch_ip(entry.to.split.ip); in amd_pmu_lbr_read() 188 br[out].mispred = entry.from.split.mispredict; in amd_pmu_lbr_read() 207 idx = (entry.to.split.valid << 1) | entry.to.split in amd_pmu_lbr_read() [all...] |
H A D | uncore.c | 73 } split; member 455 return info->split.cid; in amd_uncore_ctx_cid() 462 return info->split.gid; in amd_uncore_ctx_gid() 469 return info->split.num_pmcs; in amd_uncore_ctx_num_pmcs() 703 info.split.aux_data = 0; in amd_uncore_df_ctx_scan() 704 info.split.num_pmcs = NUM_COUNTERS_NB; in amd_uncore_df_ctx_scan() 705 info.split.gid = 0; in amd_uncore_df_ctx_scan() 706 info.split.cid = topology_logical_package_id(cpu); in amd_uncore_df_ctx_scan() 710 info.split.num_pmcs = ebx.split in amd_uncore_df_ctx_scan() [all...] |
/linux/arch/x86/kernel/cpu/resctrl/ |
H A D | internal.h | 133 } split; member 141 } split; member 150 } split; member 158 } split; member
|
/linux/arch/x86/include/asm/ |
H A D | perf_event.h | 164 } split; member 177 } split; member 188 } split; member 210 } split; member 221 } split; member 237 } split; member 249 } split; member 264 } split; member 282 } split; member
|
/linux/lib/zstd/compress/ |
H A D | zstd_ldm.c | 90 * Registers in the splits array all the split points found in the first 311 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_fillHashTable() local 312 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_fillHashTable() 316 entry.offset = (U32)(split - base); in ZSTD_ldm_fillHashTable() 391 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_generateSequences_internal() local 392 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_generateSequences_internal() 395 candidates[n].split = split; in ZSTD_ldm_generateSequences_internal() 406 BYTE const* const split = candidates[n].split; in ZSTD_ldm_generateSequences_internal() local [all...] |
/linux/Documentation/arch/x86/ |
H A D | buslock.rst | 16 A split lock is any atomic operation whose operand crosses two cache lines. 20 A bus lock is acquired through either split locked access to writeback (WB) 29 mechanisms to detect split locks and bus locks. Some AMD processors also 32 #AC exception for split lock detection 35 Beginning with the Tremont Atom CPU split lock operations may raise an 36 Alignment Check (#AC) exception when a split lock operation is attempted. 52 |split_lock_detect=|#AC for split lock |#DB for bus lock | 61 | |split lock in parallel. | | 90 generating split lock and bus lock to block the hard real time code to 103 Disable checking for split loc [all...] |
/linux/tools/hv/ |
H A D | vmbus_testing | 154 f_name = f_path.split("/")[-1] 169 .format(device.split("/")[5])) 172 .format(device.split("/")[5])) 209 .format(state_path.split("/")[5])) 213 .format(state_path.split("/")[5])) 238 interrupt.split("/")[5])) 241 message.split("/")[5])) 263 print("ALL testing now OFF for {}".format(device.split("/")[-1]))
|
/linux/drivers/thermal/intel/ |
H A D | intel_hfi.c | 64 } split; member 74 } split; member 337 info->index = edx.split.index; in init_hfi_cpu_index() 549 if (!edx.split.capabilities.split.performance) { in hfi_parse_features() 558 edx.split.capabilities.split.__reserved = 0; in hfi_parse_features() 559 nr_capabilities = hweight8(edx.split.capabilities.bits); in hfi_parse_features() 562 hfi_features.nr_table_pages = edx.split.table_pages + 1; in hfi_parse_features()
|
/linux/security/apparmor/ |
H A D | lib.c | 193 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local 197 if (split) { in aa_splitn_fqname() 198 *ns_len = split - *ns_name; in aa_splitn_fqname() 201 split++; in aa_splitn_fqname() 202 if (end - split > 1 && strncmp(split, "//", 2) == 0) in aa_splitn_fqname() 203 split += 2; in aa_splitn_fqname() 204 name = skipn_spaces(split, end - split); in aa_splitn_fqname() 359 * TODO: split int [all...] |
H A D | policy.c | 482 char *split; in __lookup_parent() local 486 for (split = strstr(hname, "//"); split;) { in __lookup_parent() 488 split - hname); in __lookup_parent() 492 hname = split + 2; in __lookup_parent() 493 split = strstr(hname, "//"); in __lookup_parent() 517 char *split; in __create_missing_ancestors() local 524 for (split = strstr(hname, "//"); split;) { in __create_missing_ancestors() 527 split in __create_missing_ancestors() 565 const char *split; __lookupn_profile() local [all...] |
/linux/drivers/md/dm-vdo/indexer/ |
H A D | volume-index.c | 319 struct split_config *split) in split_configuration() argument 325 split->hook_config = *config; in split_configuration() 326 split->hook_geometry = *config->geometry; in split_configuration() 327 split->hook_config.geometry = &split->hook_geometry; in split_configuration() 328 split->non_hook_config = *config; in split_configuration() 329 split->non_hook_geometry = *config->geometry; in split_configuration() 330 split->non_hook_config.geometry = &split->non_hook_geometry; in split_configuration() 338 split in split_configuration() 351 struct split_config split; compute_volume_index_save_bytes() local 1226 struct split_config split; uds_make_volume_index() local [all...] |
/linux/scripts/ |
H A D | jobserver-exec | 23 opts = [x for x in flags.split(" ") if x.startswith("--jobserver")] 28 fds = opts[-1].split("=", 1)[1] 38 reader, writer = [int(x) for x in fds.split(",", 1)]
|
/linux/tools/testing/selftests/turbostat/ |
H A D | added_perf_counters.py | 157 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t') 176 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t')
|
/linux/arch/x86/kvm/vmx/ |
H A D | pmu_intel.c | 531 pmu->version = eax.split.version_id; in intel_pmu_refresh() 535 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh() 537 eax.split.bit_width = min_t(int, eax.split.bit_width, in intel_pmu_refresh() 539 pmu->counter_bitmask[KVM_PMC_GP] = ((u64)1 << eax.split.bit_width) - 1; in intel_pmu_refresh() 540 eax.split.mask_length = min_t(int, eax.split.mask_length, in intel_pmu_refresh() 543 ((1ull << eax.split.mask_length) - 1); in intel_pmu_refresh() 548 pmu->nr_arch_fixed_counters = min_t(int, edx.split.num_counters_fixed, in intel_pmu_refresh() 550 edx.split in intel_pmu_refresh() [all...] |
/linux/fs/iomap/ |
H A D | ioend.c | 369 * If @is_append is set, the split will be based on the hardware limits for 384 struct bio *split; in iomap_split_ioend() local 403 /* ensure the split ioend is still block size aligned */ in iomap_split_ioend() 407 split = bio_split(bio, sector_offset, GFP_NOFS, &iomap_ioend_bioset); in iomap_split_ioend() 408 if (IS_ERR(split)) in iomap_split_ioend() 409 return ERR_CAST(split); in iomap_split_ioend() 410 split->bi_private = bio->bi_private; in iomap_split_ioend() 411 split->bi_end_io = bio->bi_end_io; in iomap_split_ioend() 413 split_ioend = iomap_init_ioend(ioend->io_inode, split, ioend->io_offset, in iomap_split_ioend()
|
/linux/Documentation/sphinx/ |
H A D | maintainers_include.py | 122 field, details = line.split(':', 1) 146 # Re-split on any added newlines in any above parsing. 148 for separated in output.split('\n'): 163 for separated in field_content.split('\n'): 183 (path, tail) = os.path.split(path)
|
/linux/security/apparmor/include/ |
H A D | lib.h | 22 * split individual debug cases out in preparation for finer grained 180 char *split; in basename() local 183 for (split = strstr(hname, "//"); split; split = strstr(hname, "//")) in basename() 184 hname = split + 2; in basename()
|
/linux/fs/jfs/ |
H A D | jfs_dtree.c | 32 * it is extended rather than split (its size is doubled), 33 * until its length becoms 4 KBytes, from then the extent is split 63 * (e.g., if split occurs <abc> and <aBd>, <ABD> trather than <aB> 64 * should be made the router key for the split) 101 /* dtree split parameter */ 141 struct dtsplit * split, struct btstack * btstack); 143 static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split, 147 struct dtsplit * split, struct btstack * btstack); 150 struct dtsplit * split, struct metapage ** rmpp); 603 /* init level count for max pages to split */ in dtSearch() 823 struct dtsplit split; /* split information */ dtInsert() local 927 dtSplitUp(tid_t tid,struct inode * ip,struct dtsplit * split,struct btstack * btstack) dtSplitUp() argument 1326 dtSplitPage(tid_t tid,struct inode * ip,struct dtsplit * split,struct metapage ** rmpp,dtpage_t ** rpp,pxd_t * rpxdp) dtSplitPage() argument 1639 dtExtendPage(tid_t tid,struct inode * ip,struct dtsplit * split,struct btstack * btstack) dtExtendPage() argument 1872 dtSplitRoot(tid_t tid,struct inode * ip,struct dtsplit * split,struct metapage ** rmpp) dtSplitRoot() argument [all...] |