| /linux/drivers/media/dvb-core/ |
| H A D | dvb_ringbuffer.c | 137 size_t split; in dvb_ringbuffer_read_user() local 139 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user() 140 if (split > 0) { in dvb_ringbuffer_read_user() 141 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user() 143 buf += split; in dvb_ringbuffer_read_user() 144 todo -= split; in dvb_ringbuffer_read_user() 163 size_t split; in dvb_ringbuffer_read() local 165 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read() 166 if (split > 0) { in dvb_ringbuffer_read() 167 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read() [all …]
|
| /linux/lib/ |
| H A D | sg_split.c | 81 struct sg_splitter *split; in sg_split_phys() local 83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys() 84 in_sg = split->in_sg0; in sg_split_phys() 85 out_sg = split->out_sg; in sg_split_phys() 86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys() 89 out_sg->offset += split->skip_sg0; in sg_split_phys() 90 out_sg->length -= split->skip_sg0; in sg_split_phys() 96 out_sg[-1].length = split->length_last_sg; in sg_split_phys() 105 struct sg_splitter *split; in sg_split_mapped() local 107 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_mapped() [all …]
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | cacheinfo.c | 53 } split; member 62 } split; member 69 } split; member 207 eax->split.is_self_initializing = 1; in legacy_amd_cpuid4() 208 eax->split.type = types[index]; in legacy_amd_cpuid4() 209 eax->split.level = levels[index]; in legacy_amd_cpuid4() 210 eax->split.num_threads_sharing = 0; in legacy_amd_cpuid4() 211 eax->split.num_cores_on_die = topology_num_cores_per_package(); in legacy_amd_cpuid4() 214 eax->split.is_fully_associative = 1; in legacy_amd_cpuid4() 216 ebx->split.coherency_line_size = line_size - 1; in legacy_amd_cpuid4() [all …]
|
| /linux/fs/jfs/ |
| H A D | jfs_xtree.c | 76 uint split; member 89 struct xtsplit * split, struct btstack * btstack); 91 static int xtSplitPage(tid_t tid, struct inode *ip, struct xtsplit * split, 95 struct xtsplit * split, struct metapage ** rmpp); 539 struct xtsplit split; /* split information */ in xtInsert() local 601 split.mp = mp; in xtInsert() 602 split.index = index; in xtInsert() 603 split.flag = xflag; in xtInsert() 604 split.off = xoff; in xtInsert() 605 split.len = xlen; in xtInsert() [all …]
|
| H A D | jfs_dtree.c | 141 struct dtsplit * split, struct btstack * btstack); 143 static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split, 147 struct dtsplit * split, struct btstack * btstack); 150 struct dtsplit * split, struct metapage ** rmpp); 823 struct dtsplit split; /* split information */ in dtInsert() local 866 split.mp = mp; in dtInsert() 867 split.index = index; in dtInsert() 868 split.nslot = n; in dtInsert() 869 split.key = name; in dtInsert() 870 split.data = &data; in dtInsert() [all …]
|
| /linux/fs/btrfs/ |
| H A D | extent_map.c | 814 struct extent_map *split; in btrfs_drop_extent_map_range() local 840 split = btrfs_alloc_extent_map(); in btrfs_drop_extent_map_range() 888 if (!split) { in btrfs_drop_extent_map_range() 889 split = split2; in btrfs_drop_extent_map_range() 891 if (!split) in btrfs_drop_extent_map_range() 894 split->start = em->start; in btrfs_drop_extent_map_range() 895 split->len = start - em->start; in btrfs_drop_extent_map_range() 898 split->disk_bytenr = em->disk_bytenr; in btrfs_drop_extent_map_range() 899 split->disk_num_bytes = em->disk_num_bytes; in btrfs_drop_extent_map_range() 900 split->offset = em->offset; in btrfs_drop_extent_map_range() [all …]
|
| /linux/tools/verification/rvgen/rvgen/ |
| H A D | automata.py | 55 line = dot_lines[cursor].split() 65 while self.__dot_lines[cursor].split()[0] != "{node": 71 while self.__dot_lines[cursor].split()[0] != "{node": 73 while self.__dot_lines[cursor].split()[0] == "{node": 88 while self.__dot_lines[cursor].split()[0] == "{node": 89 line = self.__dot_lines[cursor].split() 128 if self.__dot_lines[cursor].split()[1] == "->": 129 line = self.__dot_lines[cursor].split() 137 for i in event.split(): 166 if self.__dot_lines[cursor].split()[1] == "->": [all …]
|
| /linux/drivers/virtio/ |
| H A D | virtio_ring.c | 251 struct vring_virtqueue_split split; member 639 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split() 651 desc = vq->split.vring.desc; in virtqueue_add_split() 652 extra = vq->split.desc_extra; in virtqueue_add_split() 721 virtqueue_add_desc_split(vq, vq->split.vring.desc, in virtqueue_add_split() 722 vq->split.desc_extra, in virtqueue_add_split() 734 if (vq->free_head >= vq->split.vring.num) in virtqueue_add_split() 735 vq->free_head -= vq->split.vring.num; in virtqueue_add_split() 736 vq->split.desc_state[head].total_in_len = total_in_len; in virtqueue_add_split() 738 vq->free_head = vq->split.desc_extra[head].next; in virtqueue_add_split() [all …]
|
| /linux/tools/testing/selftests/vDSO/ |
| H A D | vdso_test_chacha.c | 99 for (unsigned int split = 0; split < BLOCKS; ++split) { in main() local 102 if (split) in main() 103 __arch_chacha20_blocks_nostack(output2, key, counter2, split); in main() 104 __arch_chacha20_blocks_nostack(output2 + split * BLOCK_SIZE, key, counter2, BLOCKS - split); in main() 106 ksft_exit_fail_msg("Main loop outputs do not match on trial %u, split %u\n", trial, split); in main() 108 ksft_exit_fail_msg("Main loop counters do not match on trial %u, split %u\n", trial, split); in main()
|
| /linux/Documentation/mm/ |
| H A D | split_page_table_lock.rst | 8 scalability, split page table lock was introduced. 10 With split page table lock we have separate per-table lock to serialize 11 access to the table. At the moment we use split lock for PTE and PMD 42 If split lock is disabled, all tables are guarded by mm->page_table_lock. 47 Hugetlb and split page table lock 50 Hugetlb can support several page sizes. We use split lock only for PMD 56 takes pmd split lock for PMD_SIZE page, mm->page_table_lock 61 Support of split page table lock by an architecture 64 There's no need in special enabling of PTE split page table lock: everything 72 PMD split lock only makes sense if you have more than two page table [all …]
|
| /linux/arch/x86/events/amd/ |
| H A D | lbr.c | 47 } split; member 58 } split; member 180 if ((!entry.to.split.valid && !entry.to.split.spec) || in amd_pmu_lbr_read() 181 entry.to.split.reserved) in amd_pmu_lbr_read() 186 br[out].from = sign_ext_branch_ip(entry.from.split.ip); in amd_pmu_lbr_read() 187 br[out].to = sign_ext_branch_ip(entry.to.split.ip); in amd_pmu_lbr_read() 188 br[out].mispred = entry.from.split.mispredict; in amd_pmu_lbr_read() 207 idx = (entry.to.split.valid << 1) | entry.to.split.spec; in amd_pmu_lbr_read() 431 x86_pmu.lbr_nr = ebx.split.lbr_v2_stack_sz; in amd_pmu_lbr_init()
|
| H A D | uncore.c | 73 } split; member 455 return info->split.cid; in amd_uncore_ctx_cid() 462 return info->split.gid; in amd_uncore_ctx_gid() 469 return info->split.num_pmcs; in amd_uncore_ctx_num_pmcs() 700 info.split.aux_data = 0; in amd_uncore_df_ctx_scan() 701 info.split.num_pmcs = NUM_COUNTERS_NB; in amd_uncore_df_ctx_scan() 702 info.split.gid = 0; in amd_uncore_df_ctx_scan() 703 info.split.cid = topology_logical_package_id(cpu); in amd_uncore_df_ctx_scan() 707 info.split.num_pmcs = ebx.split.num_df_pmc; in amd_uncore_df_ctx_scan() 836 info.split.aux_data = 0; in amd_uncore_l3_ctx_scan() [all …]
|
| /linux/lib/zstd/compress/ |
| H A D | zstd_ldm.c | 311 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_fillHashTable() local 312 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_fillHashTable() 316 entry.offset = (U32)(split - base); in ZSTD_ldm_fillHashTable() 391 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_generateSequences_internal() local 392 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_generateSequences_internal() 395 candidates[n].split = split; in ZSTD_ldm_generateSequences_internal() 406 BYTE const* const split = candidates[n].split; in ZSTD_ldm_generateSequences_internal() local 414 newEntry.offset = (U32)(split - base); in ZSTD_ldm_generateSequences_internal() 420 if (split < anchor) { in ZSTD_ldm_generateSequences_internal() 440 ZSTD_count_2segments(split, pMatch, iend, matchEnd, lowPrefixPtr); in ZSTD_ldm_generateSequences_internal() [all …]
|
| /linux/arch/x86/kernel/cpu/resctrl/ |
| H A D | internal.h | 160 } split; member 168 } split; member 177 } split; member 185 } split; member 221 } split; member
|
| /linux/Documentation/arch/x86/ |
| H A D | buslock.rst | 16 A split lock is any atomic operation whose operand crosses two cache lines. 20 A bus lock is acquired through either split locked access to writeback (WB) 29 mechanisms to detect split locks and bus locks. Some AMD processors also 32 #AC exception for split lock detection 35 Beginning with the Tremont Atom CPU split lock operations may raise an 36 Alignment Check (#AC) exception when a split lock operation is attempted. 52 |split_lock_detect=|#AC for split lock |#DB for bus lock | 61 | |split lock in parallel. | | 90 generating split lock and bus lock to block the hard real time code to 103 Disable checking for split lock and bus lock. This option can be useful if
|
| /linux/security/apparmor/ |
| H A D | lib.c | 195 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local 199 if (split) { in aa_splitn_fqname() 200 *ns_len = split - *ns_name; in aa_splitn_fqname() 203 split++; in aa_splitn_fqname() 204 if (end - split > 1 && strncmp(split, "//", 2) == 0) in aa_splitn_fqname() 205 split += 2; in aa_splitn_fqname() 206 name = skipn_spaces(split, end - split); in aa_splitn_fqname()
|
| H A D | policy_ns.c | 177 const char *split; in __aa_lookupn_ns() local 179 for (split = strnstr(hname, "//", n); split; in __aa_lookupn_ns() 180 split = strnstr(hname, "//", n)) { in __aa_lookupn_ns() 181 ns = __aa_findn_ns(&ns->sub_ns, hname, split - hname); in __aa_lookupn_ns() 185 n -= split + 2 - hname; in __aa_lookupn_ns() 186 hname = split + 2; in __aa_lookupn_ns()
|
| H A D | policy.c | 520 char *split; in __lookup_parent() local 524 for (split = strstr(hname, "//"); split;) { in __lookup_parent() 526 split - hname); in __lookup_parent() 530 hname = split + 2; in __lookup_parent() 531 split = strstr(hname, "//"); in __lookup_parent() 555 char *split; in __create_missing_ancestors() local 562 for (split = strstr(hname, "//"); split;) { in __create_missing_ancestors() 565 split - hname); in __create_missing_ancestors() 567 const char *name = kstrndup(hname, split - hname, in __create_missing_ancestors() 579 hname = split + 2; in __create_missing_ancestors() [all …]
|
| /linux/drivers/thermal/intel/ |
| H A D | intel_hfi.c | 64 } split; member 74 } split; member 337 info->index = edx.split.index; in init_hfi_cpu_index() 549 if (!edx.split.capabilities.split.performance) { in hfi_parse_features() 558 edx.split.capabilities.split.__reserved = 0; in hfi_parse_features() 559 nr_capabilities = hweight8(edx.split.capabilities.bits); in hfi_parse_features() 562 hfi_features.nr_table_pages = edx.split.table_pages + 1; in hfi_parse_features()
|
| /linux/tools/hv/ |
| H A D | vmbus_testing | 154 f_name = f_path.split("/")[-1] 169 .format(device.split("/")[5])) 172 .format(device.split("/")[5])) 209 .format(state_path.split("/")[5])) 213 .format(state_path.split("/")[5])) 238 interrupt.split("/")[5])) 241 message.split("/")[5])) 263 print("ALL testing now OFF for {}".format(device.split("/")[-1]))
|
| /linux/arch/x86/include/asm/ |
| H A D | perf_event.h | 170 } split; member 183 } split; member 194 } split; member 221 } split; member 234 } split; member 250 } split; member 262 } split; member 277 } split; member 295 } split; member
|
| /linux/drivers/md/dm-vdo/indexer/ |
| H A D | volume-index.c | 319 struct split_config *split) in split_configuration() argument 325 split->hook_config = *config; in split_configuration() 326 split->hook_geometry = *config->geometry; in split_configuration() 327 split->hook_config.geometry = &split->hook_geometry; in split_configuration() 328 split->non_hook_config = *config; in split_configuration() 329 split->non_hook_geometry = *config->geometry; in split_configuration() 330 split->non_hook_config.geometry = &split->non_hook_geometry; in split_configuration() 338 split->hook_geometry.records_per_chapter = sample_records; in split_configuration() 339 split->non_hook_geometry.records_per_chapter -= sample_records; in split_configuration() 342 split->hook_geometry.sparse_chapters_per_volume = 0; in split_configuration() [all …]
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | pmu_intel.c | 520 pmu->version = eax.split.version_id; in intel_pmu_refresh() 524 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh() 526 eax.split.bit_width = min_t(int, eax.split.bit_width, in intel_pmu_refresh() 528 pmu->counter_bitmask[KVM_PMC_GP] = BIT_ULL(eax.split.bit_width) - 1; in intel_pmu_refresh() 529 eax.split.mask_length = min_t(int, eax.split.mask_length, in intel_pmu_refresh() 531 pmu->available_event_types = ~entry->ebx & (BIT_ULL(eax.split.mask_length) - 1); in intel_pmu_refresh() 554 pmu->nr_arch_fixed_counters = min_t(int, edx.split.num_counters_fixed, in intel_pmu_refresh() 556 edx.split.bit_width_fixed = min_t(int, edx.split.bit_width_fixed, in intel_pmu_refresh() 558 pmu->counter_bitmask[KVM_PMC_FIXED] = BIT_ULL(edx.split.bit_width_fixed) - 1; in intel_pmu_refresh()
|
| /linux/tools/testing/selftests/turbostat/ |
| H A D | added_perf_counters.py | 157 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t') 176 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t')
|
| /linux/Documentation/sphinx/ |
| H A D | maintainers_include.py | 124 field, details = line.split(':', 1) 150 for separated in output.split('\n'): 165 for separated in field_content.split('\n'): 185 (path, tail) = os.path.split(path)
|