Home
last modified time | relevance | path

Searched refs:split (Results 1 – 25 of 532) sorted by relevance

12345678910>>...22

/linux/drivers/media/dvb-core/
H A Ddvb_ringbuffer.c137 size_t split; in dvb_ringbuffer_read_user() local
139 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user()
140 if (split > 0) { in dvb_ringbuffer_read_user()
141 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user()
143 buf += split; in dvb_ringbuffer_read_user()
144 todo -= split; in dvb_ringbuffer_read_user()
163 size_t split; in dvb_ringbuffer_read() local
165 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read()
166 if (split > 0) { in dvb_ringbuffer_read()
167 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read()
[all …]
/linux/lib/
H A Dsg_split.c81 struct sg_splitter *split; in sg_split_phys() local
83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys()
84 in_sg = split->in_sg0; in sg_split_phys()
85 out_sg = split->out_sg; in sg_split_phys()
86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys()
89 out_sg->offset += split->skip_sg0; in sg_split_phys()
90 out_sg->length -= split->skip_sg0; in sg_split_phys()
96 out_sg[-1].length = split->length_last_sg; in sg_split_phys()
105 struct sg_splitter *split; in sg_split_mapped() local
107 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_mapped()
[all …]
/linux/arch/x86/kernel/cpu/
H A Dcacheinfo.c53 } split; member
62 } split; member
69 } split; member
207 eax->split.is_self_initializing = 1; in legacy_amd_cpuid4()
208 eax->split.type = types[index]; in legacy_amd_cpuid4()
209 eax->split.level = levels[index]; in legacy_amd_cpuid4()
210 eax->split.num_threads_sharing = 0; in legacy_amd_cpuid4()
211 eax->split.num_cores_on_die = topology_num_cores_per_package(); in legacy_amd_cpuid4()
214 eax->split.is_fully_associative = 1; in legacy_amd_cpuid4()
216 ebx->split.coherency_line_size = line_size - 1; in legacy_amd_cpuid4()
[all …]
/linux/fs/jfs/
H A Djfs_xtree.c76 uint split; member
89 struct xtsplit * split, struct btstack * btstack);
91 static int xtSplitPage(tid_t tid, struct inode *ip, struct xtsplit * split,
95 struct xtsplit * split, struct metapage ** rmpp);
539 struct xtsplit split; /* split information */ in xtInsert() local
601 split.mp = mp; in xtInsert()
602 split.index = index; in xtInsert()
603 split.flag = xflag; in xtInsert()
604 split.off = xoff; in xtInsert()
605 split.len = xlen; in xtInsert()
[all …]
H A Djfs_dtree.c141 struct dtsplit * split, struct btstack * btstack);
143 static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split,
147 struct dtsplit * split, struct btstack * btstack);
150 struct dtsplit * split, struct metapage ** rmpp);
823 struct dtsplit split; /* split information */ in dtInsert() local
866 split.mp = mp; in dtInsert()
867 split.index = index; in dtInsert()
868 split.nslot = n; in dtInsert()
869 split.key = name; in dtInsert()
870 split.data = &data; in dtInsert()
[all …]
/linux/fs/btrfs/
H A Dextent_map.c814 struct extent_map *split; in btrfs_drop_extent_map_range() local
840 split = btrfs_alloc_extent_map(); in btrfs_drop_extent_map_range()
888 if (!split) { in btrfs_drop_extent_map_range()
889 split = split2; in btrfs_drop_extent_map_range()
891 if (!split) in btrfs_drop_extent_map_range()
894 split->start = em->start; in btrfs_drop_extent_map_range()
895 split->len = start - em->start; in btrfs_drop_extent_map_range()
898 split->disk_bytenr = em->disk_bytenr; in btrfs_drop_extent_map_range()
899 split->disk_num_bytes = em->disk_num_bytes; in btrfs_drop_extent_map_range()
900 split->offset = em->offset; in btrfs_drop_extent_map_range()
[all …]
/linux/tools/verification/rvgen/rvgen/
H A Dautomata.py55 line = dot_lines[cursor].split()
65 while self.__dot_lines[cursor].split()[0] != "{node":
71 while self.__dot_lines[cursor].split()[0] != "{node":
73 while self.__dot_lines[cursor].split()[0] == "{node":
88 while self.__dot_lines[cursor].split()[0] == "{node":
89 line = self.__dot_lines[cursor].split()
128 if self.__dot_lines[cursor].split()[1] == "->":
129 line = self.__dot_lines[cursor].split()
137 for i in event.split():
166 if self.__dot_lines[cursor].split()[1] == "->":
[all …]
/linux/drivers/virtio/
H A Dvirtio_ring.c251 struct vring_virtqueue_split split; member
639 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split()
651 desc = vq->split.vring.desc; in virtqueue_add_split()
652 extra = vq->split.desc_extra; in virtqueue_add_split()
721 virtqueue_add_desc_split(vq, vq->split.vring.desc, in virtqueue_add_split()
722 vq->split.desc_extra, in virtqueue_add_split()
734 if (vq->free_head >= vq->split.vring.num) in virtqueue_add_split()
735 vq->free_head -= vq->split.vring.num; in virtqueue_add_split()
736 vq->split.desc_state[head].total_in_len = total_in_len; in virtqueue_add_split()
738 vq->free_head = vq->split.desc_extra[head].next; in virtqueue_add_split()
[all …]
/linux/tools/testing/selftests/vDSO/
H A Dvdso_test_chacha.c99 for (unsigned int split = 0; split < BLOCKS; ++split) { in main() local
102 if (split) in main()
103 __arch_chacha20_blocks_nostack(output2, key, counter2, split); in main()
104 __arch_chacha20_blocks_nostack(output2 + split * BLOCK_SIZE, key, counter2, BLOCKS - split); in main()
106 ksft_exit_fail_msg("Main loop outputs do not match on trial %u, split %u\n", trial, split); in main()
108 ksft_exit_fail_msg("Main loop counters do not match on trial %u, split %u\n", trial, split); in main()
/linux/Documentation/mm/
H A Dsplit_page_table_lock.rst8 scalability, split page table lock was introduced.
10 With split page table lock we have separate per-table lock to serialize
11 access to the table. At the moment we use split lock for PTE and PMD
42 If split lock is disabled, all tables are guarded by mm->page_table_lock.
47 Hugetlb and split page table lock
50 Hugetlb can support several page sizes. We use split lock only for PMD
56 takes pmd split lock for PMD_SIZE page, mm->page_table_lock
61 Support of split page table lock by an architecture
64 There's no need in special enabling of PTE split page table lock: everything
72 PMD split lock only makes sense if you have more than two page table
[all …]
/linux/arch/x86/events/amd/
H A Dlbr.c47 } split; member
58 } split; member
180 if ((!entry.to.split.valid && !entry.to.split.spec) || in amd_pmu_lbr_read()
181 entry.to.split.reserved) in amd_pmu_lbr_read()
186 br[out].from = sign_ext_branch_ip(entry.from.split.ip); in amd_pmu_lbr_read()
187 br[out].to = sign_ext_branch_ip(entry.to.split.ip); in amd_pmu_lbr_read()
188 br[out].mispred = entry.from.split.mispredict; in amd_pmu_lbr_read()
207 idx = (entry.to.split.valid << 1) | entry.to.split.spec; in amd_pmu_lbr_read()
431 x86_pmu.lbr_nr = ebx.split.lbr_v2_stack_sz; in amd_pmu_lbr_init()
H A Duncore.c73 } split; member
455 return info->split.cid; in amd_uncore_ctx_cid()
462 return info->split.gid; in amd_uncore_ctx_gid()
469 return info->split.num_pmcs; in amd_uncore_ctx_num_pmcs()
700 info.split.aux_data = 0; in amd_uncore_df_ctx_scan()
701 info.split.num_pmcs = NUM_COUNTERS_NB; in amd_uncore_df_ctx_scan()
702 info.split.gid = 0; in amd_uncore_df_ctx_scan()
703 info.split.cid = topology_logical_package_id(cpu); in amd_uncore_df_ctx_scan()
707 info.split.num_pmcs = ebx.split.num_df_pmc; in amd_uncore_df_ctx_scan()
836 info.split.aux_data = 0; in amd_uncore_l3_ctx_scan()
[all …]
/linux/lib/zstd/compress/
H A Dzstd_ldm.c311 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_fillHashTable() local
312 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_fillHashTable()
316 entry.offset = (U32)(split - base); in ZSTD_ldm_fillHashTable()
391 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_generateSequences_internal() local
392 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_generateSequences_internal()
395 candidates[n].split = split; in ZSTD_ldm_generateSequences_internal()
406 BYTE const* const split = candidates[n].split; in ZSTD_ldm_generateSequences_internal() local
414 newEntry.offset = (U32)(split - base); in ZSTD_ldm_generateSequences_internal()
420 if (split < anchor) { in ZSTD_ldm_generateSequences_internal()
440 ZSTD_count_2segments(split, pMatch, iend, matchEnd, lowPrefixPtr); in ZSTD_ldm_generateSequences_internal()
[all …]
/linux/arch/x86/kernel/cpu/resctrl/
H A Dinternal.h160 } split; member
168 } split; member
177 } split; member
185 } split; member
221 } split; member
/linux/Documentation/arch/x86/
H A Dbuslock.rst16 A split lock is any atomic operation whose operand crosses two cache lines.
20 A bus lock is acquired through either split locked access to writeback (WB)
29 mechanisms to detect split locks and bus locks. Some AMD processors also
32 #AC exception for split lock detection
35 Beginning with the Tremont Atom CPU split lock operations may raise an
36 Alignment Check (#AC) exception when a split lock operation is attempted.
52 |split_lock_detect=|#AC for split lock |#DB for bus lock |
61 | |split lock in parallel. | |
90 generating split lock and bus lock to block the hard real time code to
103 Disable checking for split lock and bus lock. This option can be useful if
/linux/security/apparmor/
H A Dlib.c195 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local
199 if (split) { in aa_splitn_fqname()
200 *ns_len = split - *ns_name; in aa_splitn_fqname()
203 split++; in aa_splitn_fqname()
204 if (end - split > 1 && strncmp(split, "//", 2) == 0) in aa_splitn_fqname()
205 split += 2; in aa_splitn_fqname()
206 name = skipn_spaces(split, end - split); in aa_splitn_fqname()
H A Dpolicy_ns.c177 const char *split; in __aa_lookupn_ns() local
179 for (split = strnstr(hname, "//", n); split; in __aa_lookupn_ns()
180 split = strnstr(hname, "//", n)) { in __aa_lookupn_ns()
181 ns = __aa_findn_ns(&ns->sub_ns, hname, split - hname); in __aa_lookupn_ns()
185 n -= split + 2 - hname; in __aa_lookupn_ns()
186 hname = split + 2; in __aa_lookupn_ns()
H A Dpolicy.c520 char *split; in __lookup_parent() local
524 for (split = strstr(hname, "//"); split;) { in __lookup_parent()
526 split - hname); in __lookup_parent()
530 hname = split + 2; in __lookup_parent()
531 split = strstr(hname, "//"); in __lookup_parent()
555 char *split; in __create_missing_ancestors() local
562 for (split = strstr(hname, "//"); split;) { in __create_missing_ancestors()
565 split - hname); in __create_missing_ancestors()
567 const char *name = kstrndup(hname, split - hname, in __create_missing_ancestors()
579 hname = split + 2; in __create_missing_ancestors()
[all …]
/linux/drivers/thermal/intel/
H A Dintel_hfi.c64 } split; member
74 } split; member
337 info->index = edx.split.index; in init_hfi_cpu_index()
549 if (!edx.split.capabilities.split.performance) { in hfi_parse_features()
558 edx.split.capabilities.split.__reserved = 0; in hfi_parse_features()
559 nr_capabilities = hweight8(edx.split.capabilities.bits); in hfi_parse_features()
562 hfi_features.nr_table_pages = edx.split.table_pages + 1; in hfi_parse_features()
/linux/tools/hv/
H A Dvmbus_testing154 f_name = f_path.split("/")[-1]
169 .format(device.split("/")[5]))
172 .format(device.split("/")[5]))
209 .format(state_path.split("/")[5]))
213 .format(state_path.split("/")[5]))
238 interrupt.split("/")[5]))
241 message.split("/")[5]))
263 print("ALL testing now OFF for {}".format(device.split("/")[-1]))
/linux/arch/x86/include/asm/
H A Dperf_event.h170 } split; member
183 } split; member
194 } split; member
221 } split; member
234 } split; member
250 } split; member
262 } split; member
277 } split; member
295 } split; member
/linux/drivers/md/dm-vdo/indexer/
H A Dvolume-index.c319 struct split_config *split) in split_configuration() argument
325 split->hook_config = *config; in split_configuration()
326 split->hook_geometry = *config->geometry; in split_configuration()
327 split->hook_config.geometry = &split->hook_geometry; in split_configuration()
328 split->non_hook_config = *config; in split_configuration()
329 split->non_hook_geometry = *config->geometry; in split_configuration()
330 split->non_hook_config.geometry = &split->non_hook_geometry; in split_configuration()
338 split->hook_geometry.records_per_chapter = sample_records; in split_configuration()
339 split->non_hook_geometry.records_per_chapter -= sample_records; in split_configuration()
342 split->hook_geometry.sparse_chapters_per_volume = 0; in split_configuration()
[all …]
/linux/arch/x86/kvm/vmx/
H A Dpmu_intel.c520 pmu->version = eax.split.version_id; in intel_pmu_refresh()
524 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh()
526 eax.split.bit_width = min_t(int, eax.split.bit_width, in intel_pmu_refresh()
528 pmu->counter_bitmask[KVM_PMC_GP] = BIT_ULL(eax.split.bit_width) - 1; in intel_pmu_refresh()
529 eax.split.mask_length = min_t(int, eax.split.mask_length, in intel_pmu_refresh()
531 pmu->available_event_types = ~entry->ebx & (BIT_ULL(eax.split.mask_length) - 1); in intel_pmu_refresh()
554 pmu->nr_arch_fixed_counters = min_t(int, edx.split.num_counters_fixed, in intel_pmu_refresh()
556 edx.split.bit_width_fixed = min_t(int, edx.split.bit_width_fixed, in intel_pmu_refresh()
558 pmu->counter_bitmask[KVM_PMC_FIXED] = BIT_ULL(edx.split.bit_width_fixed) - 1; in intel_pmu_refresh()
/linux/tools/testing/selftests/turbostat/
H A Dadded_perf_counters.py157 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t')
176 actual_columns = proc_turbostat.stdout.split(b'\n')[0].split(b'\t')
/linux/Documentation/sphinx/
H A Dmaintainers_include.py124 field, details = line.split(':', 1)
150 for separated in output.split('\n'):
165 for separated in field_content.split('\n'):
185 (path, tail) = os.path.split(path)

12345678910>>...22