Home
last modified time | relevance | path

Searched refs:split (Results 1 – 25 of 809) sorted by relevance

12345678910>>...33

/kernel/linux/linux-5.10/drivers/media/dvb-core/
Ddvb_ringbuffer.c136 size_t split; in dvb_ringbuffer_read_user() local
138 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user()
139 if (split > 0) { in dvb_ringbuffer_read_user()
140 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user()
142 buf += split; in dvb_ringbuffer_read_user()
143 todo -= split; in dvb_ringbuffer_read_user()
162 size_t split; in dvb_ringbuffer_read() local
164 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read()
165 if (split > 0) { in dvb_ringbuffer_read()
166 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read()
[all …]
/kernel/linux/linux-6.6/drivers/media/dvb-core/
Ddvb_ringbuffer.c136 size_t split; in dvb_ringbuffer_read_user() local
138 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user()
139 if (split > 0) { in dvb_ringbuffer_read_user()
140 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user()
142 buf += split; in dvb_ringbuffer_read_user()
143 todo -= split; in dvb_ringbuffer_read_user()
162 size_t split; in dvb_ringbuffer_read() local
164 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read()
165 if (split > 0) { in dvb_ringbuffer_read()
166 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read()
[all …]
/kernel/linux/linux-6.6/lib/
Dsg_split.c81 struct sg_splitter *split; in sg_split_phys() local
83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys()
84 in_sg = split->in_sg0; in sg_split_phys()
85 out_sg = split->out_sg; in sg_split_phys()
86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys()
89 out_sg->offset += split->skip_sg0; in sg_split_phys()
90 out_sg->length -= split->skip_sg0; in sg_split_phys()
98 out_sg[-1].length = split->length_last_sg; in sg_split_phys()
107 struct sg_splitter *split; in sg_split_mapped() local
109 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_mapped()
[all …]
/kernel/linux/linux-5.10/lib/
Dsg_split.c81 struct sg_splitter *split; in sg_split_phys() local
83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys()
84 in_sg = split->in_sg0; in sg_split_phys()
85 out_sg = split->out_sg; in sg_split_phys()
86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys()
89 out_sg->offset += split->skip_sg0; in sg_split_phys()
90 out_sg->length -= split->skip_sg0; in sg_split_phys()
98 out_sg[-1].length = split->length_last_sg; in sg_split_phys()
107 struct sg_splitter *split; in sg_split_mapped() local
109 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_mapped()
[all …]
/kernel/linux/linux-6.6/fs/btrfs/
Dextent_map.c710 struct extent_map *split; in btrfs_drop_extent_map_range() local
736 split = alloc_extent_map(); in btrfs_drop_extent_map_range()
787 if (!split) { in btrfs_drop_extent_map_range()
788 split = split2; in btrfs_drop_extent_map_range()
790 if (!split) in btrfs_drop_extent_map_range()
793 split->start = em->start; in btrfs_drop_extent_map_range()
794 split->len = start - em->start; in btrfs_drop_extent_map_range()
797 split->orig_start = em->orig_start; in btrfs_drop_extent_map_range()
798 split->block_start = em->block_start; in btrfs_drop_extent_map_range()
801 split->block_len = em->block_len; in btrfs_drop_extent_map_range()
[all …]
/kernel/linux/linux-6.6/tools/verification/dot2/
Dautomata.py54 line = dot_lines[cursor].split()
64 while self.__dot_lines[cursor].split()[0] != "{node":
70 while self.__dot_lines[cursor].split()[0] != "{node":
72 while self.__dot_lines[cursor].split()[0] == "{node":
87 while self.__dot_lines[cursor].split()[0] == "{node":
88 line = self.__dot_lines[cursor].split()
127 if self.__dot_lines[cursor].split()[1] == "->":
128 line = self.__dot_lines[cursor].split()
136 for i in event.split():
165 if self.__dot_lines[cursor].split()[1] == "->":
[all …]
/kernel/linux/linux-5.10/drivers/virtio/
Dvirtio_ring.c137 } split; member
451 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split()
462 desc = vq->split.vring.desc; in virtqueue_add_split()
518 vq->split.vring.desc[head].flags = cpu_to_virtio16(_vq->vdev, in virtqueue_add_split()
520 vq->split.vring.desc[head].addr = cpu_to_virtio64(_vq->vdev, in virtqueue_add_split()
523 vq->split.vring.desc[head].len = cpu_to_virtio32(_vq->vdev, in virtqueue_add_split()
533 vq->split.vring.desc[head].next); in virtqueue_add_split()
538 vq->split.desc_state[head].data = data; in virtqueue_add_split()
540 vq->split.desc_state[head].indir_desc = desc; in virtqueue_add_split()
542 vq->split.desc_state[head].indir_desc = ctx; in virtqueue_add_split()
[all …]
/kernel/linux/linux-6.6/fs/jfs/
Djfs_xtree.c96 uint split; member
109 struct xtsplit * split, struct btstack * btstack);
111 static int xtSplitPage(tid_t tid, struct inode *ip, struct xtsplit * split,
115 struct xtsplit * split, struct metapage ** rmpp);
524 struct xtsplit split; /* split information */ in xtInsert() local
586 split.mp = mp; in xtInsert()
587 split.index = index; in xtInsert()
588 split.flag = xflag; in xtInsert()
589 split.off = xoff; in xtInsert()
590 split.len = xlen; in xtInsert()
[all …]
/kernel/linux/linux-5.10/arch/x86/kernel/cpu/
Dcacheinfo.c137 } split; member
146 } split; member
153 } split; member
285 eax->split.is_self_initializing = 1; in amd_cpuid4()
286 eax->split.type = types[leaf]; in amd_cpuid4()
287 eax->split.level = levels[leaf]; in amd_cpuid4()
288 eax->split.num_threads_sharing = 0; in amd_cpuid4()
289 eax->split.num_cores_on_die = __this_cpu_read(cpu_info.x86_max_cores) - 1; in amd_cpuid4()
293 eax->split.is_fully_associative = 1; in amd_cpuid4()
294 ebx->split.coherency_line_size = line_size - 1; in amd_cpuid4()
[all …]
/kernel/linux/linux-6.6/security/apparmor/
Dlib.c68 char *split = strchr(&name[1], ':'); in aa_split_fqname() local
70 if (split) { in aa_split_fqname()
72 *split++ = 0; in aa_split_fqname()
73 if (strncmp(split, "//", 2) == 0) in aa_split_fqname()
74 split += 2; in aa_split_fqname()
75 name = skip_spaces(split); in aa_split_fqname()
117 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local
121 if (split) { in aa_splitn_fqname()
122 *ns_len = split - *ns_name; in aa_splitn_fqname()
125 split++; in aa_splitn_fqname()
[all …]
/kernel/linux/linux-5.10/fs/jfs/
Djfs_xtree.c96 uint split; member
109 struct xtsplit * split, struct btstack * btstack);
111 static int xtSplitPage(tid_t tid, struct inode *ip, struct xtsplit * split,
115 struct xtsplit * split, struct metapage ** rmpp);
535 struct xtsplit split; /* split information */ in xtInsert() local
597 split.mp = mp; in xtInsert()
598 split.index = index; in xtInsert()
599 split.flag = xflag; in xtInsert()
600 split.off = xoff; in xtInsert()
601 split.len = xlen; in xtInsert()
[all …]
/kernel/linux/linux-6.6/drivers/virtio/
Dvirtio_ring.c201 struct vring_virtqueue_split split; member
463 struct vring_desc_extra *extra = vq->split.desc_extra; in vring_unmap_one_split()
524 struct vring_desc_extra *extra = vring->split.desc_extra; in virtqueue_add_desc_split()
580 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split()
591 desc = vq->split.vring.desc; in virtqueue_add_split()
647 vq->split.desc_extra[prev & (vq->split.vring.num - 1)].flags &= in virtqueue_add_split()
662 virtqueue_add_desc_split(_vq, vq->split.vring.desc, in virtqueue_add_split()
674 vq->free_head = vq->split.desc_extra[head].next; in virtqueue_add_split()
679 vq->split.desc_state[head].data = data; in virtqueue_add_split()
681 vq->split.desc_state[head].indir_desc = desc; in virtqueue_add_split()
[all …]
/kernel/linux/linux-6.6/arch/x86/kernel/cpu/
Dcacheinfo.c152 } split; member
161 } split; member
168 } split; member
300 eax->split.is_self_initializing = 1; in amd_cpuid4()
301 eax->split.type = types[leaf]; in amd_cpuid4()
302 eax->split.level = levels[leaf]; in amd_cpuid4()
303 eax->split.num_threads_sharing = 0; in amd_cpuid4()
304 eax->split.num_cores_on_die = __this_cpu_read(cpu_info.x86_max_cores) - 1; in amd_cpuid4()
308 eax->split.is_fully_associative = 1; in amd_cpuid4()
309 ebx->split.coherency_line_size = line_size - 1; in amd_cpuid4()
[all …]
/kernel/liteos_m/tools/
Dmem_analysis.py204 function_name = line.split(".text.")[1]
206 function_name = line.split(".rodata.")[1]
208 function_name = line.split(".bss.")[1]
210 function_name = line.split(".sram.text.")[1]
212 function_name = line.split(".data.")[1]
230 if len(function_name.split()) > 1:
231 strs = function_name.split()
241 char_temp = strs[3].split('(')
243 lib_name = char_temp[0].split('/')[-1]
254 if len(line.strip().split()) != 3:
[all …]
/kernel/linux/linux-6.6/Documentation/mm/
Dsplit_page_table_lock.rst8 scalability, split page table lock was introduced.
10 With split page table lock we have separate per-table lock to serialize
11 access to the table. At the moment we use split lock for PTE and PMD
38 If split lock is disabled, all tables are guarded by mm->page_table_lock.
43 Hugetlb and split page table lock
46 Hugetlb can support several page sizes. We use split lock only for PMD
52 takes pmd split lock for PMD_SIZE page, mm->page_table_lock
57 Support of split page table lock by an architecture
60 There's no need in special enabling of PTE split page table lock: everything
68 PMD split lock only makes sense if you have more than two page table
[all …]
/kernel/linux/linux-5.10/Documentation/vm/
Dsplit_page_table_lock.rst10 scalability, split page table lock was introduced.
12 With split page table lock we have separate per-table lock to serialize
13 access to the table. At the moment we use split lock for PTE and PMD
35 If split lock is disabled, all tables guaded by mm->page_table_lock.
40 Hugetlb and split page table lock
43 Hugetlb can support several page sizes. We use split lock only for PMD
49 takes pmd split lock for PMD_SIZE page, mm->page_table_lock
54 Support of split page table lock by an architecture
57 There's no need in special enabling of PTE split page table lock: everything
65 PMD split lock only makes sense if you have more than two page table
[all …]
/kernel/liteos_a/tools/scripts/parse_exc/
Dparse_excinfo.py64 strlist = line.split(' ')
78 strlist = lines.split()
84 ret = ret.split('\n')
100 strlist = lines.split()
103 ret = ret.split('\n')
126 strlist = lines.split()
135 ret = ret.split('\n')
151 strlist = lines.split()
157 res = res.split('\n')
181 strlist = lines.split()
[all …]
/kernel/linux/linux-5.10/security/apparmor/
Dlib.c48 char *split = strchr(&name[1], ':'); in aa_split_fqname() local
50 if (split) { in aa_split_fqname()
52 *split++ = 0; in aa_split_fqname()
53 if (strncmp(split, "//", 2) == 0) in aa_split_fqname()
54 split += 2; in aa_split_fqname()
55 name = skip_spaces(split); in aa_split_fqname()
96 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local
100 if (split) { in aa_splitn_fqname()
101 *ns_len = split - *ns_name; in aa_splitn_fqname()
104 split++; in aa_splitn_fqname()
[all …]
/kernel/linux/linux-5.10/arch/x86/oprofile/
Dop_model_ppro.c86 if (!(eax.split.version_id == 0 && in ppro_setup_ctrs()
90 if (counter_width < eax.split.bit_width) in ppro_setup_ctrs()
91 counter_width = eax.split.bit_width; in ppro_setup_ctrs()
215 if (eax.split.version_id == 0 && boot_cpu_data.x86 == 6 && in arch_perfmon_setup_counters()
217 eax.split.version_id = 2; in arch_perfmon_setup_counters()
218 eax.split.num_counters = 2; in arch_perfmon_setup_counters()
219 eax.split.bit_width = 40; in arch_perfmon_setup_counters()
222 num_counters = min((int)eax.split.num_counters, OP_MAX_COUNTER); in arch_perfmon_setup_counters()
/kernel/linux/linux-6.6/arch/x86/events/amd/
Dlbr.c46 } split; member
57 } split; member
178 if (!entry.to.split.valid && !entry.to.split.spec) in amd_pmu_lbr_read()
183 br[out].from = sign_ext_branch_ip(entry.from.split.ip); in amd_pmu_lbr_read()
184 br[out].to = sign_ext_branch_ip(entry.to.split.ip); in amd_pmu_lbr_read()
185 br[out].mispred = entry.from.split.mispredict; in amd_pmu_lbr_read()
204 idx = (entry.to.split.valid << 1) | entry.to.split.spec; in amd_pmu_lbr_read()
434 x86_pmu.lbr_nr = ebx.split.lbr_v2_stack_sz; in amd_pmu_lbr_init()
/kernel/linux/linux-6.6/drivers/thermal/intel/
Dintel_hfi.c64 } split; member
74 } split; member
336 info->index = edx.split.index; in init_hfi_cpu_index()
542 if (!edx.split.capabilities.split.performance) { in hfi_parse_features()
551 edx.split.capabilities.split.__reserved = 0; in hfi_parse_features()
552 nr_capabilities = hweight8(edx.split.capabilities.bits); in hfi_parse_features()
555 hfi_features.nr_table_pages = edx.split.table_pages + 1; in hfi_parse_features()
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/
Dpmu_intel.c351 pmu->version = eax.split.version_id; in intel_pmu_refresh()
357 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh()
359 eax.split.bit_width = min_t(int, eax.split.bit_width, x86_pmu.bit_width_gp); in intel_pmu_refresh()
360 pmu->counter_bitmask[KVM_PMC_GP] = ((u64)1 << eax.split.bit_width) - 1; in intel_pmu_refresh()
361 eax.split.mask_length = min_t(int, eax.split.mask_length, x86_pmu.events_mask_len); in intel_pmu_refresh()
363 ((1ull << eax.split.mask_length) - 1); in intel_pmu_refresh()
369 min_t(int, edx.split.num_counters_fixed, in intel_pmu_refresh()
371 edx.split.bit_width_fixed = min_t(int, in intel_pmu_refresh()
372 edx.split.bit_width_fixed, x86_pmu.bit_width_fixed); in intel_pmu_refresh()
374 ((u64)1 << edx.split.bit_width_fixed) - 1; in intel_pmu_refresh()
/kernel/linux/linux-6.6/lib/zstd/compress/
Dzstd_ldm.c291 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_fillHashTable() local
292 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_fillHashTable()
296 entry.offset = (U32)(split - base); in ZSTD_ldm_fillHashTable()
369 BYTE const* const split = ip + splits[n] - minMatchLength; in ZSTD_ldm_generateSequences_internal() local
370 U64 const xxhash = xxh64(split, minMatchLength, 0); in ZSTD_ldm_generateSequences_internal()
373 candidates[n].split = split; in ZSTD_ldm_generateSequences_internal()
384 BYTE const* const split = candidates[n].split; in ZSTD_ldm_generateSequences_internal() local
392 newEntry.offset = (U32)(split - base); in ZSTD_ldm_generateSequences_internal()
398 if (split < anchor) { in ZSTD_ldm_generateSequences_internal()
418 ZSTD_count_2segments(split, pMatch, iend, matchEnd, lowPrefixPtr); in ZSTD_ldm_generateSequences_internal()
[all …]
/kernel/linux/linux-6.6/Documentation/arch/x86/
Dbuslock.rst16 A split lock is any atomic operation whose operand crosses two cache lines.
20 A bus lock is acquired through either split locked access to writeback (WB)
29 mechanisms to detect split locks and bus locks.
31 #AC exception for split lock detection
34 Beginning with the Tremont Atom CPU split lock operations may raise an
35 Alignment Check (#AC) exception when a split lock operation is attempted.
51 |split_lock_detect=|#AC for split lock |#DB for bus lock |
60 | |split lock in parallel. | |
89 generating split lock and bus lock to block the hard real time code to
102 Disable checking for split lock and bus lock. This option can be useful if
/kernel/linux/linux-6.6/tools/hv/
Dvmbus_testing154 f_name = f_path.split("/")[-1]
169 .format(device.split("/")[5]))
172 .format(device.split("/")[5]))
209 .format(state_path.split("/")[5]))
213 .format(state_path.split("/")[5]))
238 interrupt.split("/")[5]))
241 message.split("/")[5]))
263 print("ALL testing now OFF for {}".format(device.split("/")[-1]))

12345678910>>...33