Home
last modified time | relevance | path

Searched full:split (Results 1 – 25 of 1782) sorted by relevance

12345678910>>...72

/kernel/linux/linux-5.10/drivers/media/dvb-core/
Ddvb_ringbuffer.c136 size_t split; in dvb_ringbuffer_read_user() local
138 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read_user()
139 if (split > 0) { in dvb_ringbuffer_read_user()
140 if (copy_to_user(buf, rbuf->data+rbuf->pread, split)) in dvb_ringbuffer_read_user()
142 buf += split; in dvb_ringbuffer_read_user()
143 todo -= split; in dvb_ringbuffer_read_user()
162 size_t split; in dvb_ringbuffer_read() local
164 split = (rbuf->pread + len > rbuf->size) ? rbuf->size - rbuf->pread : 0; in dvb_ringbuffer_read()
165 if (split > 0) { in dvb_ringbuffer_read()
166 memcpy(buf, rbuf->data+rbuf->pread, split); in dvb_ringbuffer_read()
[all …]
/kernel/linux/linux-5.10/tools/testing/selftests/net/
Ddevlink_port_split.py12 # Test port split configuration using devlink-port lanes attribute.
15 # First, check that all the ports with 1 lane fail to split.
16 # Second, check that all the ports with more than 1 lane can be split
17 # to all valid configurations (e.g., split to 2, split to 4 etc.)
91 Get the $port split ability.
92 Return: split ability, true or false.
103 def split(k, port, should_fail=False): function
105 Split $port into $k ports.
106 If should_fail == True, the split should fail. Otherwise, should pass.
108 If the $port wasn't split, the array will be empty.
[all …]
/kernel/linux/linux-5.10/Documentation/vm/
Dsplit_page_table_lock.rst4 Split page table lock
10 scalability, split page table lock was introduced.
12 With split page table lock we have separate per-table lock to serialize
13 access to the table. At the moment we use split lock for PTE and PMD
33 Split page table lock for PTE tables is enabled compile-time if
35 If split lock is disabled, all tables guaded by mm->page_table_lock.
37 Split page table lock for PMD tables is enabled, if it's enabled for PTE
40 Hugetlb and split page table lock
43 Hugetlb can support several page sizes. We use split lock only for PMD
49 takes pmd split lock for PMD_SIZE page, mm->page_table_lock
[all …]
Dtranshuge.rst16 if necessary, split a transparent hugepage. Therefore these components
48 prevent the page from being split by anyone.
58 follow_page, so that it will split the hugepages before returning
74 that you can't handle natively in your code, you can split it by
144 For anonymous pages, we set PG_double_map when a PMD of the page is split
156 requests to split pinned huge pages: it expects page count to be equal to
172 clear where references should go after split: it will stay on the head page.
175 pmd can be split at any point and never fails.
/kernel/linux/linux-5.10/lib/
Dsg_split.c81 struct sg_splitter *split; in sg_split_phys() local
83 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_phys()
84 in_sg = split->in_sg0; in sg_split_phys()
85 out_sg = split->out_sg; in sg_split_phys()
86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys()
89 out_sg->offset += split->skip_sg0; in sg_split_phys()
90 out_sg->length -= split->skip_sg0; in sg_split_phys()
98 out_sg[-1].length = split->length_last_sg; in sg_split_phys()
107 struct sg_splitter *split; in sg_split_mapped() local
109 for (i = 0, split = splitters; i < nb_splits; i++, split++) { in sg_split_mapped()
[all …]
/kernel/linux/linux-5.10/fs/jfs/
Djfs_xtree.c96 uint split; member
109 struct xtsplit * split, struct btstack * btstack);
111 static int xtSplitPage(tid_t tid, struct inode *ip, struct xtsplit * split,
115 struct xtsplit * split, struct metapage ** rmpp);
238 int nsplit = 0; /* number of pages to split */ in xtSearch()
335 /* compute number of pages to split */ in xtSearch()
382 /* compute number of pages to split */ in xtSearch()
440 /* compute number of pages to split */ in xtSearch()
482 /* update number of pages to split */ in xtSearch()
535 struct xtsplit split; /* split information */ in xtInsert() local
[all …]
Djfs_dtree.c32 * it is extended rather than split (its size is doubled),
33 * until its length becoms 4 KBytes, from then the extent is split
63 * (e.g., if split occurs <abc> and <aBd>, <ABD> trather than <aB>
64 * should be made the router key for the split)
101 /* dtree split parameter */
140 struct dtsplit * split, struct btstack * btstack);
142 static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split,
146 struct dtsplit * split, struct btstack * btstack);
149 struct dtsplit * split, struct metapage ** rmpp);
602 /* init level count for max pages to split */ in dtSearch()
[all …]
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/
Dsubcore.c28 * Split/unsplit procedure:
30 * A core can be in one of three states, unsplit, 2-way split, and 4-way split.
37 * 2-way split | 2
38 * 4-way split | 4
40 * The core is split along thread boundaries, the mapping between subcores and
50 * 2-way split:
57 * 4-way split:
68 * It is not possible to transition between either of the split states, the
72 * | | <----> | 2-way split |
76 * | | <----> | 4-way split |
[all …]
/kernel/linux/linux-5.10/drivers/virtio/
Dvirtio_ring.c117 /* Available for split ring */
137 } split; member
364 * Split ring specific functions - *_split().
451 WARN_ON_ONCE(total_sg > vq->split.vring.num && !vq->indirect); in virtqueue_add_split()
462 desc = vq->split.vring.desc; in virtqueue_add_split()
518 vq->split.vring.desc[head].flags = cpu_to_virtio16(_vq->vdev, in virtqueue_add_split()
520 vq->split.vring.desc[head].addr = cpu_to_virtio64(_vq->vdev, in virtqueue_add_split()
523 vq->split.vring.desc[head].len = cpu_to_virtio32(_vq->vdev, in virtqueue_add_split()
533 vq->split.vring.desc[head].next); in virtqueue_add_split()
538 vq->split.desc_state[head].data = data; in virtqueue_add_split()
[all …]
/kernel/linux/linux-5.10/fs/btrfs/tests/
Dextent-buffer-tests.c72 * space in this leaf to split the item without having to split the in test_btrfs_split_item()
77 test_err("split item failed %d", ret); in test_btrfs_split_item()
95 test_err("invalid len in the first split"); in test_btrfs_split_item()
104 "data in the buffer doesn't match what it should in the first split have='%.*s' want '%s'", in test_btrfs_split_item()
120 test_err("invalid len in the second split"); in test_btrfs_split_item()
129 "data in the buffer doesn't match what it should in the second split"); in test_btrfs_split_item()
138 test_err("second split item failed %d", ret); in test_btrfs_split_item()
152 test_err("invalid len in the first split"); in test_btrfs_split_item()
161 "data in the buffer doesn't match what it should in the third split"); in test_btrfs_split_item()
176 test_err("invalid len in the second split"); in test_btrfs_split_item()
[all …]
/kernel/liteos_m/tools/
Dmem_analysis.py204 function_name = line.split(".text.")[1]
206 function_name = line.split(".rodata.")[1]
208 function_name = line.split(".bss.")[1]
210 function_name = line.split(".sram.text.")[1]
212 function_name = line.split(".data.")[1]
230 if len(function_name.split()) > 1:
231 strs = function_name.split()
241 char_temp = strs[3].split('(')
243 lib_name = char_temp[0].split('/')[-1]
254 if len(line.strip().split()) != 3:
[all …]
/kernel/linux/linux-5.10/security/apparmor/
Dlib.c29 * aa_split_fqname - split a fqname into a profile and namespace name
35 * Split a namespace name from a profile name (see policy.c for naming
48 char *split = strchr(&name[1], ':'); in aa_split_fqname() local
50 if (split) { in aa_split_fqname()
52 *split++ = 0; in aa_split_fqname()
53 if (strncmp(split, "//", 2) == 0) in aa_split_fqname()
54 split += 2; in aa_split_fqname()
55 name = skip_spaces(split); in aa_split_fqname()
96 char *split = strnchr(&name[1], end - &name[1], ':'); in aa_splitn_fqname() local
100 if (split) { in aa_splitn_fqname()
[all …]
/kernel/linux/linux-5.10/arch/x86/kernel/cpu/
Dcacheinfo.c137 } split; member
146 } split; member
153 } split; member
285 eax->split.is_self_initializing = 1; in amd_cpuid4()
286 eax->split.type = types[leaf]; in amd_cpuid4()
287 eax->split.level = levels[leaf]; in amd_cpuid4()
288 eax->split.num_threads_sharing = 0; in amd_cpuid4()
289 eax->split.num_cores_on_die = __this_cpu_read(cpu_info.x86_max_cores) - 1; in amd_cpuid4()
293 eax->split.is_fully_associative = 1; in amd_cpuid4()
294 ebx->split.coherency_line_size = line_size - 1; in amd_cpuid4()
[all …]
/kernel/liteos_a/tools/scripts/parse_exc/
Dparse_excinfo.py64 strlist = line.split(' ')
78 strlist = lines.split()
84 ret = ret.split('\n')
100 strlist = lines.split()
103 ret = ret.split('\n')
126 strlist = lines.split()
135 ret = ret.split('\n')
151 strlist = lines.split()
157 res = res.split('\n')
181 strlist = lines.split()
[all …]
/kernel/linux/linux-5.10/block/
Dblk-merge.c179 * bvec_split_segs - verify whether or not a bvec should be split in the middle
192 * big to fit in a single segment and hence that it has to be split in the
223 /* tell the caller to split the bvec if it is too big to fit */ in bvec_split_segs()
228 * blk_bio_segment_split - split a bio in two bios
230 * @bio: [in] bio to be split
244 * split bio has finished.
263 goto split; in blk_bio_segment_split()
272 goto split; in blk_bio_segment_split()
281 split: in blk_bio_segment_split()
287 * __blk_queue_split - split a bio and submit the second half
[all …]
/kernel/linux/linux-5.10/tools/perf/pmu-events/arch/x86/goldmont/
Dmemory.json5 …"PublicDescription": "Counts when a memory load of a uop spans a page boundary (a split) is retire…
11 "BriefDescription": "Load uops that split a page (Precise event capable)"
16 …"PublicDescription": "Counts when a memory store of a uop spans a page boundary (a split) is retir…
22 "BriefDescription": "Store uops that split a page (Precise event capable)"
/kernel/linux/linux-5.10/arch/nds32/
DKconfig.cpu168 Depending on the selected kernel/user memory split, minimum
191 prompt "Memory split"
195 Select the desired split between kernel and user memory.
201 bool "3G/1G user/kernel split"
203 bool "3G/1G user/kernel split (for full 1G low memory)"
205 bool "2G/2G user/kernel split"
207 bool "1G/3G user/kernel split"
/kernel/linux/linux-5.10/tools/perf/pmu-events/arch/x86/goldmontplus/
Dmemory.json5 …"PublicDescription": "Counts when a memory load of a uop spans a page boundary (a split) is retire…
12 "BriefDescription": "Load uops that split a page (Precise event capable)"
17 …"PublicDescription": "Counts when a memory store of a uop spans a page boundary (a split) is retir…
24 "BriefDescription": "Store uops that split a page (Precise event capable)"
/kernel/linux/linux-5.10/drivers/crypto/caam/
Dkey_gen.h10 * split_key_len - Compute MDHA split key length for a given algorithm
14 * Return: MDHA split key length
28 * split_key_pad_len - Compute MDHA split key pad length for a given algorithm
32 * Return: MDHA split key pad length
Dkey_gen.c31 get a split ipad/opad key
33 Split key generation-----------------------------------------------
59 dev_dbg(jrdev, "split keylen %d split keylen padded %d\n", in gen_split_key()
97 * FIFO_STORE with the explicit split-key content store in gen_split_key()
/kernel/linux/linux-5.10/drivers/gpu/drm/arm/display/komeda/
Dkomeda_format_caps.c69 /* SPLIT + SPARSE + YTR RGB only */
70 /* split mode is only allowed for sparse mode */
76 /* TILED + SC + (SPLIT+SPARSE | SPARSE) + (YTR) */
84 /* SPLIT + SPARSE + (YTR) */
85 /* split mode is only allowed for sparse mode */
87 /* TILED + SC + (SPLIT+SPARSE | SPARSE) + YTR */
/kernel/linux/linux-5.10/arch/x86/oprofile/
Dop_model_ppro.c86 if (!(eax.split.version_id == 0 && in ppro_setup_ctrs()
90 if (counter_width < eax.split.bit_width) in ppro_setup_ctrs()
91 counter_width = eax.split.bit_width; in ppro_setup_ctrs()
215 if (eax.split.version_id == 0 && boot_cpu_data.x86 == 6 && in arch_perfmon_setup_counters()
217 eax.split.version_id = 2; in arch_perfmon_setup_counters()
218 eax.split.num_counters = 2; in arch_perfmon_setup_counters()
219 eax.split.bit_width = 40; in arch_perfmon_setup_counters()
222 num_counters = min((int)eax.split.num_counters, OP_MAX_COUNTER); in arch_perfmon_setup_counters()
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/
Dpmu_intel.c351 pmu->version = eax.split.version_id; in intel_pmu_refresh()
357 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh()
359 eax.split.bit_width = min_t(int, eax.split.bit_width, x86_pmu.bit_width_gp); in intel_pmu_refresh()
360 pmu->counter_bitmask[KVM_PMC_GP] = ((u64)1 << eax.split.bit_width) - 1; in intel_pmu_refresh()
361 eax.split.mask_length = min_t(int, eax.split.mask_length, x86_pmu.events_mask_len); in intel_pmu_refresh()
363 ((1ull << eax.split.mask_length) - 1); in intel_pmu_refresh()
369 min_t(int, edx.split.num_counters_fixed, in intel_pmu_refresh()
371 edx.split.bit_width_fixed = min_t(int, in intel_pmu_refresh()
372 edx.split.bit_width_fixed, x86_pmu.bit_width_fixed); in intel_pmu_refresh()
374 ((u64)1 << edx.split.bit_width_fixed) - 1; in intel_pmu_refresh()
/kernel/linux/linux-5.10/fs/btrfs/
Dfile.c515 * [start, end]. Existing extents are split as required.
521 struct extent_map *split = NULL; in btrfs_drop_extent_cache() local
541 if (!split) in btrfs_drop_extent_cache()
542 split = alloc_extent_map(); in btrfs_drop_extent_cache()
545 if (!split || !split2) in btrfs_drop_extent_cache()
577 split->start = em->start; in btrfs_drop_extent_cache()
578 split->len = start - em->start; in btrfs_drop_extent_cache()
581 split->orig_start = em->orig_start; in btrfs_drop_extent_cache()
582 split->block_start = em->block_start; in btrfs_drop_extent_cache()
585 split->block_len = em->block_len; in btrfs_drop_extent_cache()
[all …]
/kernel/linux/linux-5.10/net/sched/
Dsch_cbq.c95 struct cbq_class *split; /* Ptr to split node */ member
199 * so that it resolves to split nodes. Then packets are classified
201 * to the split node.
914 struct cbq_class *split = cl->split; in cbq_sync_defmap() local
918 if (split == NULL) in cbq_sync_defmap()
922 if (split->defaults[i] == cl && !(cl->defmap & (1<<i))) in cbq_sync_defmap()
923 split->defaults[i] = NULL; in cbq_sync_defmap()
927 int level = split->level; in cbq_sync_defmap()
929 if (split->defaults[i]) in cbq_sync_defmap()
937 if (c->split == split && c->level < level && in cbq_sync_defmap()
[all …]

12345678910>>...72