Home
last modified time | relevance | path

Searched full:chunks (Results 1 – 25 of 1111) sorted by relevance

12345678910>>...45

/kernel/linux/linux-4.19/sound/firewire/motu/
Dmotu-protocol-v3.c196 // Packets to v3 units include 2 chunks for phone 1/2, except in calculate_fixed_part()
208 * At least, packets have two data chunks for S/PDIF on coaxial in calculate_fixed_part()
215 * Fixed part consists of PCM chunks multiple of 4, with msg chunks. As in calculate_fixed_part()
216 * a result, this part can includes empty data chunks. in calculate_fixed_part()
236 * Additional two data chunks for S/PDIF on optical in calculate_differed_part()
237 * interface A. This includes empty data chunks. in calculate_differed_part()
243 * Additional data chunks for ADAT on optical interface in calculate_differed_part()
254 * Additional two data chunks for S/PDIF on optical in calculate_differed_part()
255 * interface B. This includes empty data chunks. in calculate_differed_part()
261 * Additional data chunks for ADAT on optical interface in calculate_differed_part()
/kernel/linux/linux-5.10/net/sctp/
Dchunk.c43 INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init()
65 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
81 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy()
140 * down any such message into smaller chunks. Opportunistically, fragment
141 * the chunks down to the current MTU constraints. We may get refragmented
182 /* If the peer requested that we authenticate DATA chunks in sctp_datamsg_from_user()
183 * we need to account for bundling of the AUTH chunks along with in sctp_datamsg_from_user()
235 /* Create chunks for all DATA chunks. */ in sctp_datamsg_from_user()
280 list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user()
289 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_from_user()
Doutqueue.c11 * bundling and queueing of outgoing SCTP chunks.
204 /* Free the outqueue structure and any related pending chunks.
212 /* Throw away unacknowledged chunks. */ in __sctp_outq_teardown()
224 /* Throw away chunks that have been gap ACKed. */ in __sctp_outq_teardown()
233 /* Throw away any chunks in the retransmit queue. */ in __sctp_outq_teardown()
242 /* Throw away any chunks that are in the abandoned queue. */ in __sctp_outq_teardown()
251 /* Throw away any leftover data chunks. */ in __sctp_outq_teardown()
260 /* Throw away any leftover control chunks. */ in __sctp_outq_teardown()
273 /* Free the outqueue structure and any related pending chunks. */
276 /* Throw away leftover chunks. */ in sctp_outq_free()
[all …]
Dauth.c175 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO
186 struct sctp_chunks_param *chunks, in sctp_auth_make_key_vector() argument
197 if (chunks) in sctp_auth_make_key_vector()
198 chunks_len = ntohs(chunks->param_hdr.length); in sctp_auth_make_key_vector()
209 if (chunks) { in sctp_auth_make_key_vector()
210 memcpy(new->data + offset, chunks, chunks_len); in sctp_auth_make_key_vector()
301 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO in sctp_auth_asoc_create_secret()
651 * chunks MUST NOT be listed in the CHUNKS parameter. However, if in __sctp_auth_cid()
652 * a CHUNKS parameter is received then the types for INIT, INIT-ACK, in __sctp_auth_cid()
653 * SHUTDOWN-COMPLETE and AUTH chunks MUST be ignored. in __sctp_auth_cid()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/
Dradeon_cs.c284 /* get chunks */ in radeon_cs_parser_init()
296 chunk_array_ptr = (uint64_t *)(unsigned long)(cs->chunks); in radeon_cs_parser_init()
303 p->chunks = kcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL); in radeon_cs_parser_init()
304 if (p->chunks == NULL) { in radeon_cs_parser_init()
317 p->chunks[i].length_dw = user_chunk.length_dw; in radeon_cs_parser_init()
319 p->chunk_relocs = &p->chunks[i]; in radeon_cs_parser_init()
322 p->chunk_ib = &p->chunks[i]; in radeon_cs_parser_init()
324 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init()
328 p->chunk_const_ib = &p->chunks[i]; in radeon_cs_parser_init()
330 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init()
[all …]
/kernel/linux/linux-4.19/drivers/gpu/drm/radeon/
Dradeon_cs.c278 /* get chunks */ in radeon_cs_parser_init()
290 chunk_array_ptr = (uint64_t *)(unsigned long)(cs->chunks); in radeon_cs_parser_init()
297 p->chunks = kcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL); in radeon_cs_parser_init()
298 if (p->chunks == NULL) { in radeon_cs_parser_init()
311 p->chunks[i].length_dw = user_chunk.length_dw; in radeon_cs_parser_init()
313 p->chunk_relocs = &p->chunks[i]; in radeon_cs_parser_init()
316 p->chunk_ib = &p->chunks[i]; in radeon_cs_parser_init()
318 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init()
322 p->chunk_const_ib = &p->chunks[i]; in radeon_cs_parser_init()
324 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init()
[all …]
/kernel/linux/linux-5.10/arch/mips/ar7/
Dprom.c102 Name=Value pair in 2 chunks (len is the number of chunks)
149 struct psp_env_chunk *chunks = (struct psp_env_chunk *)psp_env_data; in parse_psp_env() local
151 memcpy_fromio(chunks, psp_env_base, PSP_ENV_SIZE); in parse_psp_env()
156 if ((chunks[i].num == 0xff) || ((i + chunks[i].len) > n)) in parse_psp_env()
158 value = chunks[i].data; in parse_psp_env()
159 if (chunks[i].num) { in parse_psp_env()
160 name = lookup_psp_var_map(chunks[i].num); in parse_psp_env()
167 i += chunks[i].len; in parse_psp_env()
/kernel/linux/linux-4.19/net/sctp/
Dchunk.c58 INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init()
80 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
100 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy()
159 * down any such message into smaller chunks. Opportunistically, fragment
160 * the chunks down to the current MTU constraints. We may get refragmented
201 /* If the the peer requested that we authenticate DATA chunks in sctp_datamsg_from_user()
202 * we need to account for bundling of the AUTH chunks along with in sctp_datamsg_from_user()
254 /* Create chunks for all DATA chunks. */ in sctp_datamsg_from_user()
299 list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user()
308 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_from_user()
Doutqueue.c10 * bundling and queueing of outgoing SCTP chunks.
219 /* Free the outqueue structure and any related pending chunks.
227 /* Throw away unacknowledged chunks. */ in __sctp_outq_teardown()
239 /* Throw away chunks that have been gap ACKed. */ in __sctp_outq_teardown()
248 /* Throw away any chunks in the retransmit queue. */ in __sctp_outq_teardown()
257 /* Throw away any chunks that are in the abandoned queue. */ in __sctp_outq_teardown()
266 /* Throw away any leftover data chunks. */ in __sctp_outq_teardown()
275 /* Throw away any leftover control chunks. */ in __sctp_outq_teardown()
288 /* Free the outqueue structure and any related pending chunks. */
291 /* Throw away leftover chunks. */ in sctp_outq_free()
[all …]
/kernel/linux/linux-4.19/arch/mips/ar7/
Dprom.c114 Name=Value pair in 2 chunks (len is the number of chunks)
161 struct psp_env_chunk *chunks = (struct psp_env_chunk *)psp_env_data; in parse_psp_env() local
163 memcpy_fromio(chunks, psp_env_base, PSP_ENV_SIZE); in parse_psp_env()
168 if ((chunks[i].num == 0xff) || ((i + chunks[i].len) > n)) in parse_psp_env()
170 value = chunks[i].data; in parse_psp_env()
171 if (chunks[i].num) { in parse_psp_env()
172 name = lookup_psp_var_map(chunks[i].num); in parse_psp_env()
179 i += chunks[i].len; in parse_psp_env()
/kernel/linux/linux-5.10/drivers/net/ethernet/netronome/nfp/nfpcore/
Dnfp_nsp.c504 } *chunks; in nfp_nsp_command_buf_dma_sg() local
516 chunks = kzalloc(array_size(sizeof(*chunks), nseg), GFP_KERNEL); in nfp_nsp_command_buf_dma_sg()
517 if (!chunks) in nfp_nsp_command_buf_dma_sg()
525 chunks[i].chunk = kmalloc(chunk_size, in nfp_nsp_command_buf_dma_sg()
527 if (!chunks[i].chunk) in nfp_nsp_command_buf_dma_sg()
530 chunks[i].len = min_t(u64, chunk_size, max_size - off); in nfp_nsp_command_buf_dma_sg()
535 memcpy(chunks[i].chunk, arg->in_buf + off, coff); in nfp_nsp_command_buf_dma_sg()
537 memset(chunks[i].chunk + coff, 0, chunk_size - coff); in nfp_nsp_command_buf_dma_sg()
539 off += chunks[i].len; in nfp_nsp_command_buf_dma_sg()
547 addr = dma_map_single(dev, chunks[i].chunk, chunks[i].len, in nfp_nsp_command_buf_dma_sg()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/ulp/rtrs/
DREADME28 session. A session is associated with a set of memory chunks reserved on the
36 chunks reserved for him on the server side. Their number, size and addresses
45 which of the memory chunks has been accessed and at which offset the message
80 the server (number of memory chunks which are going to be allocated for that
122 1. When processing a write request client selects one of the memory chunks
139 1. When processing a write request client selects one of the memory chunks
144 using the IMM field, Server invalidate rkey associated to the memory chunks
162 1. When processing a read request client selects one of the memory chunks
181 1. When processing a read request client selects one of the memory chunks
186 Server invalidate rkey associated to the memory chunks first, when it finishes,
/kernel/linux/linux-4.19/mm/
Dzbud.c30 * zbud pages are divided into "chunks". The size of the chunks is fixed at
32 * into chunks allows organizing unbuddied zbud pages into a manageable number
33 * of unbuddied lists according to the number of free chunks available in the
63 * allocation granularity will be in chunks of size PAGE_SIZE/64. As one chunk
65 * 63 which shows the max number of free chunks in zbud page, also there will be
111 * @first_chunks: the size of the first buddy in chunks, 0 if free
112 * @last_chunks: the size of the last buddy in chunks, 0 if free
228 /* Converts an allocation size in bytes to size in zbud chunks */
284 /* Returns the number of free chunks in a zbud page */
356 int chunks, i, freechunks; in zbud_alloc() local
[all …]
Dz3fold.c17 * As in zbud, pages are divided into "chunks". The size of the chunks is
54 * struct z3fold_header - z3fold page metadata occupying first chunks of each
63 * @first_chunks: the size of the first buddy in chunks, 0 if free
64 * @middle_chunks: the size of the middle buddy in chunks, 0 if free
65 * @last_chunks: the size of the last buddy in chunks, 0 if free
86 * allocation granularity will be in chunks of size PAGE_SIZE/64. Some chunks
89 * which shows the max number of free chunks in z3fold page, also there will
156 /* Converts an allocation size in bytes to size in z3fold chunks */
329 * Returns the number of free chunks in a z3fold page.
338 * of chunks occupied by the first and the last objects. in num_free_chunks()
[all …]
/kernel/linux/linux-5.10/mm/
Dzbud.c31 * zbud pages are divided into "chunks". The size of the chunks is fixed at
33 * into chunks allows organizing unbuddied zbud pages into a manageable number
34 * of unbuddied lists according to the number of free chunks available in the
64 * allocation granularity will be in chunks of size PAGE_SIZE/64. As one chunk
66 * 63 which shows the max number of free chunks in zbud page, also there will be
112 * @first_chunks: the size of the first buddy in chunks, 0 if free
113 * @last_chunks: the size of the last buddy in chunks, 0 if free
230 /* Converts an allocation size in bytes to size in zbud chunks */
286 /* Returns the number of free chunks in a zbud page */
358 int chunks, i, freechunks; in zbud_alloc() local
[all …]
/kernel/linux/linux-5.10/tools/testing/selftests/drivers/net/mlxsw/spectrum/
Ddevlink_lib_spectrum.sh15 KVDL_CHILDREN="singles chunks large_chunks"
92 devlink_resource_size_set 32000 kvd linear chunks
101 devlink_resource_size_set 32000 kvd linear chunks
110 devlink_resource_size_set 49152 kvd linear chunks
/kernel/linux/linux-4.19/tools/testing/selftests/drivers/net/mlxsw/spectrum/
Ddevlink_lib_spectrum.sh15 KVDL_CHILDREN="singles chunks large_chunks"
92 devlink_resource_size_set 32000 kvd linear chunks
101 devlink_resource_size_set 32000 kvd linear chunks
110 devlink_resource_size_set 49152 kvd linear chunks
/kernel/linux/linux-4.19/arch/x86/kernel/cpu/
Dintel_rdt_monitor.c230 u64 shift = 64 - MBM_CNTR_WIDTH, chunks; in mbm_overflow_count() local
232 chunks = (cur_msr << shift) - (prev_msr << shift); in mbm_overflow_count()
233 return chunks >>= shift; in mbm_overflow_count()
239 u64 chunks, tval; in __mon_event_count() local
270 chunks = mbm_overflow_count(m->prev_msr, tval); in __mon_event_count()
271 m->chunks += chunks; in __mon_event_count()
274 rr->val += m->chunks; in __mon_event_count()
286 u64 tval, cur_bw, chunks; in mbm_bw_count() local
292 chunks = mbm_overflow_count(m->prev_bw_msr, tval); in mbm_bw_count()
293 m->chunks_bw += chunks; in mbm_bw_count()
[all …]
/kernel/linux/linux-4.19/include/net/sctp/
Dstructs.h374 struct sctp_chunks_param *chunks; member
531 /* Chunks waiting to be submitted to lower layer. */
532 struct list_head chunks; member
575 /* This field is used by chunks that hold fragmented data.
652 * spec violates the principle premis that all chunks are processed
712 /* This structure holds lists of chunks as we are assembling for
721 /* This contains the payload chunks. */
726 /* This is the total size of all chunks INCLUDING padding. */
791 * chunks sent to this address is currently being
956 /* This is the list of transports that have chunks to send. */
[all …]
/kernel/linux/linux-5.10/include/net/sctp/
Dstructs.h368 struct sctp_chunks_param *chunks; member
515 /* Chunks waiting to be submitted to lower layer. */
516 struct list_head chunks; member
559 /* This field is used by chunks that hold fragmented data.
636 * spec violates the principle premis that all chunks are processed
696 /* This structure holds lists of chunks as we are assembling for
705 /* This contains the payload chunks. */
710 /* This is the total size of all chunks INCLUDING padding. */
775 * chunks sent to this address is currently being
942 /* This is the list of transports that have chunks to send. */
[all …]
/kernel/linux/linux-5.10/fs/ocfs2/
Dquota.h40 struct list_head rc_list; /* List of chunks */
46 struct list_head r_list[OCFS2_MAXQUOTAS]; /* List of chunks to recover */
53 unsigned int dqi_chunks; /* Number of chunks in local quota file */
56 struct list_head dqi_chunk; /* List of chunks */
78 struct list_head qc_chunk; /* List of quotafile chunks */
/kernel/linux/linux-4.19/fs/ocfs2/
Dquota.h40 struct list_head rc_list; /* List of chunks */
46 struct list_head r_list[OCFS2_MAXQUOTAS]; /* List of chunks to recover */
53 unsigned int dqi_chunks; /* Number of chunks in local quota file */
56 struct list_head dqi_chunk; /* List of chunks */
78 struct list_head qc_chunk; /* List of quotafile chunks */
/kernel/linux/linux-5.10/drivers/infiniband/hw/usnic/
Dusnic_vnic.c45 struct usnic_vnic_res_chunk chunks[USNIC_VNIC_RES_TYPE_MAX]; member
118 for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) { in usnic_vnic_dump()
119 chunk = &vnic->chunks[i]; in usnic_vnic_dump()
223 return vnic->chunks[type].cnt; in usnic_vnic_res_cnt()
229 return vnic->chunks[type].free_cnt; in usnic_vnic_res_free_cnt()
255 src = &vnic->chunks[type]; in usnic_vnic_get_resources()
287 vnic->chunks[res->type].free_cnt++; in usnic_vnic_put_resources()
383 &vnic->chunks[res_type]); in usnic_vnic_discover_resources()
392 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_discover_resources()
428 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_release_resources()
/kernel/linux/linux-4.19/drivers/infiniband/hw/usnic/
Dusnic_vnic.c45 struct usnic_vnic_res_chunk chunks[USNIC_VNIC_RES_TYPE_MAX]; member
118 for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) { in usnic_vnic_dump()
119 chunk = &vnic->chunks[i]; in usnic_vnic_dump()
223 return vnic->chunks[type].cnt; in usnic_vnic_res_cnt()
229 return vnic->chunks[type].free_cnt; in usnic_vnic_res_free_cnt()
255 src = &vnic->chunks[type]; in usnic_vnic_get_resources()
287 vnic->chunks[res->type].free_cnt++; in usnic_vnic_put_resources()
383 &vnic->chunks[res_type]); in usnic_vnic_discover_resources()
392 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_discover_resources()
428 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_release_resources()
/kernel/linux/linux-5.10/kernel/
Daudit_tree.c17 struct list_head chunks; member
68 * tree.chunks anchors chunk.owners[].list hash_lock
101 INIT_LIST_HEAD(&tree->chunks); in alloc_tree()
435 list_add(&chunk->owners[0].list, &tree->chunks); in create_chunk()
507 list_add(&p->list, &tree->chunks); in tag_chunk()
565 * Remove tree from chunks. If 'tagged' is set, remove tree only from tagged
566 * chunks. The function expects tagged chunks are all at the beginning of the
567 * chunks list.
572 while (!list_empty(&victim->chunks)) { in prune_tree_chunks()
577 p = list_first_entry(&victim->chunks, struct node, list); in prune_tree_chunks()
[all …]

12345678910>>...45