/external/libdrm/amdgpu/ |
D | amdgpu_cs.c | 181 struct drm_amdgpu_cs_chunk_data *chunk_data; in amdgpu_cs_submit_one() local 207 chunk_data = alloca(sizeof(struct drm_amdgpu_cs_chunk_data) * size); in amdgpu_cs_submit_one() 221 chunks[i].chunk_data = (uint64_t)(uintptr_t)&chunk_data[i]; in amdgpu_cs_submit_one() 225 chunk_data[i].ib_data._pad = 0; in amdgpu_cs_submit_one() 226 chunk_data[i].ib_data.va_start = ib->ib_mc_address; in amdgpu_cs_submit_one() 227 chunk_data[i].ib_data.ib_bytes = ib->size * 4; in amdgpu_cs_submit_one() 228 chunk_data[i].ib_data.ip_type = ibs_request->ip_type; in amdgpu_cs_submit_one() 229 chunk_data[i].ib_data.ip_instance = ibs_request->ip_instance; in amdgpu_cs_submit_one() 230 chunk_data[i].ib_data.ring = ibs_request->ring; in amdgpu_cs_submit_one() 231 chunk_data[i].ib_data.flags = ib->flags; in amdgpu_cs_submit_one() [all …]
|
/external/u-boot/fs/yaffs2/ |
D | yaffs_yaffs1.c | 38 u8 *chunk_data; in yaffs1_scan() local 44 chunk_data = yaffs_get_temp_buffer(dev); in yaffs1_scan() 183 chunk_data, NULL); in yaffs1_scan() 185 oh = (struct yaffs_obj_hdr *)chunk_data; in yaffs1_scan() 411 yaffs_release_temp_buffer(dev, chunk_data); in yaffs1_scan()
|
D | yaffs_yaffs2.c | 937 u8 *chunk_data, in yaffs2_scan_chunk() argument 1116 yaffs_rd_chunk_tags_nand(dev, chunk, chunk_data, NULL); in yaffs2_scan_chunk() 1118 oh = (struct yaffs_obj_hdr *)chunk_data; in yaffs2_scan_chunk() 1352 u8 *chunk_data; in yaffs2_scan_backwards() local 1383 chunk_data = yaffs_get_temp_buffer(dev); in yaffs2_scan_backwards() 1484 &found_chunks, chunk_data, in yaffs2_scan_backwards() 1518 yaffs_release_temp_buffer(dev, chunk_data); in yaffs2_scan_backwards()
|
/external/webp/src/mux/ |
D | muxread.c | 57 WebPData chunk_data; in ChunkVerifyAndAssign() local 71 chunk_data.bytes = data + CHUNK_HEADER_SIZE; in ChunkVerifyAndAssign() 72 chunk_data.size = chunk_size; in ChunkVerifyAndAssign() 73 return ChunkAssignData(chunk, &chunk_data, copy_data, GetLE32(data + 0)); in ChunkVerifyAndAssign() 426 WebPData* chunk_data) { in WebPMuxGetChunk() argument 428 if (mux == NULL || fourcc == NULL || chunk_data == NULL) { in WebPMuxGetChunk() 435 return MuxGet(mux, idx, 1, chunk_data); in WebPMuxGetChunk() 440 *chunk_data = chunk->data_; in WebPMuxGetChunk()
|
D | muxedit.c | 179 const WebPData* chunk_data, int copy_data) { in WebPMuxSetChunk() argument 182 if (mux == NULL || fourcc == NULL || chunk_data == NULL || in WebPMuxSetChunk() 183 chunk_data->bytes == NULL || chunk_data->size > MAX_CHUNK_PAYLOAD) { in WebPMuxSetChunk() 193 return MuxSet(mux, tag, chunk_data, copy_data); in WebPMuxSetChunk()
|
/external/mesa3d/src/amd/vulkan/winsys/amdgpu/ |
D | radv_amdgpu_cs.c | 1270 chunk->chunk_data = (uint64_t)(uintptr_t)syncobj; in radv_amdgpu_cs_alloc_syncobj_chunk() 1283 struct drm_amdgpu_cs_chunk_data *chunk_data; in radv_amdgpu_cs_submit() local 1296 chunk_data = alloca(sizeof(struct drm_amdgpu_cs_chunk_data) * size); in radv_amdgpu_cs_submit() 1303 chunks[i].chunk_data = (uint64_t)(uintptr_t)&chunk_data[i]; in radv_amdgpu_cs_submit() 1307 chunk_data[i].ib_data._pad = 0; in radv_amdgpu_cs_submit() 1308 chunk_data[i].ib_data.va_start = ib->ib_mc_address; in radv_amdgpu_cs_submit() 1309 chunk_data[i].ib_data.ib_bytes = ib->size * 4; in radv_amdgpu_cs_submit() 1310 chunk_data[i].ib_data.ip_type = request->ip_type; in radv_amdgpu_cs_submit() 1311 chunk_data[i].ib_data.ip_instance = request->ip_instance; in radv_amdgpu_cs_submit() 1312 chunk_data[i].ib_data.ring = request->ring; in radv_amdgpu_cs_submit() [all …]
|
/external/curl/tests/libtest/ |
D | lib576.c | 104 chunk_data_t chunk_data = {0, 0}; in test() local 116 test_setopt(handle, CURLOPT_CHUNK_DATA, &chunk_data); in test()
|
/external/webp/include/webp/ |
D | mux.h | 164 WebPMux* mux, const char fourcc[4], const WebPData* chunk_data, 180 const WebPMux* mux, const char fourcc[4], WebPData* chunk_data);
|
/external/webp/src/webp/ |
D | mux.h | 164 WebPMux* mux, const char fourcc[4], const WebPData* chunk_data, 180 const WebPMux* mux, const char fourcc[4], WebPData* chunk_data);
|
/external/libdrm/radeon/ |
D | radeon_cs_gem.c | 165 csg->chunks[0].chunk_data = (uint64_t)(uintptr_t)csg->base.packets; in cs_gem_create() 168 csg->chunks[1].chunk_data = (uint64_t)(uintptr_t)csg->relocs; in cs_gem_create() 257 csg->chunks[1].chunk_data = (uint64_t)(uintptr_t)csg->relocs; in cs_gem_write_reloc()
|
/external/mesa3d/src/gallium/winsys/radeon/drm/ |
D | radeon_drm_cs.c | 95 csc->chunks[0].chunk_data = (uint64_t)(uintptr_t)csc->buf; in radeon_init_cs_context() 98 csc->chunks[1].chunk_data = (uint64_t)(uintptr_t)csc->relocs; in radeon_init_cs_context() 101 csc->chunks[2].chunk_data = (uint64_t)(uintptr_t)&csc->flags; in radeon_init_cs_context() 264 csc->chunks[1].chunk_data = (uint64_t)(uintptr_t)csc->relocs; in radeon_lookup_or_add_real_buffer()
|
/external/tensorflow/tensorflow/python/keras/saving/ |
D | hdf5_format.py | 866 for chunk_id, chunk_data in enumerate(chunked_data): 867 group.attrs['%s%d' % (name, chunk_id)] = chunk_data
|
/external/mesa3d/src/gallium/winsys/amdgpu/drm/ |
D | amdgpu_cs.c | 1310 chunks[num_chunks].chunk_data = (uintptr_t)&cs->ib[IB_MAIN]; in amdgpu_cs_submit_ib() 1317 chunks[num_chunks].chunk_data = (uintptr_t)&acs->fence_chunk; in amdgpu_cs_submit_ib() 1345 chunks[num_chunks].chunk_data = (uintptr_t)dep_chunk; in amdgpu_cs_submit_ib() 1368 chunks[num_chunks].chunk_data = (uintptr_t)sem_chunk; in amdgpu_cs_submit_ib()
|
/external/libdrm/include/drm/ |
D | amdgpu_drm.h | 489 __u64 chunk_data; member
|
D | radeon_drm.h | 971 __u64 chunk_data; member
|
/external/kernel-headers/original/uapi/drm/ |
D | amdgpu_drm.h | 530 __u64 chunk_data; member
|
D | radeon_drm.h | 970 __u64 chunk_data; member
|
/external/libpng/ |
D | CHANGES | 1099 Check for missing profile length field in iCCP chunk and free chunk_data
|