• Home
  • Raw
  • Download

Lines Matching refs:device

103 anv_physical_device_init_heaps(struct anv_physical_device *device, int fd)  in anv_physical_device_init_heaps()  argument
109 device->supports_48bit_addresses = in anv_physical_device_init_heaps()
110 (device->info.gen >= 8) && anv_gem_supports_48b_addresses(fd); in anv_physical_device_init_heaps()
117 if (heap_size > (2ull << 30) && !device->supports_48bit_addresses) { in anv_physical_device_init_heaps()
134 device->memory.heap_count = 1; in anv_physical_device_init_heaps()
135 device->memory.heaps[0] = (struct anv_memory_heap) { in anv_physical_device_init_heaps()
151 assert(device->supports_48bit_addresses); in anv_physical_device_init_heaps()
153 device->memory.heap_count = 2; in anv_physical_device_init_heaps()
154 device->memory.heaps[0] = (struct anv_memory_heap) { in anv_physical_device_init_heaps()
159 device->memory.heaps[1] = (struct anv_memory_heap) { in anv_physical_device_init_heaps()
167 for (uint32_t heap = 0; heap < device->memory.heap_count; heap++) { in anv_physical_device_init_heaps()
177 if (device->memory.heaps[heap].supports_48bit_addresses) { in anv_physical_device_init_heaps()
182 if (device->info.has_llc) { in anv_physical_device_init_heaps()
186 device->memory.types[type_count++] = (struct anv_memory_type) { in anv_physical_device_init_heaps()
200 device->memory.types[type_count++] = (struct anv_memory_type) { in anv_physical_device_init_heaps()
207 device->memory.types[type_count++] = (struct anv_memory_type) { in anv_physical_device_init_heaps()
216 device->memory.type_count = type_count; in anv_physical_device_init_heaps()
222 anv_physical_device_init_uuids(struct anv_physical_device *device) in anv_physical_device_init_uuids() argument
227 return vk_errorf(device->instance, device, in anv_physical_device_init_uuids()
234 return vk_errorf(device->instance, device, in anv_physical_device_init_uuids()
248 _mesa_sha1_update(&sha1_ctx, &device->chipset_id, in anv_physical_device_init_uuids()
249 sizeof(device->chipset_id)); in anv_physical_device_init_uuids()
251 memcpy(device->pipeline_cache_uuid, sha1, VK_UUID_SIZE); in anv_physical_device_init_uuids()
258 memcpy(device->driver_uuid, build_id_data(note), VK_UUID_SIZE); in anv_physical_device_init_uuids()
267 _mesa_sha1_update(&sha1_ctx, &device->chipset_id, in anv_physical_device_init_uuids()
268 sizeof(device->chipset_id)); in anv_physical_device_init_uuids()
269 _mesa_sha1_update(&sha1_ctx, &device->isl_dev.has_bit6_swizzling, in anv_physical_device_init_uuids()
270 sizeof(device->isl_dev.has_bit6_swizzling)); in anv_physical_device_init_uuids()
272 memcpy(device->device_uuid, sha1, VK_UUID_SIZE); in anv_physical_device_init_uuids()
278 anv_physical_device_init(struct anv_physical_device *device, in anv_physical_device_init() argument
291 device->_loader_data.loaderMagic = ICD_LOADER_MAGIC; in anv_physical_device_init()
292 device->instance = instance; in anv_physical_device_init()
294 assert(strlen(path) < ARRAY_SIZE(device->path)); in anv_physical_device_init()
295 strncpy(device->path, path, ARRAY_SIZE(device->path)); in anv_physical_device_init()
297 device->chipset_id = anv_gem_get_param(fd, I915_PARAM_CHIPSET_ID); in anv_physical_device_init()
298 if (!device->chipset_id) { in anv_physical_device_init()
303 device->name = gen_get_device_name(device->chipset_id); in anv_physical_device_init()
304 if (!gen_get_device_info(device->chipset_id, &device->info)) { in anv_physical_device_init()
309 if (device->info.is_haswell) { in anv_physical_device_init()
311 } else if (device->info.gen == 7 && !device->info.is_baytrail) { in anv_physical_device_init()
313 } else if (device->info.gen == 7 && device->info.is_baytrail) { in anv_physical_device_init()
315 } else if (device->info.gen >= 8 && device->info.gen <= 10) { in anv_physical_device_init()
318 result = vk_errorf(device->instance, device, in anv_physical_device_init()
320 "Vulkan not yet supported on %s", device->name); in anv_physical_device_init()
324 device->cmd_parser_version = -1; in anv_physical_device_init()
325 if (device->info.gen == 7) { in anv_physical_device_init()
326 device->cmd_parser_version = in anv_physical_device_init()
328 if (device->cmd_parser_version == -1) { in anv_physical_device_init()
329 result = vk_errorf(device->instance, device, in anv_physical_device_init()
337 result = vk_errorf(device->instance, device, in anv_physical_device_init()
344 result = vk_errorf(device->instance, device, in anv_physical_device_init()
350 if (!device->info.has_llc && in anv_physical_device_init()
352 result = vk_errorf(device->instance, device, in anv_physical_device_init()
358 result = anv_physical_device_init_heaps(device, fd); in anv_physical_device_init()
362 device->has_exec_async = anv_gem_get_param(fd, I915_PARAM_HAS_EXEC_ASYNC); in anv_physical_device_init()
363 device->has_exec_capture = anv_gem_get_param(fd, I915_PARAM_HAS_EXEC_CAPTURE); in anv_physical_device_init()
364 device->has_exec_fence = anv_gem_get_param(fd, I915_PARAM_HAS_EXEC_FENCE); in anv_physical_device_init()
365 device->has_syncobj = anv_gem_get_param(fd, I915_PARAM_HAS_EXEC_FENCE_ARRAY); in anv_physical_device_init()
366 device->has_syncobj_wait = device->has_syncobj && in anv_physical_device_init()
374 if (device->info.gen >= 10) { in anv_physical_device_init()
381 device->info.timestamp_frequency = timestamp_frequency; in anv_physical_device_init()
385 if (device->info.gen >= 8) { in anv_physical_device_init()
386 device->subslice_total = anv_gem_get_param(fd, I915_PARAM_SUBSLICE_TOTAL); in anv_physical_device_init()
387 device->eu_total = anv_gem_get_param(fd, I915_PARAM_EU_TOTAL); in anv_physical_device_init()
393 if (device->subslice_total < 1 || device->eu_total < 1) { in anv_physical_device_init()
396 } else if (device->info.gen == 7) { in anv_physical_device_init()
397 device->subslice_total = 1 << (device->info.gt - 1); in anv_physical_device_init()
400 if (device->info.is_cherryview && in anv_physical_device_init()
401 device->subslice_total > 0 && device->eu_total > 0) { in anv_physical_device_init()
404 device->eu_total / device->subslice_total * device->info.num_thread_per_eu; in anv_physical_device_init()
407 if (max_cs_threads > device->info.max_cs_threads) in anv_physical_device_init()
408 device->info.max_cs_threads = max_cs_threads; in anv_physical_device_init()
411 device->compiler = brw_compiler_create(NULL, &device->info); in anv_physical_device_init()
412 if (device->compiler == NULL) { in anv_physical_device_init()
416 device->compiler->shader_debug_log = compiler_debug_log; in anv_physical_device_init()
417 device->compiler->shader_perf_log = compiler_perf_log; in anv_physical_device_init()
418 device->compiler->supports_pull_constants = false; in anv_physical_device_init()
419 device->compiler->constant_buffer_0_is_relative = true; in anv_physical_device_init()
421 isl_device_init(&device->isl_dev, &device->info, swizzled); in anv_physical_device_init()
423 result = anv_physical_device_init_uuids(device); in anv_physical_device_init()
427 result = anv_init_wsi(device); in anv_physical_device_init()
429 ralloc_free(device->compiler); in anv_physical_device_init()
433 anv_physical_device_get_supported_extensions(device, in anv_physical_device_init()
434 &device->supported_extensions); in anv_physical_device_init()
436 device->local_fd = fd; in anv_physical_device_init()
445 anv_physical_device_finish(struct anv_physical_device *device) in anv_physical_device_finish() argument
447 anv_finish_wsi(device); in anv_physical_device_finish()
448 ralloc_free(device->compiler); in anv_physical_device_finish()
449 close(device->local_fd); in anv_physical_device_finish()
1131 ANV_FROM_HANDLE(anv_device, device, _device); in anv_GetDeviceProcAddr()
1133 if (!device || !pName) in anv_GetDeviceProcAddr()
1140 return device->dispatch.entrypoints[idx]; in anv_GetDeviceProcAddr()
1181 anv_queue_init(struct anv_device *device, struct anv_queue *queue) in anv_queue_init() argument
1184 queue->device = device; in anv_queue_init()
1185 queue->pool = &device->surface_state_pool; in anv_queue_init()
1201 anv_state_flush(pool->block_pool.device, state); in anv_state_pool_emit_data()
1216 anv_device_init_border_colors(struct anv_device *device) in anv_device_init_border_colors() argument
1227 device->border_colors = anv_state_pool_emit_data(&device->dynamic_state_pool, in anv_device_init_border_colors()
1233 anv_device_init_trivial_batch(struct anv_device *device) in anv_device_init_trivial_batch() argument
1235 anv_bo_init_new(&device->trivial_batch_bo, device, 4096); in anv_device_init_trivial_batch()
1237 if (device->instance->physicalDevice.has_exec_async) in anv_device_init_trivial_batch()
1238 device->trivial_batch_bo.flags |= EXEC_OBJECT_ASYNC; in anv_device_init_trivial_batch()
1240 void *map = anv_gem_mmap(device, device->trivial_batch_bo.gem_handle, in anv_device_init_trivial_batch()
1252 if (!device->info.has_llc) in anv_device_init_trivial_batch()
1255 anv_gem_munmap(map, device->trivial_batch_bo.size); in anv_device_init_trivial_batch()
1264 ANV_FROM_HANDLE(anv_physical_device, device, physicalDevice); in anv_EnumerateDeviceExtensionProperties()
1266 (void)device; in anv_EnumerateDeviceExtensionProperties()
1269 if (device->supported_extensions.extensions[i]) { in anv_EnumerateDeviceExtensionProperties()
1280 anv_device_init_dispatch(struct anv_device *device) in anv_device_init_dispatch() argument
1283 switch (device->info.gen) { in anv_device_init_dispatch()
1294 if (device->info.is_haswell) in anv_device_init_dispatch()
1303 for (unsigned i = 0; i < ARRAY_SIZE(device->dispatch.entrypoints); i++) { in anv_device_init_dispatch()
1307 if (!anv_entrypoint_is_enabled(i, device->instance->apiVersion, in anv_device_init_dispatch()
1308 &device->instance->enabled_extensions, in anv_device_init_dispatch()
1309 &device->enabled_extensions)) { in anv_device_init_dispatch()
1310 device->dispatch.entrypoints[i] = NULL; in anv_device_init_dispatch()
1312 device->dispatch.entrypoints[i] = genX_table->entrypoints[i]; in anv_device_init_dispatch()
1314 device->dispatch.entrypoints[i] = anv_dispatch_table.entrypoints[i]; in anv_device_init_dispatch()
1327 struct anv_device *device; in anv_CreateDevice() local
1362 device = vk_alloc2(&physical_device->instance->alloc, pAllocator, in anv_CreateDevice()
1363 sizeof(*device), 8, in anv_CreateDevice()
1365 if (!device) in anv_CreateDevice()
1368 device->_loader_data.loaderMagic = ICD_LOADER_MAGIC; in anv_CreateDevice()
1369 device->instance = physical_device->instance; in anv_CreateDevice()
1370 device->chipset_id = physical_device->chipset_id; in anv_CreateDevice()
1371 device->lost = false; in anv_CreateDevice()
1374 device->alloc = *pAllocator; in anv_CreateDevice()
1376 device->alloc = physical_device->instance->alloc; in anv_CreateDevice()
1379 device->fd = open(physical_device->path, O_RDWR | O_CLOEXEC); in anv_CreateDevice()
1380 if (device->fd == -1) { in anv_CreateDevice()
1385 device->context_id = anv_gem_create_context(device); in anv_CreateDevice()
1386 if (device->context_id == -1) { in anv_CreateDevice()
1391 device->info = physical_device->info; in anv_CreateDevice()
1392 device->isl_dev = physical_device->isl_dev; in anv_CreateDevice()
1399 device->can_chain_batches = device->info.gen >= 8; in anv_CreateDevice()
1401 device->robust_buffer_access = pCreateInfo->pEnabledFeatures && in anv_CreateDevice()
1403 device->enabled_extensions = enabled_extensions; in anv_CreateDevice()
1405 anv_device_init_dispatch(device); in anv_CreateDevice()
1407 if (pthread_mutex_init(&device->mutex, NULL) != 0) { in anv_CreateDevice()
1422 if (pthread_cond_init(&device->queue_submit, NULL) != 0) { in anv_CreateDevice()
1434 anv_bo_pool_init(&device->batch_bo_pool, device, bo_flags); in anv_CreateDevice()
1436 result = anv_bo_cache_init(&device->bo_cache); in anv_CreateDevice()
1444 result = anv_state_pool_init(&device->dynamic_state_pool, device, 16384, in anv_CreateDevice()
1449 result = anv_state_pool_init(&device->instruction_state_pool, device, 16384, in anv_CreateDevice()
1454 result = anv_state_pool_init(&device->surface_state_pool, device, 4096, in anv_CreateDevice()
1459 result = anv_bo_init_new(&device->workaround_bo, device, 1024); in anv_CreateDevice()
1463 anv_device_init_trivial_batch(device); in anv_CreateDevice()
1465 anv_scratch_pool_init(device, &device->scratch_pool); in anv_CreateDevice()
1467 anv_queue_init(device, &device->queue); in anv_CreateDevice()
1469 switch (device->info.gen) { in anv_CreateDevice()
1471 if (!device->info.is_haswell) in anv_CreateDevice()
1472 result = gen7_init_device_state(device); in anv_CreateDevice()
1474 result = gen75_init_device_state(device); in anv_CreateDevice()
1477 result = gen8_init_device_state(device); in anv_CreateDevice()
1480 result = gen9_init_device_state(device); in anv_CreateDevice()
1483 result = gen10_init_device_state(device); in anv_CreateDevice()
1493 anv_device_init_blorp(device); in anv_CreateDevice()
1495 anv_device_init_border_colors(device); in anv_CreateDevice()
1497 *pDevice = anv_device_to_handle(device); in anv_CreateDevice()
1502 anv_queue_finish(&device->queue); in anv_CreateDevice()
1503 anv_scratch_pool_finish(device, &device->scratch_pool); in anv_CreateDevice()
1504 anv_gem_munmap(device->workaround_bo.map, device->workaround_bo.size); in anv_CreateDevice()
1505 anv_gem_close(device, device->workaround_bo.gem_handle); in anv_CreateDevice()
1507 anv_state_pool_finish(&device->surface_state_pool); in anv_CreateDevice()
1509 anv_state_pool_finish(&device->instruction_state_pool); in anv_CreateDevice()
1511 anv_state_pool_finish(&device->dynamic_state_pool); in anv_CreateDevice()
1513 anv_bo_cache_finish(&device->bo_cache); in anv_CreateDevice()
1515 anv_bo_pool_finish(&device->batch_bo_pool); in anv_CreateDevice()
1516 pthread_cond_destroy(&device->queue_submit); in anv_CreateDevice()
1518 pthread_mutex_destroy(&device->mutex); in anv_CreateDevice()
1520 anv_gem_destroy_context(device, device->context_id); in anv_CreateDevice()
1522 close(device->fd); in anv_CreateDevice()
1524 vk_free(&device->alloc, device); in anv_CreateDevice()
1533 ANV_FROM_HANDLE(anv_device, device, _device); in anv_DestroyDevice()
1535 if (!device) in anv_DestroyDevice()
1538 anv_device_finish_blorp(device); in anv_DestroyDevice()
1540 anv_queue_finish(&device->queue); in anv_DestroyDevice()
1546 anv_state_pool_free(&device->dynamic_state_pool, device->border_colors); in anv_DestroyDevice()
1549 anv_scratch_pool_finish(device, &device->scratch_pool); in anv_DestroyDevice()
1551 anv_gem_munmap(device->workaround_bo.map, device->workaround_bo.size); in anv_DestroyDevice()
1552 anv_gem_close(device, device->workaround_bo.gem_handle); in anv_DestroyDevice()
1554 anv_gem_close(device, device->trivial_batch_bo.gem_handle); in anv_DestroyDevice()
1556 anv_state_pool_finish(&device->surface_state_pool); in anv_DestroyDevice()
1557 anv_state_pool_finish(&device->instruction_state_pool); in anv_DestroyDevice()
1558 anv_state_pool_finish(&device->dynamic_state_pool); in anv_DestroyDevice()
1560 anv_bo_cache_finish(&device->bo_cache); in anv_DestroyDevice()
1562 anv_bo_pool_finish(&device->batch_bo_pool); in anv_DestroyDevice()
1564 pthread_cond_destroy(&device->queue_submit); in anv_DestroyDevice()
1565 pthread_mutex_destroy(&device->mutex); in anv_DestroyDevice()
1567 anv_gem_destroy_context(device, device->context_id); in anv_DestroyDevice()
1569 close(device->fd); in anv_DestroyDevice()
1571 vk_free(&device->alloc, device); in anv_DestroyDevice()
1607 ANV_FROM_HANDLE(anv_device, device, _device); in anv_GetDeviceQueue()
1611 *pQueue = anv_queue_to_handle(&device->queue); in anv_GetDeviceQueue()
1615 anv_device_query_status(struct anv_device *device) in anv_device_query_status() argument
1621 if (unlikely(device->lost)) in anv_device_query_status()
1625 int ret = anv_gem_gpu_get_reset_stats(device, &active, &pending); in anv_device_query_status()
1628 device->lost = true; in anv_device_query_status()
1629 return vk_errorf(device->instance, device, VK_ERROR_DEVICE_LOST, in anv_device_query_status()
1634 device->lost = true; in anv_device_query_status()
1635 return vk_errorf(device->instance, device, VK_ERROR_DEVICE_LOST, in anv_device_query_status()
1638 device->lost = true; in anv_device_query_status()
1639 return vk_errorf(device->instance, device, VK_ERROR_DEVICE_LOST, in anv_device_query_status()
1647 anv_device_bo_busy(struct anv_device *device, struct anv_bo *bo) in anv_device_bo_busy() argument
1653 int ret = anv_gem_busy(device, bo->gem_handle); in anv_device_bo_busy()
1658 device->lost = true; in anv_device_bo_busy()
1659 return vk_errorf(device->instance, device, VK_ERROR_DEVICE_LOST, in anv_device_bo_busy()
1669 return anv_device_query_status(device); in anv_device_bo_busy()
1673 anv_device_wait(struct anv_device *device, struct anv_bo *bo, in anv_device_wait() argument
1676 int ret = anv_gem_wait(device, bo->gem_handle, &timeout); in anv_device_wait()
1681 device->lost = true; in anv_device_wait()
1682 return vk_errorf(device->instance, device, VK_ERROR_DEVICE_LOST, in anv_device_wait()
1691 return anv_device_query_status(device); in anv_device_wait()
1697 ANV_FROM_HANDLE(anv_device, device, _device); in anv_DeviceWaitIdle()
1698 if (unlikely(device->lost)) in anv_DeviceWaitIdle()
1710 return anv_device_submit_simple_batch(device, &batch); in anv_DeviceWaitIdle()
1714 anv_bo_init_new(struct anv_bo *bo, struct anv_device *device, uint64_t size) in anv_bo_init_new() argument
1716 uint32_t gem_handle = anv_gem_create(device, size); in anv_bo_init_new()
1731 ANV_FROM_HANDLE(anv_device, device, _device); in anv_AllocateMemory()
1732 struct anv_physical_device *pdevice = &device->instance->physicalDevice; in anv_AllocateMemory()
1763 mem = vk_alloc2(&device->alloc, pAllocator, sizeof(*mem), 8, in anv_AllocateMemory()
1786 result = anv_bo_cache_import(device, &device->bo_cache, in anv_AllocateMemory()
1803 result = vk_errorf(device->instance, device, in anv_AllocateMemory()
1809 anv_bo_cache_release(device, &device->bo_cache, mem->bo); in anv_AllocateMemory()
1824 result = anv_bo_cache_alloc(device, &device->bo_cache, in anv_AllocateMemory()
1852 int ret = anv_gem_set_tiling(device, mem->bo->gem_handle, in anv_AllocateMemory()
1856 anv_bo_cache_release(device, &device->bo_cache, mem->bo); in anv_AllocateMemory()
1857 return vk_errorf(device->instance, NULL, in anv_AllocateMemory()
1886 vk_free2(&device->alloc, pAllocator, mem); in anv_AllocateMemory()
1913 ANV_FROM_HANDLE(anv_device, device, _device); in anv_GetMemoryFdPropertiesKHR()
1914 struct anv_physical_device *pdevice = &device->instance->physicalDevice; in anv_GetMemoryFdPropertiesKHR()
1940 ANV_FROM_HANDLE(anv_device, device, _device); in anv_FreeMemory()
1949 anv_bo_cache_release(device, &device->bo_cache, mem->bo); in anv_FreeMemory()
1951 vk_free2(&device->alloc, pAllocator, mem); in anv_FreeMemory()
1962 ANV_FROM_HANDLE(anv_device, device, _device); in anv_MapMemory()
1991 if (!device->info.has_llc && in anv_MapMemory()
2003 void *map = anv_gem_mmap(device, mem->bo->gem_handle, in anv_MapMemory()
2032 clflush_mapped_ranges(struct anv_device *device, in clflush_mapped_ranges() argument
2051 ANV_FROM_HANDLE(anv_device, device, _device); in anv_FlushMappedMemoryRanges()
2053 if (device->info.has_llc) in anv_FlushMappedMemoryRanges()
2059 clflush_mapped_ranges(device, memoryRangeCount, pMemoryRanges); in anv_FlushMappedMemoryRanges()
2069 ANV_FROM_HANDLE(anv_device, device, _device); in anv_InvalidateMappedMemoryRanges()
2071 if (device->info.has_llc) in anv_InvalidateMappedMemoryRanges()
2074 clflush_mapped_ranges(device, memoryRangeCount, pMemoryRanges); in anv_InvalidateMappedMemoryRanges()
2088 ANV_FROM_HANDLE(anv_device, device, _device); in anv_GetBufferMemoryRequirements()
2089 struct anv_physical_device *pdevice = &device->instance->physicalDevice; in anv_GetBufferMemoryRequirements()
2147 ANV_FROM_HANDLE(anv_device, device, _device); in anv_GetImageMemoryRequirements()
2148 struct anv_physical_device *pdevice = &device->instance->physicalDevice; in anv_GetImageMemoryRequirements()
2171 ANV_FROM_HANDLE(anv_device, device, _device); in anv_GetImageMemoryRequirements2KHR()
2180 struct anv_physical_device *pdevice = &device->instance->physicalDevice; in anv_GetImageMemoryRequirements2KHR()
2239 VkDevice device, in anv_GetImageSparseMemoryRequirements() argument
2248 VkDevice device, in anv_GetImageSparseMemoryRequirements2KHR() argument
2257 VkDevice device, in anv_GetDeviceMemoryCommitment() argument
2283 VkDevice device, in anv_BindBufferMemory() argument
2300 VkDevice device, in anv_BindBufferMemory2KHR() argument
2317 if (unlikely(queue->device->lost)) in anv_QueueBindSparse()
2331 ANV_FROM_HANDLE(anv_device, device, _device); in anv_CreateEvent()
2337 state = anv_state_pool_alloc(&device->dynamic_state_pool, in anv_CreateEvent()
2343 if (!device->info.has_llc) { in anv_CreateEvent()
2359 ANV_FROM_HANDLE(anv_device, device, _device); in anv_DestroyEvent()
2365 anv_state_pool_free(&device->dynamic_state_pool, event->state); in anv_DestroyEvent()
2372 ANV_FROM_HANDLE(anv_device, device, _device); in anv_GetEventStatus()
2375 if (unlikely(device->lost)) in anv_GetEventStatus()
2378 if (!device->info.has_llc) { in anv_GetEventStatus()
2392 ANV_FROM_HANDLE(anv_device, device, _device); in anv_SetEvent()
2397 if (!device->info.has_llc) { in anv_SetEvent()
2410 ANV_FROM_HANDLE(anv_device, device, _device); in anv_ResetEvent()
2415 if (!device->info.has_llc) { in anv_ResetEvent()
2432 ANV_FROM_HANDLE(anv_device, device, _device); in anv_CreateBuffer()
2437 buffer = vk_alloc2(&device->alloc, pAllocator, sizeof(*buffer), 8, in anv_CreateBuffer()
2457 ANV_FROM_HANDLE(anv_device, device, _device); in anv_DestroyBuffer()
2463 vk_free2(&device->alloc, pAllocator, buffer); in anv_DestroyBuffer()
2467 anv_fill_buffer_surface_state(struct anv_device *device, struct anv_state state, in anv_fill_buffer_surface_state() argument
2471 isl_buffer_fill_state(&device->isl_dev, state.map, in anv_fill_buffer_surface_state()
2473 .mocs = device->default_mocs, in anv_fill_buffer_surface_state()
2478 anv_state_flush(device, state); in anv_fill_buffer_surface_state()
2486 ANV_FROM_HANDLE(anv_device, device, _device); in anv_DestroySampler()
2492 vk_free2(&device->alloc, pAllocator, sampler); in anv_DestroySampler()
2501 ANV_FROM_HANDLE(anv_device, device, _device); in anv_CreateFramebuffer()
2508 framebuffer = vk_alloc2(&device->alloc, pAllocator, size, 8, in anv_CreateFramebuffer()
2533 ANV_FROM_HANDLE(anv_device, device, _device); in anv_DestroyFramebuffer()
2539 vk_free2(&device->alloc, pAllocator, fb); in anv_DestroyFramebuffer()