Lines Matching +full:file +full:- +full:entry +full:- +full:cache
31 #include <sys/file.h>
45 #include "util/mesa-sha1.h"
53 /* The cache version should be bumped whenever a change is made to the
54 * structure of cache entries or the index. This will give any 3rd party
55 * applications reading the cache entries a chance to adjust to the changes.
57 * - The cache version is checked internally when reading a cache entry. If we
58 * ever have a mismatch we are in big trouble as this means we had a cache
62 * - There is no strict requirement that cache versions be backwards
74 disk_cache_init_queue(struct disk_cache *cache) in disk_cache_init_queue() argument
76 if (util_queue_is_initialized(&cache->cache_queue)) in disk_cache_init_queue()
82 * avoiding excessive memory use due to a backlog of cache entrys building in disk_cache_init_queue()
89 return util_queue_init(&cache->cache_queue, "disk$", 32, 4, in disk_cache_init_queue()
104 struct disk_cache *cache = NULL; in disk_cache_type_create() local
114 cache = rzalloc(NULL, struct disk_cache); in disk_cache_type_create()
115 if (cache == NULL) in disk_cache_type_create()
119 cache->path_init_failed = true; in disk_cache_type_create()
120 cache->type = DISK_CACHE_NONE; in disk_cache_type_create()
130 cache->path = ralloc_strdup(cache, path); in disk_cache_type_create()
131 if (cache->path == NULL) in disk_cache_type_create()
134 /* Cache tests that want to have a disabled cache compression are using in disk_cache_type_create()
136 * disable disk cache compression when mesa's build tests require it. in disk_cache_type_create()
139 cache->compression_disabled = true; in disk_cache_type_create()
142 if (!disk_cache_load_cache_index_foz(local, cache)) in disk_cache_type_create()
145 if (!disk_cache_db_load_cache_index(local, cache)) in disk_cache_type_create()
150 disk_cache_touch_cache_user_marker(cache->path); in disk_cache_type_create()
152 cache->type = cache_type; in disk_cache_type_create()
154 cache->stats.enabled = debug_get_bool_option("MESA_SHADER_CACHE_SHOW_STATS", in disk_cache_type_create()
157 if (!disk_cache_mmap_cache_index(local, cache, path)) in disk_cache_type_create()
160 cache->max_size = max_size; in disk_cache_type_create()
162 if (cache->type == DISK_CACHE_DATABASE) in disk_cache_type_create()
163 mesa_cache_db_multipart_set_size_limit(&cache->cache_db, cache->max_size); in disk_cache_type_create()
165 if (!disk_cache_init_queue(cache)) in disk_cache_type_create()
168 cache->path_init_failed = false; in disk_cache_type_create()
172 cache->driver_keys_blob_size = cv_size; in disk_cache_type_create()
177 cache->driver_keys_blob_size += id_size; in disk_cache_type_create()
178 cache->driver_keys_blob_size += gpu_name_size; in disk_cache_type_create()
180 /* We sometimes store entire structs that contains a pointers in the cache, in disk_cache_type_create()
185 cache->driver_keys_blob_size += ptr_size_size; in disk_cache_type_create()
188 cache->driver_keys_blob_size += driver_flags_size; in disk_cache_type_create()
190 cache->driver_keys_blob = in disk_cache_type_create()
191 ralloc_size(cache, cache->driver_keys_blob_size); in disk_cache_type_create()
192 if (!cache->driver_keys_blob) in disk_cache_type_create()
195 uint8_t *drv_key_blob = cache->driver_keys_blob; in disk_cache_type_create()
203 s_rand_xorshift128plus(cache->seed_xorshift128plus, true); in disk_cache_type_create()
207 return cache; in disk_cache_type_create()
210 if (cache) in disk_cache_type_create()
211 ralloc_free(cache); in disk_cache_type_create()
222 struct disk_cache *cache; in disk_cache_create() local
232 /* Since switching the default cache to <mesa_shader_cache_db>, remove the in disk_cache_create()
233 * old cache folder if it hasn't been modified for more than 7 days. in disk_cache_create()
280 /* Default to 1GB for maximum cache size. */ in disk_cache_create()
285 /* Create main writable cache. */ in disk_cache_create()
286 cache = disk_cache_type_create(gpu_name, driver_id, NULL, driver_flags, in disk_cache_create()
288 if (!cache) in disk_cache_create()
293 * cache. At first we will check cache entry presence in the RO caches and in disk_cache_create()
294 * if entry isn't found there, then we'll fall back to the RW cache. in disk_cache_create()
296 if (cache_type != DISK_CACHE_SINGLE_FILE && !cache->path_init_failed && in disk_cache_create()
299 /* Create read-only cache used for sharing prebuilt shaders. in disk_cache_create()
300 * If cache entry will be found in this cache, then the main cache in disk_cache_create()
303 cache->foz_ro_cache = disk_cache_type_create(gpu_name, driver_id, NULL, in disk_cache_create()
309 return cache; in disk_cache_create()
322 disk_cache_destroy(struct disk_cache *cache) in disk_cache_destroy() argument
324 if (unlikely(cache && cache->stats.enabled)) { in disk_cache_destroy()
325 printf("disk shader cache: hits = %u, misses = %u\n", in disk_cache_destroy()
326 cache->stats.hits, in disk_cache_destroy()
327 cache->stats.misses); in disk_cache_destroy()
330 if (cache && util_queue_is_initialized(&cache->cache_queue)) { in disk_cache_destroy()
331 util_queue_finish(&cache->cache_queue); in disk_cache_destroy()
332 util_queue_destroy(&cache->cache_queue); in disk_cache_destroy()
334 if (cache->foz_ro_cache) in disk_cache_destroy()
335 disk_cache_destroy(cache->foz_ro_cache); in disk_cache_destroy()
337 if (cache->type == DISK_CACHE_SINGLE_FILE) in disk_cache_destroy()
338 foz_destroy(&cache->foz_db); in disk_cache_destroy()
340 if (cache->type == DISK_CACHE_DATABASE) in disk_cache_destroy()
341 mesa_cache_db_multipart_close(&cache->cache_db); in disk_cache_destroy()
343 disk_cache_destroy_mmap(cache); in disk_cache_destroy()
346 ralloc_free(cache); in disk_cache_destroy()
350 disk_cache_wait_for_idle(struct disk_cache *cache) in disk_cache_wait_for_idle() argument
352 util_queue_finish(&cache->cache_queue); in disk_cache_wait_for_idle()
356 disk_cache_remove(struct disk_cache *cache, const cache_key key) in disk_cache_remove() argument
358 if (cache->type == DISK_CACHE_DATABASE) { in disk_cache_remove()
359 mesa_cache_db_multipart_entry_remove(&cache->cache_db, key); in disk_cache_remove()
363 char *filename = disk_cache_get_cache_filename(cache, key); in disk_cache_remove()
368 disk_cache_evict_item(cache, filename); in disk_cache_remove()
372 create_put_job(struct disk_cache *cache, const cache_key key, in create_put_job() argument
381 dc_job->cache = cache; in create_put_job()
382 memcpy(dc_job->key, key, sizeof(cache_key)); in create_put_job()
384 dc_job->data = data; in create_put_job()
386 dc_job->data = dc_job + 1; in create_put_job()
387 memcpy(dc_job->data, data, size); in create_put_job()
389 dc_job->size = size; in create_put_job()
391 /* Copy the cache item metadata */ in create_put_job()
393 dc_job->cache_item_metadata.type = cache_item_metadata->type; in create_put_job()
394 if (cache_item_metadata->type == CACHE_ITEM_TYPE_GLSL) { in create_put_job()
395 dc_job->cache_item_metadata.num_keys = in create_put_job()
396 cache_item_metadata->num_keys; in create_put_job()
397 dc_job->cache_item_metadata.keys = (cache_key *) in create_put_job()
398 malloc(cache_item_metadata->num_keys * sizeof(cache_key)); in create_put_job()
400 if (!dc_job->cache_item_metadata.keys) in create_put_job()
403 memcpy(dc_job->cache_item_metadata.keys, in create_put_job()
404 cache_item_metadata->keys, in create_put_job()
405 sizeof(cache_key) * cache_item_metadata->num_keys); in create_put_job()
408 dc_job->cache_item_metadata.type = CACHE_ITEM_TYPE_UNKNOWN; in create_put_job()
409 dc_job->cache_item_metadata.keys = NULL; in create_put_job()
426 free(dc_job->cache_item_metadata.keys); in destroy_put_job()
435 free(dc_job->data); in destroy_put_job_nocopy()
440 blob_put_compressed(struct disk_cache *cache, const cache_key key,
452 if (dc_job->cache->blob_put_cb) { in cache_put()
453 blob_put_compressed(dc_job->cache, dc_job->key, dc_job->data, dc_job->size); in cache_put()
454 } else if (dc_job->cache->type == DISK_CACHE_SINGLE_FILE) { in cache_put()
456 } else if (dc_job->cache->type == DISK_CACHE_DATABASE) { in cache_put()
458 } else if (dc_job->cache->type == DISK_CACHE_MULTI_FILE) { in cache_put()
459 filename = disk_cache_get_cache_filename(dc_job->cache, dc_job->key); in cache_put()
463 /* If the cache is too large, evict something else first. */ in cache_put()
464 …while (p_atomic_read_relaxed(&dc_job->cache->size->value) + dc_job->size > dc_job->cache->max_size… in cache_put()
466 disk_cache_evict_lru_item(dc_job->cache); in cache_put()
483 blob_put_compressed(struct disk_cache *cache, const cache_key key, in blob_put_compressed() argument
489 struct blob_cache_entry *entry = malloc(max_buf + sizeof(*entry)); in blob_put_compressed() local
490 if (!entry) in blob_put_compressed()
493 entry->uncompressed_size = size; in blob_put_compressed()
496 util_compress_deflate(data, size, entry->compressed_data, max_buf); in blob_put_compressed()
500 unsigned entry_size = compressed_size + sizeof(*entry); in blob_put_compressed()
504 cache->blob_put_cb(key, CACHE_KEY_SIZE, entry, entry_size); in blob_put_compressed()
508 free(entry); in blob_put_compressed()
512 blob_get_compressed(struct disk_cache *cache, const cache_key key, in blob_get_compressed() argument
521 struct blob_cache_entry *entry = malloc(max_blob_size); in blob_get_compressed() local
522 if (!entry) in blob_get_compressed()
529 entry_size = cache->blob_get_cb(key, CACHE_KEY_SIZE, entry, max_blob_size); in blob_get_compressed()
533 free(entry); in blob_get_compressed()
537 void *data = malloc(entry->uncompressed_size); in blob_get_compressed()
539 free(entry); in blob_get_compressed()
543 unsigned compressed_size = entry_size - sizeof(*entry); in blob_get_compressed()
544 bool ret = util_compress_inflate(entry->compressed_data, compressed_size, in blob_get_compressed()
545 data, entry->uncompressed_size); in blob_get_compressed()
548 free(entry); in blob_get_compressed()
553 *size = entry->uncompressed_size; in blob_get_compressed()
555 free(entry); in blob_get_compressed()
561 disk_cache_put(struct disk_cache *cache, const cache_key key, in disk_cache_put() argument
565 if (!util_queue_is_initialized(&cache->cache_queue)) in disk_cache_put()
569 create_put_job(cache, key, (void*)data, size, cache_item_metadata, false); in disk_cache_put()
572 util_queue_fence_init(&dc_job->fence); in disk_cache_put()
573 util_queue_add_job(&cache->cache_queue, dc_job, &dc_job->fence, in disk_cache_put()
574 cache_put, destroy_put_job, dc_job->size); in disk_cache_put()
579 disk_cache_put_nocopy(struct disk_cache *cache, const cache_key key, in disk_cache_put_nocopy() argument
583 if (!util_queue_is_initialized(&cache->cache_queue)) { in disk_cache_put_nocopy()
589 create_put_job(cache, key, data, size, cache_item_metadata, true); in disk_cache_put_nocopy()
592 util_queue_fence_init(&dc_job->fence); in disk_cache_put_nocopy()
593 util_queue_add_job(&cache->cache_queue, dc_job, &dc_job->fence, in disk_cache_put_nocopy()
594 cache_put, destroy_put_job_nocopy, dc_job->size); in disk_cache_put_nocopy()
599 disk_cache_get(struct disk_cache *cache, const cache_key key, size_t *size) in disk_cache_get() argument
606 if (cache->foz_ro_cache) in disk_cache_get()
607 buf = disk_cache_load_item_foz(cache->foz_ro_cache, key, size); in disk_cache_get()
610 if (cache->blob_get_cb) { in disk_cache_get()
611 buf = blob_get_compressed(cache, key, size); in disk_cache_get()
612 } else if (cache->type == DISK_CACHE_SINGLE_FILE) { in disk_cache_get()
613 buf = disk_cache_load_item_foz(cache, key, size); in disk_cache_get()
614 } else if (cache->type == DISK_CACHE_DATABASE) { in disk_cache_get()
615 buf = disk_cache_db_load_item(cache, key, size); in disk_cache_get()
616 } else if (cache->type == DISK_CACHE_MULTI_FILE) { in disk_cache_get()
617 char *filename = disk_cache_get_cache_filename(cache, key); in disk_cache_get()
619 buf = disk_cache_load_item(cache, filename, size); in disk_cache_get()
623 if (unlikely(cache->stats.enabled)) { in disk_cache_get()
625 p_atomic_inc(&cache->stats.hits); in disk_cache_get()
627 p_atomic_inc(&cache->stats.misses); in disk_cache_get()
634 disk_cache_put_key(struct disk_cache *cache, const cache_key key) in disk_cache_put_key() argument
638 unsigned char *entry; in disk_cache_put_key() local
640 if (cache->blob_put_cb) { in disk_cache_put_key()
641 cache->blob_put_cb(key, CACHE_KEY_SIZE, key_chunk, sizeof(uint32_t)); in disk_cache_put_key()
645 if (cache->path_init_failed) in disk_cache_put_key()
648 entry = &cache->stored_keys[i * CACHE_KEY_SIZE]; in disk_cache_put_key()
650 memcpy(entry, key, CACHE_KEY_SIZE); in disk_cache_put_key()
654 * stored in the cache with disk_cache_put_key(). The implement is
656 * race-free, but the races are benign. If we race with someone else
657 * calling disk_cache_put_key, then that's just an extra cache miss and an
661 disk_cache_has_key(struct disk_cache *cache, const cache_key key) in disk_cache_has_key() argument
665 unsigned char *entry; in disk_cache_has_key() local
667 if (cache->blob_get_cb) { in disk_cache_has_key()
669 return cache->blob_get_cb(key, CACHE_KEY_SIZE, &blob, sizeof(uint32_t)); in disk_cache_has_key()
672 if (cache->path_init_failed) in disk_cache_has_key()
675 entry = &cache->stored_keys[i * CACHE_KEY_SIZE]; in disk_cache_has_key()
677 return memcmp(entry, key, CACHE_KEY_SIZE) == 0; in disk_cache_has_key()
681 disk_cache_compute_key(struct disk_cache *cache, const void *data, size_t size, in disk_cache_compute_key() argument
687 _mesa_sha1_update(&ctx, cache->driver_keys_blob, in disk_cache_compute_key()
688 cache->driver_keys_blob_size); in disk_cache_compute_key()
694 disk_cache_set_callbacks(struct disk_cache *cache, disk_cache_put_cb put, in disk_cache_set_callbacks() argument
697 cache->blob_put_cb = put; in disk_cache_set_callbacks()
698 cache->blob_get_cb = get; in disk_cache_set_callbacks()
699 disk_cache_init_queue(cache); in disk_cache_set_callbacks()