/************************************************************************** * * Copyright 2018-2019 Alyssa Rosenzweig * Copyright 2018-2019 Collabora, Ltd. * Copyright © 2015 Intel Corporation * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #ifndef PAN_DEVICE_H #define PAN_DEVICE_H #include #include "renderonly/renderonly.h" #include "util/u_dynarray.h" #include "util/bitset.h" #include "util/list.h" #include "util/sparse_array.h" #include /* Driver limits */ #define PAN_MAX_CONST_BUFFERS 16 /* Transient slab size. This is a balance between fragmentation against cache * locality and ease of bookkeeping */ #define TRANSIENT_SLAB_PAGES (16) /* 64kb */ #define TRANSIENT_SLAB_SIZE (4096 * TRANSIENT_SLAB_PAGES) /* Maximum number of transient slabs so we don't need dynamic arrays. Most * interesting Mali boards are 4GB RAM max, so if the entire RAM was filled * with transient slabs, you could never exceed (4GB / TRANSIENT_SLAB_SIZE) * allocations anyway. By capping, we can use a fixed-size bitset for tracking * free slabs, eliminating quite a bit of complexity. We can pack the free * state of 8 slabs into a single byte, so for 128kb transient slabs the bitset * occupies a cheap 4kb of memory */ #define MAX_TRANSIENT_SLABS (1024*1024 / TRANSIENT_SLAB_PAGES) /* How many power-of-two levels in the BO cache do we want? 2^12 * minimum chosen as it is the page size that all allocations are * rounded to */ #define MIN_BO_CACHE_BUCKET (12) /* 2^12 = 4KB */ #define MAX_BO_CACHE_BUCKET (22) /* 2^22 = 4MB */ /* Fencepost problem, hence the off-by-one */ #define NR_BO_CACHE_BUCKETS (MAX_BO_CACHE_BUCKET - MIN_BO_CACHE_BUCKET + 1) /* Cache for blit shaders. Defined here so they can be cached with the device */ enum pan_blit_type { PAN_BLIT_FLOAT = 0, PAN_BLIT_UINT, PAN_BLIT_INT, PAN_BLIT_NUM_TYPES, }; #define PAN_BLIT_NUM_TARGETS (12) struct pan_blit_shader { mali_ptr shader; uint32_t blend_ret_addr; }; struct pan_blit_shaders { struct panfrost_bo *bo; struct pan_blit_shader loads[PAN_BLIT_NUM_TARGETS][PAN_BLIT_NUM_TYPES][2]; }; typedef uint32_t mali_pixel_format; struct panfrost_format { mali_pixel_format hw; unsigned bind; }; struct panfrost_device { /* For ralloc */ void *memctx; int fd; /* Properties of the GPU in use */ unsigned arch; unsigned gpu_id; unsigned core_count; unsigned thread_tls_alloc; unsigned quirks; /* Table of formats, indexed by a PIPE format */ const struct panfrost_format *formats; /* Bitmask of supported compressed texture formats */ uint32_t compressed_formats; /* debug flags, see pan_util.h how to interpret */ unsigned debug; drmVersionPtr kernel_version; struct renderonly *ro; pthread_mutex_t bo_map_lock; struct util_sparse_array bo_map; struct { pthread_mutex_t lock; /* List containing all cached BOs sorted in LRU (Least * Recently Used) order. This allows us to quickly evict BOs * that are more than 1 second old. */ struct list_head lru; /* The BO cache is a set of buckets with power-of-two sizes * ranging from 2^12 (4096, the page size) to * 2^(12 + MAX_BO_CACHE_BUCKETS). * Each bucket is a linked list of free panfrost_bo objects. */ struct list_head buckets[NR_BO_CACHE_BUCKETS]; } bo_cache; struct pan_blit_shaders blit_shaders; /* Tiler heap shared across all tiler jobs, allocated against the * device since there's only a single tiler. Since this is invisible to * the CPU, it's okay for multiple contexts to reference it * simultaneously; by keeping on the device struct, we eliminate a * costly per-context allocation. */ struct panfrost_bo *tiler_heap; }; void panfrost_open_device(void *memctx, int fd, struct panfrost_device *dev); void panfrost_close_device(struct panfrost_device *dev); bool panfrost_supports_compressed_format(struct panfrost_device *dev, unsigned fmt); static inline struct panfrost_bo * pan_lookup_bo(struct panfrost_device *dev, uint32_t gem_handle) { return util_sparse_array_get(&dev->bo_map, gem_handle); } #endif