/* * Copyright (C) 2014-2015 Etnaviv Project * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * * Authors: * Christian Gmeiner */ #ifndef ETNAVIV_PRIV_H_ #define ETNAVIV_PRIV_H_ #include #include #include #include #include #include #include #include #include #include #include "util/list.h" #include "util/macros.h" #include "util/simple_mtx.h" #include "util/timespec.h" #include "util/u_atomic.h" #include "util/u_debug.h" #include "util/vma.h" #include "etnaviv_drmif.h" #include "drm-uapi/etnaviv_drm.h" extern simple_mtx_t etna_drm_table_lock; struct etna_bo_bucket { uint32_t size; struct list_head list; }; struct etna_bo_cache { struct etna_bo_bucket cache_bucket[14 * 4]; unsigned num_buckets; time_t time; }; struct etna_device { int fd; uint32_t drm_version; int refcnt; /* tables to keep track of bo's, to avoid "evil-twin" etna_bo objects: * * handle_table: maps handle to etna_bo * name_table: maps flink name to etna_bo * * We end up needing two tables, because DRM_IOCTL_GEM_OPEN always * returns a new handle. So we need to figure out if the bo is already * open in the process first, before calling gem-open. */ void *handle_table, *name_table; struct etna_bo_cache bo_cache; int use_softpin; struct util_vma_heap address_space; int closefd; /* call close(fd) upon destruction */ }; void etna_bo_cache_init(struct etna_bo_cache *cache); void etna_bo_cache_cleanup(struct etna_bo_cache *cache, time_t time); struct etna_bo *etna_bo_cache_alloc(struct etna_bo_cache *cache, uint32_t *size, uint32_t flags); int etna_bo_cache_free(struct etna_bo_cache *cache, struct etna_bo *bo); /* for where @etna_drm_table_lock is already held: */ void etna_device_del_locked(struct etna_device *dev); /* a GEM buffer object allocated from the DRM device */ struct etna_bo { struct etna_device *dev; void *map; /* userspace mmap'ing (if there is one) */ uint32_t size; uint32_t handle; uint32_t flags; uint32_t name; /* flink global handle (DRI2 name) */ uint64_t offset; /* offset to mmap() */ uint32_t va; /* GPU virtual address */ int refcnt; /* * To avoid excess hashtable lookups, cache the stream this bo was * last emitted on (since that will probably also be the next ring * it is emitted on). */ struct etna_cmd_stream *current_stream; uint32_t idx; int reuse; struct list_head list; /* bucket-list entry */ time_t free_time; /* time when added to bucket-list */ }; struct etna_gpu { struct etna_device *dev; uint32_t core; uint32_t model; uint32_t revision; }; struct etna_pipe { enum etna_pipe_id id; struct etna_gpu *gpu; }; struct etna_cmd_stream_priv { struct etna_cmd_stream base; struct etna_pipe *pipe; uint32_t last_timestamp; /* submit ioctl related tables: */ struct { /* bo's table: */ struct drm_etnaviv_gem_submit_bo *bos; uint32_t nr_bos, max_bos; /* reloc's table: */ struct drm_etnaviv_gem_submit_reloc *relocs; uint32_t nr_relocs, max_relocs; /* perf's table: */ struct drm_etnaviv_gem_submit_pmr *pmrs; uint32_t nr_pmrs, max_pmrs; } submit; /* should have matching entries in submit.bos: */ struct etna_bo **bos; uint32_t nr_bos, max_bos; /* notify callback if buffer reset happened */ void (*force_flush)(struct etna_cmd_stream *stream, void *priv); void *force_flush_priv; void *bo_table; }; struct etna_perfmon { struct list_head domains; struct etna_pipe *pipe; }; struct etna_perfmon_domain { struct list_head head; struct list_head signals; uint8_t id; char name[64]; }; struct etna_perfmon_signal { struct list_head head; struct etna_perfmon_domain *domain; uint8_t signal; char name[64]; }; #define ALIGN(v,a) (((v) + (a) - 1) & ~((a) - 1)) #define enable_debug 0 /* TODO make dynamic */ #define INFO_MSG(fmt, ...) \ do { debug_printf("[I] "fmt " (%s:%d)\n", \ ##__VA_ARGS__, __FUNCTION__, __LINE__); } while (0) #define DEBUG_MSG(fmt, ...) \ do if (enable_debug) { debug_printf("[D] "fmt " (%s:%d)\n", \ ##__VA_ARGS__, __FUNCTION__, __LINE__); } while (0) #define WARN_MSG(fmt, ...) \ do { debug_printf("[W] "fmt " (%s:%d)\n", \ ##__VA_ARGS__, __FUNCTION__, __LINE__); } while (0) #define ERROR_MSG(fmt, ...) \ do { debug_printf("[E] " fmt " (%s:%d)\n", \ ##__VA_ARGS__, __FUNCTION__, __LINE__); } while (0) #define VOID2U64(x) ((uint64_t)(unsigned long)(x)) static inline void get_abs_timeout(struct drm_etnaviv_timespec *tv, uint64_t ns) { struct timespec t; clock_gettime(CLOCK_MONOTONIC, &t); tv->tv_sec = t.tv_sec + ns / NSEC_PER_SEC; tv->tv_nsec = t.tv_nsec + ns % NSEC_PER_SEC; if (tv->tv_nsec >= NSEC_PER_SEC) { tv->tv_nsec -= NSEC_PER_SEC; tv->tv_sec++; } } #if HAVE_VALGRIND # include /* * For tracking the backing memory (if valgrind enabled, we force a mmap * for the purposes of tracking) */ static inline void VG_BO_ALLOC(struct etna_bo *bo) { if (bo && RUNNING_ON_VALGRIND) { VALGRIND_MALLOCLIKE_BLOCK(etna_bo_map(bo), bo->size, 0, 1); } } static inline void VG_BO_FREE(struct etna_bo *bo) { VALGRIND_FREELIKE_BLOCK(bo->map, 0); } /* * For tracking bo structs that are in the buffer-cache, so that valgrind * doesn't attribute ownership to the first one to allocate the recycled * bo. * * Note that the list_head in etna_bo is used to track the buffers in cache * so disable error reporting on the range while they are in cache so * valgrind doesn't squawk about list traversal. * */ static inline void VG_BO_RELEASE(struct etna_bo *bo) { if (RUNNING_ON_VALGRIND) { VALGRIND_DISABLE_ADDR_ERROR_REPORTING_IN_RANGE(bo, sizeof(*bo)); VALGRIND_MAKE_MEM_NOACCESS(bo, sizeof(*bo)); VALGRIND_FREELIKE_BLOCK(bo->map, 0); } } static inline void VG_BO_OBTAIN(struct etna_bo *bo) { if (RUNNING_ON_VALGRIND) { VALGRIND_MAKE_MEM_DEFINED(bo, sizeof(*bo)); VALGRIND_ENABLE_ADDR_ERROR_REPORTING_IN_RANGE(bo, sizeof(*bo)); VALGRIND_MALLOCLIKE_BLOCK(bo->map, bo->size, 0, 1); } } #else static inline void VG_BO_ALLOC(struct etna_bo *bo) {} static inline void VG_BO_FREE(struct etna_bo *bo) {} static inline void VG_BO_RELEASE(struct etna_bo *bo) {} static inline void VG_BO_OBTAIN(struct etna_bo *bo) {} #endif #endif /* ETNAVIV_PRIV_H_ */