/* * Copyright © 2012 Rob Clark * SPDX-License-Identifier: MIT * * Authors: * Rob Clark */ #ifndef FREEDRENO_RESOURCE_H_ #define FREEDRENO_RESOURCE_H_ #include "util/list.h" #include "util/simple_mtx.h" #include "util/u_dump.h" #include "util/u_range.h" #include "util/u_transfer_helper.h" #include "freedreno/fdl/freedreno_layout.h" #include "freedreno_batch.h" #include "freedreno_util.h" BEGINC; #define PRSC_FMT \ "p: target=%s, format=%s, %ux%ux%u, " \ "array_size=%u, last_level=%u, " \ "nr_samples=%u, usage=%u, bind=%x, flags=%x" #define PRSC_ARGS(p) \ (p), util_str_tex_target((p)->target, true), \ util_format_short_name((p)->format), (p)->width0, (p)->height0, \ (p)->depth0, (p)->array_size, (p)->last_level, (p)->nr_samples, \ (p)->usage, (p)->bind, (p)->flags enum fd_lrz_direction { FD_LRZ_UNKNOWN, /* Depth func less/less-than: */ FD_LRZ_LESS, /* Depth func greater/greater-than: */ FD_LRZ_GREATER, }; /** * State related to batch/resource tracking. * * With threaded_context we need to support replace_buffer_storage, in * which case we can end up in transfer_map with tres->latest, but other * pipe_context APIs using the original prsc pointer. This allows TC to * not have to synchronize the front-end thread with the buffer storage * replacement called on driver thread. But it complicates the batch/ * resource tracking. * * To handle this, we need to split the tracking out into it's own ref- * counted structure, so as needed both "versions" of the resource can * point to the same tracking. * * We could *almost* just push this down to fd_bo, except for a3xx/a4xx * hw queries, where we don't know up-front the size to allocate for * per-tile query results. */ struct fd_resource_tracking { struct pipe_reference reference; /* bitmask of in-flight batches which reference this resource. Note * that the batch doesn't hold reference to resources (but instead * the fd_ringbuffer holds refs to the underlying fd_bo), but in case * the resource is destroyed we need to clean up the batch's weak * references to us. */ uint32_t batch_mask; /* reference to batch that writes this resource: */ struct fd_batch *write_batch; /* Set of batches whose batch-cache key references this resource. * We need to track this to know which batch-cache entries to * invalidate if, for example, the resource is invalidated or * shadowed. */ uint32_t bc_batch_mask; }; void __fd_resource_tracking_destroy(struct fd_resource_tracking *track); static inline void fd_resource_tracking_reference(struct fd_resource_tracking **ptr, struct fd_resource_tracking *track) { struct fd_resource_tracking *old_track = *ptr; if (pipe_reference(&(*ptr)->reference, &track->reference)) { assert(!old_track->write_batch); free(old_track); } *ptr = track; } /** * A resource (any buffer/texture/image/etc) */ struct fd_resource { struct threaded_resource b; struct fd_bo *bo; /* use fd_resource_set_bo() to write */ enum pipe_format internal_format; uint32_t hash; /* _mesa_hash_pointer() on this resource's address. */ struct fdl_layout layout; /* buffer range that has been initialized */ struct util_range valid_buffer_range; bool valid; struct renderonly_scanout *scanout; /* reference to the resource holding stencil data for a z32_s8 texture */ /* TODO rename to secondary or auxiliary? */ struct fd_resource *stencil; struct fd_resource_tracking *track; simple_mtx_t lock; /* bitmask of state this resource could potentially dirty when rebound, * see rebind_resource() */ BITMASK_ENUM(fd_dirty_3d_state) dirty; /* Sequence # incremented each time bo changes: */ uint16_t seqno; /* Is this buffer a replacement created by threaded_context to avoid * a stall in PIPE_MAP_DISCARD_WHOLE_RESOURCE|PIPE_MAP_WRITE case? * If so, it no longer "owns" it's rsc->track, and so should not * invalidate when the rsc is destroyed. */ bool is_replacement : 1; /* Uninitialized resources with UBWC format need their UBWC flag data * cleared before writes, as the UBWC state is read and used during * writes, so undefined UBWC flag data results in undefined results. */ bool needs_ubwc_clear : 1; /* * LRZ * * TODO lrz width/height/pitch should probably also move to * fdl_layout */ bool lrz_valid : 1; enum fd_lrz_direction lrz_direction : 2; uint16_t lrz_width; // for lrz clear, does this differ from lrz_pitch? uint16_t lrz_height; uint16_t lrz_pitch; uint32_t lrz_fc_offset; struct fd_bo *lrz; }; struct fd_memory_object { struct pipe_memory_object b; struct fd_bo *bo; }; static inline struct fd_resource * fd_resource(struct pipe_resource *ptex) { return (struct fd_resource *)ptex; } static inline struct fd_memory_object * fd_memory_object(struct pipe_memory_object *pmemobj) { return (struct fd_memory_object *)pmemobj; } static inline bool pending(struct fd_resource *rsc, bool write) { /* if we have a pending GPU write, we are busy in any case: */ if (rsc->track->write_batch) return true; /* if CPU wants to write, but we are pending a GPU read, we are busy: */ if (write && rsc->track->batch_mask) return true; if (rsc->stencil && pending(rsc->stencil, write)) return true; return false; } static inline bool resource_busy(struct fd_resource *rsc, unsigned op) { return fd_bo_cpu_prep(rsc->bo, NULL, op | FD_BO_PREP_NOSYNC) != 0; } int __fd_resource_wait(struct fd_context *ctx, struct fd_resource *rsc, unsigned op, const char *func); #define fd_resource_wait(ctx, rsc, op) ({ \ MESA_TRACE_FUNC(); \ __fd_resource_wait(ctx, rsc, op, __func__); \ }) static inline void fd_resource_lock(struct fd_resource *rsc) { simple_mtx_lock(&rsc->lock); } static inline void fd_resource_unlock(struct fd_resource *rsc) { simple_mtx_unlock(&rsc->lock); } static inline void fd_resource_set_usage(struct pipe_resource *prsc, enum fd_dirty_3d_state usage) { if (!prsc) return; struct fd_resource *rsc = fd_resource(prsc); /* Bits are only ever ORed in, and we expect many set_usage() per * resource, so do the quick check outside of the lock. */ if (likely(rsc->dirty & usage)) return; fd_resource_lock(rsc); rsc->dirty |= usage; fd_resource_unlock(rsc); } static inline bool has_depth(enum pipe_format format) { const struct util_format_description *desc = util_format_description(format); return util_format_has_depth(desc); } static inline bool is_z32(enum pipe_format format) { switch (format) { case PIPE_FORMAT_Z32_FLOAT_S8X24_UINT: case PIPE_FORMAT_Z32_UNORM: case PIPE_FORMAT_Z32_FLOAT: return true; default: return false; } } struct fd_transfer { struct threaded_transfer b; struct pipe_resource *staging_prsc; struct pipe_box staging_box; void *upload_ptr; }; static inline struct fd_transfer * fd_transfer(struct pipe_transfer *ptrans) { return (struct fd_transfer *)ptrans; } static inline struct fdl_slice * fd_resource_slice(struct fd_resource *rsc, unsigned level) { assert(level <= rsc->b.b.last_level); return &rsc->layout.slices[level]; } static inline uint32_t fd_resource_layer_stride(struct fd_resource *rsc, unsigned level) { return fdl_layer_stride(&rsc->layout, level); } /* get pitch (in bytes) for specified mipmap level */ static inline uint32_t fd_resource_pitch(struct fd_resource *rsc, unsigned level) { if (is_a2xx(fd_screen(rsc->b.b.screen))) return fdl2_pitch(&rsc->layout, level); return fdl_pitch(&rsc->layout, level); } /* get offset for specified mipmap level and texture/array layer */ static inline uint32_t fd_resource_offset(struct fd_resource *rsc, unsigned level, unsigned layer) { uint32_t offset = fdl_surface_offset(&rsc->layout, level, layer); assert(offset < fd_bo_size(rsc->bo)); return offset; } static inline uint32_t fd_resource_ubwc_offset(struct fd_resource *rsc, unsigned level, unsigned layer) { uint32_t offset = fdl_ubwc_offset(&rsc->layout, level, layer); assert(offset < fd_bo_size(rsc->bo)); return offset; } static inline uint32_t fd_resource_tile_mode(struct pipe_resource *prsc, int level) { return fdl_tile_mode(&fd_resource(prsc)->layout, level); } static inline const char * fd_resource_tile_mode_desc(const struct fd_resource *rsc, int level) { return fdl_tile_mode_desc(&rsc->layout, level); } static inline bool fd_resource_ubwc_enabled(struct fd_resource *rsc, int level) { return fdl_ubwc_enabled(&rsc->layout, level); } /* access # of samples, with 0 normalized to 1 (which is what we care about * most of the time) */ static inline unsigned fd_resource_nr_samples(const struct pipe_resource *prsc) { return MAX2(1, prsc->nr_samples); } void fd_resource_screen_init(struct pipe_screen *pscreen); void fd_resource_context_init(struct pipe_context *pctx); uint32_t fd_setup_slices(struct fd_resource *rsc); void fd_resource_resize(struct pipe_resource *prsc, uint32_t sz); void fd_replace_buffer_storage(struct pipe_context *ctx, struct pipe_resource *dst, struct pipe_resource *src, unsigned num_rebinds, uint32_t rebind_mask, uint32_t delete_buffer_id) in_dt; bool fd_resource_busy(struct pipe_screen *pscreen, struct pipe_resource *prsc, unsigned usage); void fd_resource_uncompress(struct fd_context *ctx, struct fd_resource *rsc, bool linear) assert_dt; void fd_resource_dump(struct fd_resource *rsc, const char *name); bool fd_render_condition_check(struct pipe_context *pctx) assert_dt; static inline bool fd_batch_references_resource(struct fd_batch *batch, struct fd_resource *rsc) { return rsc->track->batch_mask & (1 << batch->idx); } static inline void fd_batch_write_prep(struct fd_batch *batch, struct fd_resource *rsc) assert_dt { if (unlikely(rsc->needs_ubwc_clear)) { batch->ctx->clear_ubwc(batch, rsc); rsc->needs_ubwc_clear = false; } } static inline void fd_batch_resource_read(struct fd_batch *batch, struct fd_resource *rsc) assert_dt { /* Fast path: if we hit this then we know we don't have anyone else * writing to it (since both _write and _read flush other writers), and * that we've already recursed for stencil. */ if (unlikely(!fd_batch_references_resource(batch, rsc))) fd_batch_resource_read_slowpath(batch, rsc); } static inline bool needs_dirty_resource(struct fd_context *ctx, struct pipe_resource *prsc, bool write) assert_dt { if (!prsc) return false; struct fd_resource *rsc = fd_resource(prsc); /* Switching between draw and non_draw will dirty all state, so if * we pick the wrong one, all the bits in the dirty_resource state * will be set anyways.. so no harm, no foul. */ struct fd_batch *batch = ctx->batch_nondraw ? ctx->batch_nondraw : ctx->batch; if (!batch) return false; if (write) return rsc->track->write_batch != batch; return !fd_batch_references_resource(batch, rsc); } static inline void fd_dirty_resource(struct fd_context *ctx, struct pipe_resource *prsc, BITMASK_ENUM(fd_dirty_3d_state) dirty, bool write) assert_dt { fd_context_dirty(ctx, dirty); if (ctx->dirty_resource & dirty) return; if (!needs_dirty_resource(ctx, prsc, write)) return; ctx->dirty_resource |= dirty; } static inline void fd_dirty_shader_resource(struct fd_context *ctx, struct pipe_resource *prsc, enum pipe_shader_type shader, BITMASK_ENUM(fd_dirty_shader_state) dirty, bool write) assert_dt { fd_context_dirty_shader(ctx, shader, dirty); if (ctx->dirty_shader_resource[shader] & dirty) return; if (!needs_dirty_resource(ctx, prsc, write)) return; ctx->dirty_shader_resource[shader] |= dirty; ctx->dirty_resource |= dirty_shader_to_dirty_state(dirty); } static inline enum fdl_view_type fdl_type_from_pipe_target(enum pipe_texture_target target) { switch (target) { case PIPE_TEXTURE_1D: case PIPE_TEXTURE_1D_ARRAY: return FDL_VIEW_TYPE_1D; case PIPE_TEXTURE_2D: case PIPE_TEXTURE_RECT: case PIPE_TEXTURE_2D_ARRAY: return FDL_VIEW_TYPE_2D; case PIPE_TEXTURE_CUBE: case PIPE_TEXTURE_CUBE_ARRAY: return FDL_VIEW_TYPE_CUBE; case PIPE_TEXTURE_3D: return FDL_VIEW_TYPE_3D; case PIPE_MAX_TEXTURE_TYPES: default: unreachable("bad texture type"); } } ENDC; #endif /* FREEDRENO_RESOURCE_H_ */