/* * Copyright (C) 2017-2019 Lyude Paul * Copyright (C) 2017-2019 Alyssa Rosenzweig * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * */ #ifndef __PAN_DECODE_H__ #define __PAN_DECODE_H__ #include "genxml/gen_macros.h" #include "util/rb_tree.h" #include "util/simple_mtx.h" #include "util/u_dynarray.h" #include "wrap.h" struct pandecode_context { int id; /* only used for the filename */ FILE *dump_stream; unsigned indent; struct rb_tree mmap_tree; struct util_dynarray ro_mappings; int dump_frame_count; simple_mtx_t lock; /* On CSF context, set to true if the root CS ring buffer * is managed in userspace. The blob does that, and mesa might use * usermode queues too at some point. */ bool usermode_queue; }; void pandecode_dump_file_open(struct pandecode_context *ctx); struct pandecode_mapped_memory { struct rb_node node; size_t length; void *addr; uint64_t gpu_va; bool ro; char name[32]; }; char *pointer_as_memory_reference(struct pandecode_context *ctx, uint64_t ptr); struct pandecode_mapped_memory * pandecode_find_mapped_gpu_mem_containing(struct pandecode_context *ctx, uint64_t addr); void pandecode_map_read_write(struct pandecode_context *ctx); void pandecode_dump_mappings(struct pandecode_context *ctx); static inline void * __pandecode_fetch_gpu_mem(struct pandecode_context *ctx, uint64_t gpu_va, size_t size, int line, const char *filename) { const struct pandecode_mapped_memory *mem = pandecode_find_mapped_gpu_mem_containing(ctx, gpu_va); if (!mem) { fprintf(stderr, "Access to unknown memory %" PRIx64 " in %s:%d\n", gpu_va, filename, line); assert(0); } assert(size + (gpu_va - mem->gpu_va) <= mem->length); return mem->addr + gpu_va - mem->gpu_va; } #define pandecode_fetch_gpu_mem(ctx, gpu_va, size) \ __pandecode_fetch_gpu_mem(ctx, gpu_va, size, __LINE__, __FILE__) /* Returns a validated pointer to mapped GPU memory with the given pointer type, * size automatically determined from the pointer type */ #define PANDECODE_PTR(ctx, gpu_va, type) \ ((type *)(__pandecode_fetch_gpu_mem(ctx, gpu_va, sizeof(type), __LINE__, \ __FILE__))) /* Usage: PANDECODE_PTR_VAR(name, gpu_va) */ #define PANDECODE_PTR_VAR(ctx, name, gpu_va) \ name = __pandecode_fetch_gpu_mem(ctx, gpu_va, sizeof(*name), __LINE__, \ __FILE__) void pandecode_validate_buffer(struct pandecode_context *ctx, mali_ptr addr, size_t sz); /* Forward declare for all supported gens to permit thunking */ void pandecode_jc_v4(struct pandecode_context *ctx, mali_ptr jc_gpu_va, unsigned gpu_id); void pandecode_jc_v5(struct pandecode_context *ctx, mali_ptr jc_gpu_va, unsigned gpu_id); void pandecode_jc_v6(struct pandecode_context *ctx, mali_ptr jc_gpu_va, unsigned gpu_id); void pandecode_jc_v7(struct pandecode_context *ctx, mali_ptr jc_gpu_va, unsigned gpu_id); void pandecode_jc_v9(struct pandecode_context *ctx, mali_ptr jc_gpu_va, unsigned gpu_id); void pandecode_abort_on_fault_v4(struct pandecode_context *ctx, mali_ptr jc_gpu_va); void pandecode_abort_on_fault_v5(struct pandecode_context *ctx, mali_ptr jc_gpu_va); void pandecode_abort_on_fault_v6(struct pandecode_context *ctx, mali_ptr jc_gpu_va); void pandecode_abort_on_fault_v7(struct pandecode_context *ctx, mali_ptr jc_gpu_va); void pandecode_abort_on_fault_v9(struct pandecode_context *ctx, mali_ptr jc_gpu_va); void pandecode_cs_v10(struct pandecode_context *ctx, mali_ptr queue, uint32_t size, unsigned gpu_id, uint32_t *regs); /* Logging infrastructure */ static void pandecode_make_indent(struct pandecode_context *ctx) { for (unsigned i = 0; i < ctx->indent; ++i) fprintf(ctx->dump_stream, " "); } static inline void PRINTFLIKE(2, 3) pandecode_log(struct pandecode_context *ctx, const char *format, ...) { va_list ap; pandecode_make_indent(ctx); va_start(ap, format); vfprintf(ctx->dump_stream, format, ap); va_end(ap); } static inline void pandecode_log_cont(struct pandecode_context *ctx, const char *format, ...) { va_list ap; va_start(ap, format); vfprintf(ctx->dump_stream, format, ap); va_end(ap); } /* Convenience methods */ #define DUMP_UNPACKED(ctx, T, var, ...) \ { \ pandecode_log(ctx, __VA_ARGS__); \ pan_print(ctx->dump_stream, T, var, (ctx->indent + 1) * 2); \ } #define DUMP_CL(ctx, T, cl, ...) \ { \ pan_unpack(cl, T, temp); \ DUMP_UNPACKED(ctx, T, temp, __VA_ARGS__); \ } #define DUMP_SECTION(ctx, A, S, cl, ...) \ { \ pan_section_unpack(cl, A, S, temp); \ pandecode_log(ctx, __VA_ARGS__); \ pan_section_print(ctx->dump_stream, A, S, temp, (ctx->indent + 1) * 2); \ } #define MAP_ADDR(ctx, T, addr, cl) \ const uint8_t *cl = pandecode_fetch_gpu_mem(ctx, addr, pan_size(T)); #define DUMP_ADDR(ctx, T, addr, ...) \ { \ MAP_ADDR(ctx, T, addr, cl) \ DUMP_CL(ctx, T, cl, __VA_ARGS__); \ } void pandecode_shader_disassemble(struct pandecode_context *ctx, mali_ptr shader_ptr, unsigned gpu_id); #ifdef PAN_ARCH /* Information about the framebuffer passed back for additional analysis */ struct pandecode_fbd { unsigned rt_count; bool has_extra; }; struct pandecode_fbd GENX(pandecode_fbd)(struct pandecode_context *ctx, uint64_t gpu_va, bool is_fragment, unsigned gpu_id); #if PAN_ARCH >= 9 void GENX(pandecode_dcd)(struct pandecode_context *ctx, const struct MALI_DRAW *p, unsigned unused, unsigned gpu_id); #else void GENX(pandecode_dcd)(struct pandecode_context *ctx, const struct MALI_DRAW *p, enum mali_job_type job_type, unsigned gpu_id); #endif #if PAN_ARCH <= 5 void GENX(pandecode_texture)(struct pandecode_context *ctx, mali_ptr u, unsigned tex); #else void GENX(pandecode_texture)(struct pandecode_context *ctx, const void *cl, unsigned tex); #endif #if PAN_ARCH >= 5 mali_ptr GENX(pandecode_blend)(struct pandecode_context *ctx, void *descs, int rt_no, mali_ptr frag_shader); #endif #if PAN_ARCH >= 6 void GENX(pandecode_tiler)(struct pandecode_context *ctx, mali_ptr gpu_va, unsigned gpu_id); #endif #if PAN_ARCH >= 9 void GENX(pandecode_shader_environment)(struct pandecode_context *ctx, const struct MALI_SHADER_ENVIRONMENT *p, unsigned gpu_id); void GENX(pandecode_resource_tables)(struct pandecode_context *ctx, mali_ptr addr, const char *label); void GENX(pandecode_fau)(struct pandecode_context *ctx, mali_ptr addr, unsigned count, const char *name); mali_ptr GENX(pandecode_shader)(struct pandecode_context *ctx, mali_ptr addr, const char *label, unsigned gpu_id); void GENX(pandecode_blend_descs)(struct pandecode_context *ctx, mali_ptr blend, unsigned count, mali_ptr frag_shader, unsigned gpu_id); void GENX(pandecode_depth_stencil)(struct pandecode_context *ctx, mali_ptr addr); #endif #endif #endif /* __MMAP_TRACE_H__ */