/* * Copyright © 2016 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ /* * NOTE: The header can be included multiple times, from the same file. */ /* * Gen-specific function declarations. This header must *not* be included * directly. Instead, it is included multiple times by anv_private.h. * * In this header file, the usual genx() macro is available. */ #ifndef ANV_PRIVATE_H #error This file is included by means other than anv_private.h #endif struct intel_sample_positions; struct intel_urb_config; struct anv_async_submit; struct anv_embedded_sampler; struct anv_pipeline_embedded_sampler_binding; struct anv_trtt_bind; typedef struct nir_builder nir_builder; typedef struct nir_shader nir_shader; extern const uint32_t genX(vk_to_intel_cullmode)[]; extern const uint32_t genX(vk_to_intel_front_face)[]; extern const uint32_t genX(vk_to_intel_primitive_type)[]; extern const uint32_t genX(vk_to_intel_compare_op)[]; extern const uint32_t genX(vk_to_intel_stencil_op)[]; extern const uint32_t genX(vk_to_intel_logic_op)[]; extern const uint32_t genX(vk_to_intel_fillmode)[]; void genX(init_physical_device_state)(struct anv_physical_device *device); VkResult genX(init_device_state)(struct anv_device *device); void genX(init_cps_device_state)(struct anv_device *device); nir_shader *genX(load_libanv_shader)(struct anv_device *device, void *mem_ctx); uint32_t genX(call_internal_shader)(nir_builder *b, enum anv_internal_kernel_name shader_name); void genX(set_fast_clear_state)(struct anv_cmd_buffer *cmd_buffer, const struct anv_image *image, const enum isl_format format, union isl_color_value clear_color); void genX(load_image_clear_color)(struct anv_cmd_buffer *cmd_buffer, struct anv_state surface_state, const struct anv_image *image); void genX(cmd_buffer_emit_bt_pool_base_address)(struct anv_cmd_buffer *cmd_buffer); void genX(cmd_buffer_emit_state_base_address)(struct anv_cmd_buffer *cmd_buffer); void genX(cmd_buffer_apply_pipe_flushes)(struct anv_cmd_buffer *cmd_buffer); void genX(cmd_buffer_update_color_aux_op)(struct anv_cmd_buffer *cmd_buffer, enum isl_aux_op aux_op); void genX(cmd_buffer_emit_gfx12_depth_wa)(struct anv_cmd_buffer *cmd_buffer, const struct isl_surf *surf); void genX(cmd_buffer_set_binding_for_gfx8_vb_flush)(struct anv_cmd_buffer *cmd_buffer, int vb_index, struct anv_address vb_address, uint32_t vb_size); void genX(cmd_buffer_update_dirty_vbs_for_gfx8_vb_flush)(struct anv_cmd_buffer *cmd_buffer, uint32_t access_type, uint64_t vb_used); void genX(cmd_buffer_emit_hashing_mode)(struct anv_cmd_buffer *cmd_buffer, unsigned width, unsigned height, unsigned scale); void genX(urb_workaround)(struct anv_cmd_buffer *cmd_buffer, const struct intel_urb_config *urb_cfg); void genX(flush_pipeline_select_3d)(struct anv_cmd_buffer *cmd_buffer); void genX(flush_pipeline_select_gpgpu)(struct anv_cmd_buffer *cmd_buffer); void genX(emit_pipeline_select)(struct anv_batch *batch, uint32_t pipeline, const struct anv_device *device); void genX(apply_task_urb_workaround)(struct anv_cmd_buffer *cmd_buffer); void genX(emit_vertex_input)(struct anv_batch *batch, uint32_t *vertex_element_dws, struct anv_graphics_pipeline *pipeline, const struct vk_vertex_input_state *vi, bool emit_in_pipeline); enum anv_pipe_bits genX(emit_apply_pipe_flushes)(struct anv_batch *batch, struct anv_device *device, uint32_t current_pipeline, enum anv_pipe_bits bits, enum anv_pipe_bits *emitted_flush_bits); void genX(invalidate_aux_map)(struct anv_batch *batch, struct anv_device *device, enum intel_engine_class engine_class, enum anv_pipe_bits bits); #if INTEL_WA_14018283232_GFX_VER void genX(batch_emit_wa_14018283232)(struct anv_batch *batch); static inline void genX(cmd_buffer_ensure_wa_14018283232)(struct anv_cmd_buffer *cmd_buffer, bool toggle) { struct anv_gfx_dynamic_state *hw_state = &cmd_buffer->state.gfx.dyn_state; if (intel_needs_workaround(cmd_buffer->device->info, 14018283232) && hw_state->wa_14018283232_toggle != toggle) { hw_state->wa_14018283232_toggle = toggle; BITSET_SET(hw_state->dirty, ANV_GFX_STATE_WA_14018283232); genX(batch_emit_wa_14018283232)(&cmd_buffer->batch); } } #endif static inline bool genX(cmd_buffer_set_coarse_pixel_active)(struct anv_cmd_buffer *cmd_buffer, enum anv_coarse_pixel_state state) { #if INTEL_WA_18038825448_GFX_VER struct anv_cmd_graphics_state *gfx = &cmd_buffer->state.gfx; if (intel_needs_workaround(cmd_buffer->device->info, 18038825448) && gfx->coarse_pixel_active != state) { gfx->coarse_pixel_active = state; gfx->dirty |= ANV_CMD_DIRTY_COARSE_PIXEL_ACTIVE; return true; } return false; #else return false; #endif } void genX(emit_so_memcpy_init)(struct anv_memcpy_state *state, struct anv_device *device, struct anv_cmd_buffer *cmd_buffer, struct anv_batch *batch); void genX(emit_so_memcpy_fini)(struct anv_memcpy_state *state); void genX(emit_so_memcpy_end)(struct anv_memcpy_state *state); void genX(emit_so_memcpy)(struct anv_memcpy_state *state, struct anv_address dst, struct anv_address src, uint32_t size); void genX(emit_l3_config)(struct anv_batch *batch, const struct anv_device *device, const struct intel_l3_config *cfg); void genX(cmd_buffer_config_l3)(struct anv_cmd_buffer *cmd_buffer, const struct intel_l3_config *cfg); void genX(flush_descriptor_buffers)(struct anv_cmd_buffer *cmd_buffer, struct anv_cmd_pipeline_state *pipe_state); uint32_t genX(cmd_buffer_flush_descriptor_sets)(struct anv_cmd_buffer *cmd_buffer, struct anv_cmd_pipeline_state *pipe_state, const VkShaderStageFlags dirty, struct anv_shader_bin **shaders, uint32_t num_shaders); void genX(cmd_buffer_flush_gfx_hw_state)(struct anv_cmd_buffer *cmd_buffer); anv_cmd_dirty_mask_t genX(cmd_buffer_flush_gfx_runtime_state)(struct anv_cmd_buffer *cmd_buffer); void genX(cmd_buffer_flush_gfx_hw_state)(struct anv_cmd_buffer *cmd_buffer); void genX(cmd_buffer_enable_pma_fix)(struct anv_cmd_buffer *cmd_buffer, bool enable); void genX(cmd_buffer_mark_image_written)(struct anv_cmd_buffer *cmd_buffer, const struct anv_image *image, VkImageAspectFlagBits aspect, enum isl_aux_usage aux_usage, uint32_t level, uint32_t base_layer, uint32_t layer_count); void genX(cmd_emit_conditional_render_predicate)(struct anv_cmd_buffer *cmd_buffer); struct anv_address genX(cmd_buffer_ray_query_globals)(struct anv_cmd_buffer *cmd_buffer); void genX(cmd_buffer_ensure_cfe_state)(struct anv_cmd_buffer *cmd_buffer, uint32_t total_scratch); void genX(emit_urb_setup)(struct anv_device *device, struct anv_batch *batch, const struct intel_l3_config *l3_config, VkShaderStageFlags active_stages, const struct intel_urb_config *urb_cfg_in, struct intel_urb_config *urb_cfg_out, enum intel_urb_deref_block_size *deref_block_size); void genX(emit_sample_pattern)(struct anv_batch *batch, const struct vk_sample_locations_state *sl); void genX(cmd_buffer_so_memcpy)(struct anv_cmd_buffer *cmd_buffer, struct anv_address dst, struct anv_address src, uint32_t size); void genX(cmd_buffer_dispatch_kernel)(struct anv_cmd_buffer *cmd_buffer, struct anv_kernel *kernel, const uint32_t *global_size, /* NULL for indirect */ uint32_t arg_count, const struct anv_kernel_arg *args); void genX(blorp_init_dynamic_states)(struct blorp_context *context); void genX(blorp_exec)(struct blorp_batch *batch, const struct blorp_params *params); void genX(batch_emit_secondary_call)(struct anv_batch *batch, struct anv_device *device, struct anv_address secondary_addr, struct anv_address secondary_return_addr); void *genX(batch_emit_return)(struct anv_batch *batch); void genX(cmd_emit_timestamp)(struct anv_batch *batch, struct anv_device *device, struct anv_address addr, enum anv_timestamp_capture_type type, void *data); void genX(cmd_capture_data)(struct anv_batch *batch, struct anv_device *device, struct anv_address dst_addr, struct anv_address src_addr, uint32_t size_B); void genX(batch_emit_post_3dprimitive_was)(struct anv_batch *batch, const struct anv_device *device, uint32_t primitive_topology, uint32_t vertex_count); void genX(batch_emit_fast_color_dummy_blit)(struct anv_batch *batch, struct anv_device *device); VkPolygonMode genX(raster_polygon_mode)(const struct anv_graphics_pipeline *pipeline, VkPolygonMode polygon_mode, VkPrimitiveTopology primitive_topology); void genX(graphics_pipeline_emit)(struct anv_graphics_pipeline *pipeline, const struct vk_graphics_pipeline_state *state); void genX(compute_pipeline_emit)(struct anv_compute_pipeline *pipeline); void genX(ray_tracing_pipeline_emit)(struct anv_ray_tracing_pipeline *pipeline); #define anv_shader_bin_get_bsr(bin, local_arg_offset) ({ \ assert((local_arg_offset) % 8 == 0); \ const struct brw_bs_prog_data *prog_data = \ brw_bs_prog_data_const(bin->prog_data); \ assert(prog_data->simd_size == 8 || prog_data->simd_size == 16); \ \ (struct GENX(BINDLESS_SHADER_RECORD)) { \ .OffsetToLocalArguments = (local_arg_offset) / 8, \ .BindlessShaderDispatchMode = \ prog_data->simd_size == 16 ? RT_SIMD16 : RT_SIMD8, \ .KernelStartPointer = bin->kernel.offset, \ }; \ }) void genX(batch_set_preemption)(struct anv_batch *batch, const struct intel_device_info *devinfo, uint32_t current_pipeline, bool value); void genX(cmd_buffer_set_preemption)(struct anv_cmd_buffer *cmd_buffer, bool value); void genX(batch_emit_pipe_control)(struct anv_batch *batch, const struct intel_device_info *devinfo, uint32_t current_pipeline, enum anv_pipe_bits bits, const char *reason); void genX(batch_emit_pipe_control_write)(struct anv_batch *batch, const struct intel_device_info *devinfo, uint32_t current_pipeline, uint32_t post_sync_op, struct anv_address address, uint32_t imm_data, enum anv_pipe_bits bits, const char *reason); #define genx_batch_emit_pipe_control(a, b, c, d) \ genX(batch_emit_pipe_control) (a, b, c, d, __func__) #define genx_batch_emit_pipe_control_write(a, b, c, d, e, f, g) \ genX(batch_emit_pipe_control_write) (a, b, c, d, e, f, g, __func__) void genX(batch_emit_breakpoint)(struct anv_batch *batch, struct anv_device *device, bool emit_before_draw); static inline void genX(emit_breakpoint)(struct anv_batch *batch, struct anv_device *device, bool emit_before_draw) { if (INTEL_DEBUG(DEBUG_DRAW_BKP)) genX(batch_emit_breakpoint)(batch, device, emit_before_draw); } void genX(cmd_buffer_begin_companion)(struct anv_cmd_buffer *buffer, VkCommandBufferLevel level); struct anv_state genX(cmd_buffer_begin_companion_rcs_syncpoint)(struct anv_cmd_buffer *cmd_buffer); void genX(cmd_buffer_end_companion_rcs_syncpoint)(struct anv_cmd_buffer *cmd_buffer, struct anv_state syncpoint); void genX(emit_simple_shader_init)(struct anv_simple_shader *state); void genX(emit_simple_shader_dispatch)(struct anv_simple_shader *state, uint32_t num_threads, struct anv_state push_state); struct anv_state genX(simple_shader_alloc_push)(struct anv_simple_shader *state, uint32_t size); struct anv_address genX(simple_shader_push_state_address)(struct anv_simple_shader *state, struct anv_state push_state); void genX(emit_simple_shader_end)(struct anv_simple_shader *state); VkResult genX(init_trtt_context_state)(struct anv_async_submit *submit); void genX(write_trtt_entries)(struct anv_async_submit *submit, struct anv_trtt_bind *l3l2_binds, uint32_t n_l3l2_binds, struct anv_trtt_bind *l1_binds, uint32_t n_l1_binds); void genX(async_submit_end)(struct anv_async_submit *submit); void genX(cmd_buffer_emit_push_descriptor_buffer_surface)(struct anv_cmd_buffer *cmd_buffer, struct anv_descriptor_set *set); void genX(cmd_buffer_emit_push_descriptor_surfaces)(struct anv_cmd_buffer *cmd_buffer, struct anv_descriptor_set *set); static inline VkShaderStageFlags genX(cmd_buffer_flush_push_descriptors)(struct anv_cmd_buffer *cmd_buffer, struct anv_cmd_pipeline_state *state, struct anv_pipeline *pipeline) { if (!pipeline->use_push_descriptor && !pipeline->use_push_descriptor_buffer) return 0; assert(pipeline->layout.push_descriptor_set_index != -1); struct anv_descriptor_set *set = state->descriptors[pipeline->layout.push_descriptor_set_index]; assert(set->is_push); const VkShaderStageFlags push_buffer_dirty = cmd_buffer->state.push_descriptors_dirty & pipeline->use_push_descriptor_buffer; if (push_buffer_dirty) { if (set->desc_surface_state.map == NULL) genX(cmd_buffer_emit_push_descriptor_buffer_surface)(cmd_buffer, set); /* Force the next push descriptor update to allocate a new descriptor set. */ state->push_descriptor.set_used_on_gpu = true; } const VkShaderStageFlags push_descriptor_dirty = cmd_buffer->state.push_descriptors_dirty & pipeline->use_push_descriptor; if (push_descriptor_dirty) { genX(cmd_buffer_emit_push_descriptor_surfaces)(cmd_buffer, set); /* Force the next push descriptor update to allocate a new descriptor set. */ state->push_descriptor.set_used_on_gpu = true; } /* Clear the dirty stages now that we've generated the surface states for * them. */ cmd_buffer->state.push_descriptors_dirty &= ~(push_descriptor_dirty | push_buffer_dirty); /* Return the binding table stages that need to be updated */ return push_buffer_dirty | push_descriptor_dirty; } void genX(emit_embedded_sampler)(struct anv_device *device, struct anv_embedded_sampler *sampler, struct anv_pipeline_embedded_sampler_binding *binding);