/* * Copyright © 2022 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #include "i915/intel_gem.h" #include "common/intel_gem.h" #include "i915/intel_engine.h" #include "drm-uapi/i915_drm.h" bool i915_gem_create_context(int fd, uint32_t *context_id) { struct drm_i915_gem_context_create create = {}; if (intel_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_CREATE, &create)) return false; *context_id = create.ctx_id; return true; } bool i915_gem_destroy_context(int fd, uint32_t context_id) { struct drm_i915_gem_context_destroy destroy = { .ctx_id = context_id, }; return intel_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_DESTROY, &destroy) == 0; } bool i915_gem_create_context_engines(int fd, enum intel_gem_create_context_flags flags, const struct intel_query_engine_info *info, int num_engines, enum intel_engine_class *engine_classes, uint32_t vm_id, uint32_t *context_id) { assert(info != NULL); assert(num_engines <= 64); I915_DEFINE_CONTEXT_PARAM_ENGINES(engines_param, 64); engines_param.extensions = 0; /* For each type of intel_engine_class of interest, we keep track of * the previous engine instance used. */ int last_engine_idx[] = { [INTEL_ENGINE_CLASS_RENDER] = -1, [INTEL_ENGINE_CLASS_COPY] = -1, [INTEL_ENGINE_CLASS_COMPUTE] = -1, [INTEL_ENGINE_CLASS_VIDEO] = -1, }; int engine_counts[] = { [INTEL_ENGINE_CLASS_RENDER] = intel_engines_count(info, INTEL_ENGINE_CLASS_RENDER), [INTEL_ENGINE_CLASS_COPY] = intel_engines_count(info, INTEL_ENGINE_CLASS_COPY), [INTEL_ENGINE_CLASS_COMPUTE] = intel_engines_count(info, INTEL_ENGINE_CLASS_COMPUTE), [INTEL_ENGINE_CLASS_VIDEO] = intel_engines_count(info, INTEL_ENGINE_CLASS_VIDEO), }; /* For each queue, we look for the next instance that matches the class we * need. */ for (int i = 0; i < num_engines; i++) { enum intel_engine_class engine_class = engine_classes[i]; assert(engine_class == INTEL_ENGINE_CLASS_RENDER || engine_class == INTEL_ENGINE_CLASS_COPY || engine_class == INTEL_ENGINE_CLASS_COMPUTE || engine_class == INTEL_ENGINE_CLASS_VIDEO); if (engine_counts[engine_class] <= 0) return false; /* Run through the engines reported by the kernel looking for the next * matching instance. We loop in case we want to create multiple * contexts on an engine instance. */ int engine_instance = -1; for (int i = 0; i < info->num_engines; i++) { int *idx = &last_engine_idx[engine_class]; if (++(*idx) >= info->num_engines) *idx = 0; if (info->engines[*idx].engine_class == engine_class) { engine_instance = info->engines[*idx].engine_instance; break; } } if (engine_instance < 0) return false; engines_param.engines[i].engine_class = intel_engine_class_to_i915(engine_class); engines_param.engines[i].engine_instance = engine_instance; } uint32_t size = sizeof(engines_param.extensions); size += sizeof(engines_param.engines[0]) * num_engines; struct drm_i915_gem_context_create_ext_setparam set_engines = { .param = { .param = I915_CONTEXT_PARAM_ENGINES, .value = (uintptr_t)&engines_param, .size = size, } }; struct drm_i915_gem_context_create_ext_setparam protected_param = { .param = { .param = I915_CONTEXT_PARAM_PROTECTED_CONTENT, .value = flags & INTEL_GEM_CREATE_CONTEXT_EXT_PROTECTED_FLAG, }, }; struct drm_i915_gem_context_create_ext_setparam recoverable_param = { .param = { .param = I915_CONTEXT_PARAM_RECOVERABLE, .value = flags & INTEL_GEM_CREATE_CONTEXT_EXT_RECOVERABLE_FLAG, }, }; struct drm_i915_gem_context_create_ext_setparam low_latency_param = { .param = { .param = I915_CONTEXT_PARAM_LOW_LATENCY, .value = flags & INTEL_GEM_CREATE_CONTEXT_EXT_LOW_LATENCY_FLAG, } }; struct drm_i915_gem_context_create_ext create = { .flags = I915_CONTEXT_CREATE_FLAGS_USE_EXTENSIONS, }; struct drm_i915_gem_context_create_ext_setparam vm_param = { .param = { .param = I915_CONTEXT_PARAM_VM, .value = vm_id, }, }; intel_i915_gem_add_ext(&create.extensions, I915_CONTEXT_CREATE_EXT_SETPARAM, &set_engines.base); intel_i915_gem_add_ext(&create.extensions, I915_CONTEXT_CREATE_EXT_SETPARAM, &recoverable_param.base); if (vm_id != 0) { intel_i915_gem_add_ext(&create.extensions, I915_CONTEXT_CREATE_EXT_SETPARAM, &vm_param.base); } if (flags & INTEL_GEM_CREATE_CONTEXT_EXT_PROTECTED_FLAG) { intel_i915_gem_add_ext(&create.extensions, I915_CONTEXT_CREATE_EXT_SETPARAM, &protected_param.base); } if (flags & INTEL_GEM_CREATE_CONTEXT_EXT_LOW_LATENCY_FLAG) { intel_i915_gem_add_ext(&create.extensions, I915_CONTEXT_CREATE_EXT_SETPARAM, &low_latency_param.base); } if (intel_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_CREATE_EXT, &create) == -1) return false; *context_id = create.ctx_id; return true; } bool i915_gem_set_context_param(int fd, uint32_t context, uint32_t param, uint64_t value) { struct drm_i915_gem_context_param p = { .ctx_id = context, .param = param, .value = value, }; return intel_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_SETPARAM, &p) == 0; } bool i915_gem_get_context_param(int fd, uint32_t context, uint32_t param, uint64_t *value) { struct drm_i915_gem_context_param gp = { .ctx_id = context, .param = param, }; if (intel_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_GETPARAM, &gp)) return false; *value = gp.value; return true; } bool i915_gem_read_render_timestamp(int fd, uint64_t *value) { struct drm_i915_reg_read reg_read = { .offset = RCS_TIMESTAMP | I915_REG_READ_8B_WA, }; int ret = intel_ioctl(fd, DRM_IOCTL_I915_REG_READ, ®_read); if (ret == 0) *value = reg_read.val; return ret == 0; } bool i915_gem_create_context_ext(int fd, enum intel_gem_create_context_flags flags, uint32_t *ctx_id) { struct drm_i915_gem_context_create_ext_setparam recoverable_param = { .param = { .param = I915_CONTEXT_PARAM_RECOVERABLE, .value = flags & INTEL_GEM_CREATE_CONTEXT_EXT_RECOVERABLE_FLAG, }, }; struct drm_i915_gem_context_create_ext_setparam protected_param = { .param = { .param = I915_CONTEXT_PARAM_PROTECTED_CONTENT, .value = flags & INTEL_GEM_CREATE_CONTEXT_EXT_PROTECTED_FLAG, }, }; struct drm_i915_gem_context_create_ext create = { .flags = I915_CONTEXT_CREATE_FLAGS_USE_EXTENSIONS, }; intel_i915_gem_add_ext(&create.extensions, I915_CONTEXT_CREATE_EXT_SETPARAM, &recoverable_param.base); intel_i915_gem_add_ext(&create.extensions, I915_CONTEXT_CREATE_EXT_SETPARAM, &protected_param.base); if (intel_ioctl(fd, DRM_IOCTL_I915_GEM_CONTEXT_CREATE_EXT, &create)) return false; *ctx_id = create.ctx_id; return true; } bool i915_gem_supports_protected_context(int fd) { int val = 0; uint32_t ctx_id; bool ret; errno = 0; if (!i915_gem_get_param(fd, I915_PARAM_PXP_STATUS, &val)) { if (errno == ENODEV) return false; } else { return (val > 0); } /* failed without ENODEV, so older kernels require a creation test */ ret = i915_gem_create_context_ext(fd, INTEL_GEM_CREATE_CONTEXT_EXT_PROTECTED_FLAG, &ctx_id); if (!ret) return ret; i915_gem_destroy_context(fd, ctx_id); return ret; } bool i915_gem_get_param(int fd, uint32_t param, int *value) { drm_i915_getparam_t gp = { .param = param, .value = value, }; return intel_ioctl(fd, DRM_IOCTL_I915_GETPARAM, &gp) == 0; } bool i915_gem_can_render_on_fd(int fd) { int val; return intel_gem_get_param(fd, I915_PARAM_CHIPSET_ID, &val) && val > 0; }