2023-08-30 17:31:07 +02:00
|
|
|
// SPDX-License-Identifier: MIT
|
|
|
|
/*
|
|
|
|
* Copyright © 2020 Intel Corporation
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <drm/drm_fourcc.h>
|
|
|
|
|
|
|
|
#include "display/intel_display.h"
|
|
|
|
#include "gem/i915_gem_ioctls.h"
|
|
|
|
#include "gem/i915_gem_lmem.h"
|
|
|
|
#include "gem/i915_gem_region.h"
|
|
|
|
#include "pxp/intel_pxp.h"
|
|
|
|
|
|
|
|
#include "i915_drv.h"
|
|
|
|
#include "i915_gem_create.h"
|
|
|
|
#include "i915_trace.h"
|
|
|
|
#include "i915_user_extensions.h"
|
|
|
|
|
|
|
|
static u32 object_max_page_size(struct intel_memory_region **placements,
|
|
|
|
unsigned int n_placements)
|
|
|
|
{
|
|
|
|
u32 max_page_size = 0;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < n_placements; i++) {
|
|
|
|
struct intel_memory_region *mr = placements[i];
|
|
|
|
|
|
|
|
GEM_BUG_ON(!is_power_of_2(mr->min_page_size));
|
|
|
|
max_page_size = max_t(u32, max_page_size, mr->min_page_size);
|
|
|
|
}
|
|
|
|
|
|
|
|
GEM_BUG_ON(!max_page_size);
|
|
|
|
return max_page_size;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int object_set_placements(struct drm_i915_gem_object *obj,
|
|
|
|
struct intel_memory_region **placements,
|
|
|
|
unsigned int n_placements)
|
|
|
|
{
|
|
|
|
struct intel_memory_region **arr;
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
GEM_BUG_ON(!n_placements);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* For the common case of one memory region, skip storing an
|
|
|
|
* allocated array and just point at the region directly.
|
|
|
|
*/
|
|
|
|
if (n_placements == 1) {
|
|
|
|
struct intel_memory_region *mr = placements[0];
|
|
|
|
struct drm_i915_private *i915 = mr->i915;
|
|
|
|
|
|
|
|
obj->mm.placements = &i915->mm.regions[mr->id];
|
|
|
|
obj->mm.n_placements = 1;
|
|
|
|
} else {
|
|
|
|
arr = kmalloc_array(n_placements,
|
|
|
|
sizeof(struct intel_memory_region *),
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!arr)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
for (i = 0; i < n_placements; i++)
|
|
|
|
arr[i] = placements[i];
|
|
|
|
|
|
|
|
obj->mm.placements = arr;
|
|
|
|
obj->mm.n_placements = n_placements;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int i915_gem_publish(struct drm_i915_gem_object *obj,
|
|
|
|
struct drm_file *file,
|
|
|
|
u64 *size_p,
|
|
|
|
u32 *handle_p)
|
|
|
|
{
|
|
|
|
u64 size = obj->base.size;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = drm_gem_handle_create(file, &obj->base, handle_p);
|
|
|
|
/* drop reference from allocate - handle holds it now */
|
|
|
|
i915_gem_object_put(obj);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
*size_p = size;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drm_i915_gem_object *
|
|
|
|
__i915_gem_object_create_user_ext(struct drm_i915_private *i915, u64 size,
|
|
|
|
struct intel_memory_region **placements,
|
|
|
|
unsigned int n_placements,
|
|
|
|
unsigned int ext_flags)
|
|
|
|
{
|
|
|
|
struct intel_memory_region *mr = placements[0];
|
|
|
|
struct drm_i915_gem_object *obj;
|
|
|
|
unsigned int flags;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
i915_gem_flush_free_objects(i915);
|
|
|
|
|
|
|
|
size = round_up(size, object_max_page_size(placements, n_placements));
|
|
|
|
if (size == 0)
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
|
|
/* For most of the ABI (e.g. mmap) we think in system pages */
|
|
|
|
GEM_BUG_ON(!IS_ALIGNED(size, PAGE_SIZE));
|
|
|
|
|
|
|
|
if (i915_gem_object_size_2big(size))
|
|
|
|
return ERR_PTR(-E2BIG);
|
|
|
|
|
|
|
|
obj = i915_gem_object_alloc();
|
|
|
|
if (!obj)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
ret = object_set_placements(obj, placements, n_placements);
|
|
|
|
if (ret)
|
|
|
|
goto object_free;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* I915_BO_ALLOC_USER will make sure the object is cleared before
|
|
|
|
* any user access.
|
|
|
|
*/
|
|
|
|
flags = I915_BO_ALLOC_USER;
|
|
|
|
|
|
|
|
ret = mr->ops->init_object(mr, obj, I915_BO_INVALID_OFFSET, size, 0, flags);
|
|
|
|
if (ret)
|
|
|
|
goto object_free;
|
|
|
|
|
|
|
|
GEM_BUG_ON(size != obj->base.size);
|
|
|
|
|
|
|
|
/* Add any flag set by create_ext options */
|
|
|
|
obj->flags |= ext_flags;
|
|
|
|
|
|
|
|
trace_i915_gem_object_create(obj);
|
|
|
|
return obj;
|
|
|
|
|
|
|
|
object_free:
|
|
|
|
if (obj->mm.n_placements > 1)
|
|
|
|
kfree(obj->mm.placements);
|
|
|
|
i915_gem_object_free(obj);
|
|
|
|
return ERR_PTR(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2023-10-24 12:59:35 +02:00
|
|
|
* __i915_gem_object_create_user - Creates a new object using the same path as
|
|
|
|
* DRM_I915_GEM_CREATE_EXT
|
2023-08-30 17:31:07 +02:00
|
|
|
* @i915: i915 private
|
|
|
|
* @size: size of the buffer, in bytes
|
|
|
|
* @placements: possible placement regions, in priority order
|
|
|
|
* @n_placements: number of possible placement regions
|
|
|
|
*
|
|
|
|
* This function is exposed primarily for selftests and does very little
|
|
|
|
* error checking. It is assumed that the set of placement regions has
|
|
|
|
* already been verified to be valid.
|
|
|
|
*/
|
|
|
|
struct drm_i915_gem_object *
|
|
|
|
__i915_gem_object_create_user(struct drm_i915_private *i915, u64 size,
|
|
|
|
struct intel_memory_region **placements,
|
|
|
|
unsigned int n_placements)
|
|
|
|
{
|
|
|
|
return __i915_gem_object_create_user_ext(i915, size, placements,
|
|
|
|
n_placements, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
i915_gem_dumb_create(struct drm_file *file,
|
|
|
|
struct drm_device *dev,
|
|
|
|
struct drm_mode_create_dumb *args)
|
|
|
|
{
|
|
|
|
struct drm_i915_gem_object *obj;
|
|
|
|
struct intel_memory_region *mr;
|
|
|
|
enum intel_memory_type mem_type;
|
|
|
|
int cpp = DIV_ROUND_UP(args->bpp, 8);
|
|
|
|
u32 format;
|
|
|
|
|
|
|
|
switch (cpp) {
|
|
|
|
case 1:
|
|
|
|
format = DRM_FORMAT_C8;
|
|
|
|
break;
|
|
|
|
case 2:
|
|
|
|
format = DRM_FORMAT_RGB565;
|
|
|
|
break;
|
|
|
|
case 4:
|
|
|
|
format = DRM_FORMAT_XRGB8888;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* have to work out size/pitch and return them */
|
|
|
|
args->pitch = ALIGN(args->width * cpp, 64);
|
|
|
|
|
|
|
|
/* align stride to page size so that we can remap */
|
|
|
|
if (args->pitch > intel_plane_fb_max_stride(to_i915(dev), format,
|
|
|
|
DRM_FORMAT_MOD_LINEAR))
|
|
|
|
args->pitch = ALIGN(args->pitch, 4096);
|
|
|
|
|
|
|
|
if (args->pitch < args->width)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
args->size = mul_u32_u32(args->pitch, args->height);
|
|
|
|
|
|
|
|
mem_type = INTEL_MEMORY_SYSTEM;
|
|
|
|
if (HAS_LMEM(to_i915(dev)))
|
|
|
|
mem_type = INTEL_MEMORY_LOCAL;
|
|
|
|
|
|
|
|
mr = intel_memory_region_by_type(to_i915(dev), mem_type);
|
|
|
|
|
|
|
|
obj = __i915_gem_object_create_user(to_i915(dev), args->size, &mr, 1);
|
|
|
|
if (IS_ERR(obj))
|
|
|
|
return PTR_ERR(obj);
|
|
|
|
|
|
|
|
return i915_gem_publish(obj, file, &args->size, &args->handle);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2023-10-24 12:59:35 +02:00
|
|
|
* i915_gem_create_ioctl - Creates a new mm object and returns a handle to it.
|
2023-08-30 17:31:07 +02:00
|
|
|
* @dev: drm device pointer
|
|
|
|
* @data: ioctl data blob
|
|
|
|
* @file: drm file pointer
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
i915_gem_create_ioctl(struct drm_device *dev, void *data,
|
|
|
|
struct drm_file *file)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *i915 = to_i915(dev);
|
|
|
|
struct drm_i915_gem_create *args = data;
|
|
|
|
struct drm_i915_gem_object *obj;
|
|
|
|
struct intel_memory_region *mr;
|
|
|
|
|
|
|
|
mr = intel_memory_region_by_type(i915, INTEL_MEMORY_SYSTEM);
|
|
|
|
|
|
|
|
obj = __i915_gem_object_create_user(i915, args->size, &mr, 1);
|
|
|
|
if (IS_ERR(obj))
|
|
|
|
return PTR_ERR(obj);
|
|
|
|
|
|
|
|
return i915_gem_publish(obj, file, &args->size, &args->handle);
|
|
|
|
}
|
|
|
|
|
|
|
|
struct create_ext {
|
|
|
|
struct drm_i915_private *i915;
|
|
|
|
struct intel_memory_region *placements[INTEL_REGION_UNKNOWN];
|
|
|
|
unsigned int n_placements;
|
|
|
|
unsigned int placement_mask;
|
|
|
|
unsigned long flags;
|
2023-10-24 12:59:35 +02:00
|
|
|
unsigned int pat_index;
|
2023-08-30 17:31:07 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
static void repr_placements(char *buf, size_t size,
|
|
|
|
struct intel_memory_region **placements,
|
|
|
|
int n_placements)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
buf[0] = '\0';
|
|
|
|
|
|
|
|
for (i = 0; i < n_placements; i++) {
|
|
|
|
struct intel_memory_region *mr = placements[i];
|
|
|
|
int r;
|
|
|
|
|
|
|
|
r = snprintf(buf, size, "\n %s -> { class: %d, inst: %d }",
|
|
|
|
mr->name, mr->type, mr->instance);
|
|
|
|
if (r >= size)
|
|
|
|
return;
|
|
|
|
|
|
|
|
buf += r;
|
|
|
|
size -= r;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int set_placements(struct drm_i915_gem_create_ext_memory_regions *args,
|
|
|
|
struct create_ext *ext_data)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *i915 = ext_data->i915;
|
|
|
|
struct drm_i915_gem_memory_class_instance __user *uregions =
|
|
|
|
u64_to_user_ptr(args->regions);
|
|
|
|
struct intel_memory_region *placements[INTEL_REGION_UNKNOWN];
|
|
|
|
u32 mask;
|
|
|
|
int i, ret = 0;
|
|
|
|
|
|
|
|
if (args->pad) {
|
|
|
|
drm_dbg(&i915->drm, "pad should be zero\n");
|
|
|
|
ret = -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!args->num_regions) {
|
|
|
|
drm_dbg(&i915->drm, "num_regions is zero\n");
|
|
|
|
ret = -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(i915->mm.regions) != ARRAY_SIZE(placements));
|
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(ext_data->placements) != ARRAY_SIZE(placements));
|
|
|
|
if (args->num_regions > ARRAY_SIZE(i915->mm.regions)) {
|
|
|
|
drm_dbg(&i915->drm, "num_regions is too large\n");
|
|
|
|
ret = -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
mask = 0;
|
|
|
|
for (i = 0; i < args->num_regions; i++) {
|
|
|
|
struct drm_i915_gem_memory_class_instance region;
|
|
|
|
struct intel_memory_region *mr;
|
|
|
|
|
|
|
|
if (copy_from_user(®ion, uregions, sizeof(region)))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
mr = intel_memory_region_lookup(i915,
|
|
|
|
region.memory_class,
|
|
|
|
region.memory_instance);
|
|
|
|
if (!mr || mr->private) {
|
|
|
|
drm_dbg(&i915->drm, "Device is missing region { class: %d, inst: %d } at index = %d\n",
|
|
|
|
region.memory_class, region.memory_instance, i);
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto out_dump;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (mask & BIT(mr->id)) {
|
|
|
|
drm_dbg(&i915->drm, "Found duplicate placement %s -> { class: %d, inst: %d } at index = %d\n",
|
|
|
|
mr->name, region.memory_class,
|
|
|
|
region.memory_instance, i);
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto out_dump;
|
|
|
|
}
|
|
|
|
|
|
|
|
placements[i] = mr;
|
|
|
|
mask |= BIT(mr->id);
|
|
|
|
|
|
|
|
++uregions;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ext_data->n_placements) {
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto out_dump;
|
|
|
|
}
|
|
|
|
|
|
|
|
ext_data->n_placements = args->num_regions;
|
|
|
|
for (i = 0; i < args->num_regions; i++)
|
|
|
|
ext_data->placements[i] = placements[i];
|
|
|
|
|
|
|
|
ext_data->placement_mask = mask;
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_dump:
|
|
|
|
if (1) {
|
|
|
|
char buf[256];
|
|
|
|
|
|
|
|
if (ext_data->n_placements) {
|
|
|
|
repr_placements(buf,
|
|
|
|
sizeof(buf),
|
|
|
|
ext_data->placements,
|
|
|
|
ext_data->n_placements);
|
|
|
|
drm_dbg(&i915->drm,
|
|
|
|
"Placements were already set in previous EXT. Existing placements: %s\n",
|
|
|
|
buf);
|
|
|
|
}
|
|
|
|
|
|
|
|
repr_placements(buf, sizeof(buf), placements, i);
|
|
|
|
drm_dbg(&i915->drm, "New placements(so far validated): %s\n", buf);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ext_set_placements(struct i915_user_extension __user *base,
|
|
|
|
void *data)
|
|
|
|
{
|
|
|
|
struct drm_i915_gem_create_ext_memory_regions ext;
|
|
|
|
|
|
|
|
if (copy_from_user(&ext, base, sizeof(ext)))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
return set_placements(&ext, data);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ext_set_protected(struct i915_user_extension __user *base, void *data)
|
|
|
|
{
|
|
|
|
struct drm_i915_gem_create_ext_protected_content ext;
|
|
|
|
struct create_ext *ext_data = data;
|
|
|
|
|
|
|
|
if (copy_from_user(&ext, base, sizeof(ext)))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
if (ext.flags)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (!intel_pxp_is_enabled(ext_data->i915->pxp))
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
ext_data->flags |= I915_BO_PROTECTED;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2023-10-24 12:59:35 +02:00
|
|
|
static int ext_set_pat(struct i915_user_extension __user *base, void *data)
|
|
|
|
{
|
|
|
|
struct create_ext *ext_data = data;
|
|
|
|
struct drm_i915_private *i915 = ext_data->i915;
|
|
|
|
struct drm_i915_gem_create_ext_set_pat ext;
|
|
|
|
unsigned int max_pat_index;
|
|
|
|
|
|
|
|
BUILD_BUG_ON(sizeof(struct drm_i915_gem_create_ext_set_pat) !=
|
|
|
|
offsetofend(struct drm_i915_gem_create_ext_set_pat, rsvd));
|
|
|
|
|
|
|
|
/* Limiting the extension only to Meteor Lake */
|
|
|
|
if (!IS_METEORLAKE(i915))
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
if (copy_from_user(&ext, base, sizeof(ext)))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
max_pat_index = INTEL_INFO(i915)->max_pat_index;
|
|
|
|
|
|
|
|
if (ext.pat_index > max_pat_index) {
|
|
|
|
drm_dbg(&i915->drm, "PAT index is invalid: %u\n",
|
|
|
|
ext.pat_index);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
ext_data->pat_index = ext.pat_index;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2023-08-30 17:31:07 +02:00
|
|
|
static const i915_user_extension_fn create_extensions[] = {
|
|
|
|
[I915_GEM_CREATE_EXT_MEMORY_REGIONS] = ext_set_placements,
|
|
|
|
[I915_GEM_CREATE_EXT_PROTECTED_CONTENT] = ext_set_protected,
|
2023-10-24 12:59:35 +02:00
|
|
|
[I915_GEM_CREATE_EXT_SET_PAT] = ext_set_pat,
|
2023-08-30 17:31:07 +02:00
|
|
|
};
|
|
|
|
|
2023-10-24 12:59:35 +02:00
|
|
|
#define PAT_INDEX_NOT_SET 0xffff
|
2023-08-30 17:31:07 +02:00
|
|
|
/**
|
2023-10-24 12:59:35 +02:00
|
|
|
* i915_gem_create_ext_ioctl - Creates a new mm object and returns a handle to it.
|
2023-08-30 17:31:07 +02:00
|
|
|
* @dev: drm device pointer
|
|
|
|
* @data: ioctl data blob
|
|
|
|
* @file: drm file pointer
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
i915_gem_create_ext_ioctl(struct drm_device *dev, void *data,
|
|
|
|
struct drm_file *file)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *i915 = to_i915(dev);
|
|
|
|
struct drm_i915_gem_create_ext *args = data;
|
|
|
|
struct create_ext ext_data = { .i915 = i915 };
|
|
|
|
struct drm_i915_gem_object *obj;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (args->flags & ~I915_GEM_CREATE_EXT_FLAG_NEEDS_CPU_ACCESS)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2023-10-24 12:59:35 +02:00
|
|
|
ext_data.pat_index = PAT_INDEX_NOT_SET;
|
2023-08-30 17:31:07 +02:00
|
|
|
ret = i915_user_extensions(u64_to_user_ptr(args->extensions),
|
|
|
|
create_extensions,
|
|
|
|
ARRAY_SIZE(create_extensions),
|
|
|
|
&ext_data);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (!ext_data.n_placements) {
|
|
|
|
ext_data.placements[0] =
|
|
|
|
intel_memory_region_by_type(i915, INTEL_MEMORY_SYSTEM);
|
|
|
|
ext_data.n_placements = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (args->flags & I915_GEM_CREATE_EXT_FLAG_NEEDS_CPU_ACCESS) {
|
|
|
|
if (ext_data.n_placements == 1)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We always need to be able to spill to system memory, if we
|
|
|
|
* can't place in the mappable part of LMEM.
|
|
|
|
*/
|
|
|
|
if (!(ext_data.placement_mask & BIT(INTEL_REGION_SMEM)))
|
|
|
|
return -EINVAL;
|
|
|
|
} else {
|
|
|
|
if (ext_data.n_placements > 1 ||
|
|
|
|
ext_data.placements[0]->type != INTEL_MEMORY_SYSTEM)
|
|
|
|
ext_data.flags |= I915_BO_ALLOC_GPU_ONLY;
|
|
|
|
}
|
|
|
|
|
|
|
|
obj = __i915_gem_object_create_user_ext(i915, args->size,
|
|
|
|
ext_data.placements,
|
|
|
|
ext_data.n_placements,
|
|
|
|
ext_data.flags);
|
|
|
|
if (IS_ERR(obj))
|
|
|
|
return PTR_ERR(obj);
|
|
|
|
|
2023-10-24 12:59:35 +02:00
|
|
|
if (ext_data.pat_index != PAT_INDEX_NOT_SET) {
|
|
|
|
i915_gem_object_set_pat_index(obj, ext_data.pat_index);
|
|
|
|
/* Mark pat_index is set by UMD */
|
|
|
|
obj->pat_set_by_user = true;
|
|
|
|
}
|
|
|
|
|
2023-08-30 17:31:07 +02:00
|
|
|
return i915_gem_publish(obj, file, &args->size, &args->handle);
|
|
|
|
}
|