140 lines
4.7 KiB
C
140 lines
4.7 KiB
C
|
/* SPDX-License-Identifier: MIT */
|
||
|
#ifndef __NVKM_MMU_H__
|
||
|
#define __NVKM_MMU_H__
|
||
|
#include <core/subdev.h>
|
||
|
|
||
|
struct nvkm_vma {
|
||
|
struct list_head head;
|
||
|
struct rb_node tree;
|
||
|
u64 addr;
|
||
|
u64 size:50;
|
||
|
bool mapref:1; /* PTs (de)referenced on (un)map (vs pre-allocated). */
|
||
|
bool sparse:1; /* Unmapped PDEs/PTEs will not trigger MMU faults. */
|
||
|
#define NVKM_VMA_PAGE_NONE 7
|
||
|
u8 page:3; /* Requested page type (index, or NONE for automatic). */
|
||
|
u8 refd:3; /* Current page type (index, or NONE for unreferenced). */
|
||
|
bool used:1; /* Region allocated. */
|
||
|
bool part:1; /* Region was split from an allocated region by map(). */
|
||
|
bool busy:1; /* Region busy (for temporarily preventing user access). */
|
||
|
bool mapped:1; /* Region contains valid pages. */
|
||
|
struct nvkm_memory *memory; /* Memory currently mapped into VMA. */
|
||
|
struct nvkm_tags *tags; /* Compression tag reference. */
|
||
|
};
|
||
|
|
||
|
struct nvkm_vmm {
|
||
|
const struct nvkm_vmm_func *func;
|
||
|
struct nvkm_mmu *mmu;
|
||
|
const char *name;
|
||
|
u32 debug;
|
||
|
struct kref kref;
|
||
|
struct mutex mutex;
|
||
|
|
||
|
u64 start;
|
||
|
u64 limit;
|
||
|
|
||
|
struct nvkm_vmm_pt *pd;
|
||
|
struct list_head join;
|
||
|
|
||
|
struct list_head list;
|
||
|
struct rb_root free;
|
||
|
struct rb_root root;
|
||
|
|
||
|
bool bootstrapped;
|
||
|
atomic_t engref[NVKM_SUBDEV_NR];
|
||
|
|
||
|
dma_addr_t null;
|
||
|
void *nullp;
|
||
|
|
||
|
bool replay;
|
||
|
};
|
||
|
|
||
|
int nvkm_vmm_new(struct nvkm_device *, u64 addr, u64 size, void *argv, u32 argc,
|
||
|
struct lock_class_key *, const char *name, struct nvkm_vmm **);
|
||
|
struct nvkm_vmm *nvkm_vmm_ref(struct nvkm_vmm *);
|
||
|
void nvkm_vmm_unref(struct nvkm_vmm **);
|
||
|
int nvkm_vmm_boot(struct nvkm_vmm *);
|
||
|
int nvkm_vmm_join(struct nvkm_vmm *, struct nvkm_memory *inst);
|
||
|
void nvkm_vmm_part(struct nvkm_vmm *, struct nvkm_memory *inst);
|
||
|
int nvkm_vmm_get(struct nvkm_vmm *, u8 page, u64 size, struct nvkm_vma **);
|
||
|
void nvkm_vmm_put(struct nvkm_vmm *, struct nvkm_vma **);
|
||
|
|
||
|
struct nvkm_vmm_map {
|
||
|
struct nvkm_memory *memory;
|
||
|
u64 offset;
|
||
|
|
||
|
struct nvkm_mm_node *mem;
|
||
|
struct scatterlist *sgl;
|
||
|
dma_addr_t *dma;
|
||
|
u64 *pfn;
|
||
|
u64 off;
|
||
|
|
||
|
const struct nvkm_vmm_page *page;
|
||
|
|
||
|
struct nvkm_tags *tags;
|
||
|
u64 next;
|
||
|
u64 type;
|
||
|
u64 ctag;
|
||
|
};
|
||
|
|
||
|
int nvkm_vmm_map(struct nvkm_vmm *, struct nvkm_vma *, void *argv, u32 argc,
|
||
|
struct nvkm_vmm_map *);
|
||
|
void nvkm_vmm_unmap(struct nvkm_vmm *, struct nvkm_vma *);
|
||
|
|
||
|
struct nvkm_memory *nvkm_umem_search(struct nvkm_client *, u64);
|
||
|
struct nvkm_vmm *nvkm_uvmm_search(struct nvkm_client *, u64 handle);
|
||
|
|
||
|
struct nvkm_mmu {
|
||
|
const struct nvkm_mmu_func *func;
|
||
|
struct nvkm_subdev subdev;
|
||
|
|
||
|
u8 dma_bits;
|
||
|
|
||
|
int heap_nr;
|
||
|
struct {
|
||
|
#define NVKM_MEM_VRAM 0x01
|
||
|
#define NVKM_MEM_HOST 0x02
|
||
|
#define NVKM_MEM_COMP 0x04
|
||
|
#define NVKM_MEM_DISP 0x08
|
||
|
u8 type;
|
||
|
u64 size;
|
||
|
} heap[4];
|
||
|
|
||
|
int type_nr;
|
||
|
struct {
|
||
|
#define NVKM_MEM_KIND 0x10
|
||
|
#define NVKM_MEM_MAPPABLE 0x20
|
||
|
#define NVKM_MEM_COHERENT 0x40
|
||
|
#define NVKM_MEM_UNCACHED 0x80
|
||
|
u8 type;
|
||
|
u8 heap;
|
||
|
} type[16];
|
||
|
|
||
|
struct nvkm_vmm *vmm;
|
||
|
|
||
|
struct {
|
||
|
struct mutex mutex;
|
||
|
struct list_head list;
|
||
|
} ptc, ptp;
|
||
|
|
||
|
struct mutex mutex; /* serialises mmu invalidations */
|
||
|
|
||
|
struct nvkm_device_oclass user;
|
||
|
};
|
||
|
|
||
|
int nv04_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int nv41_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int nv44_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int nv50_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int g84_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int mcp77_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gf100_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gk104_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gk20a_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gm200_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gm20b_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gp100_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gp10b_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int gv100_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
int tu102_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
|
||
|
#endif
|