Home
last modified time | relevance | path

Searched refs:nvkm_mmu (Results 1 – 25 of 47) sorted by relevance

12

/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dpriv.h4 #define nvkm_mmu(p) container_of((p), struct nvkm_mmu, subdev) macro
8 struct nvkm_mmu **);
11 struct nvkm_mmu *);
13 struct nvkm_mmu **);
16 void (*dtor)(struct nvkm_mmu *);
17 void (*init)(struct nvkm_mmu *);
27 int (*vram)(struct nvkm_mmu *, int type, u8 page, u64 size,
29 int (*umap)(struct nvkm_mmu *, struct nvkm_memory *, void *argv,
35 int (*ctor)(struct nvkm_mmu *, bool managed, u64 addr, u64 size,
42 const u8 *(*kind)(struct nvkm_mmu *, int *count, u8 *invalid);
[all …]
H A Dbase.c42 nvkm_mmu_ptp_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt *pt) in nvkm_mmu_ptp_put()
65 nvkm_mmu_ptp_get(struct nvkm_mmu *mmu, u32 size, bool zero) in nvkm_mmu_ptp_get()
120 nvkm_mmu_ptc_find(struct nvkm_mmu *mmu, u32 size) in nvkm_mmu_ptc_find()
141 nvkm_mmu_ptc_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt **ppt) in nvkm_mmu_ptc_put()
167 nvkm_mmu_ptc_get(struct nvkm_mmu *mmu, u32 size, u32 align, bool zero) in nvkm_mmu_ptc_get()
220 nvkm_mmu_ptc_dump(struct nvkm_mmu *mmu) in nvkm_mmu_ptc_dump()
234 nvkm_mmu_ptc_fini(struct nvkm_mmu *mmu) in nvkm_mmu_ptc_fini()
246 nvkm_mmu_ptc_init(struct nvkm_mmu *mmu) in nvkm_mmu_ptc_init()
255 nvkm_mmu_type(struct nvkm_mmu *mmu, int heap, u8 type) in nvkm_mmu_type()
265 nvkm_mmu_heap(struct nvkm_mmu *mmu, u8 type, u64 size) in nvkm_mmu_heap()
[all …]
H A Dmem.h5 int nvkm_mem_new_type(struct nvkm_mmu *, int type, u8 page, u64 size,
9 int nv04_mem_new(struct nvkm_mmu *, int, u8, u64, void *, u32,
11 int nv04_mem_map(struct nvkm_mmu *, struct nvkm_memory *, void *, u32,
14 int nv50_mem_new(struct nvkm_mmu *, int, u8, u64, void *, u32,
16 int nv50_mem_map(struct nvkm_mmu *, struct nvkm_memory *, void *, u32,
19 int gf100_mem_new(struct nvkm_mmu *, int, u8, u64, void *, u32,
21 int gf100_mem_map(struct nvkm_mmu *, struct nvkm_memory *, void *, u32,
H A Dummu.c35 struct nvkm_mmu *mmu = nvkm_ummu(object)->mmu; in nvkm_ummu_sclass()
59 struct nvkm_mmu *mmu = ummu->mmu; in nvkm_ummu_heap()
79 struct nvkm_mmu *mmu = ummu->mmu; in nvkm_ummu_type()
108 struct nvkm_mmu *mmu = ummu->mmu; in nvkm_ummu_kind()
159 struct nvkm_mmu *mmu = device->mmu; in nvkm_ummu_new()
H A Dnv50.c30 nv50_mmu_kind(struct nvkm_mmu *base, int *count, u8 *invalid) in nv50_mmu_kind()
75 struct nvkm_mmu **pmmu) in nv50_mmu_new()
H A Dgf100.c33 gf100_mmu_kind(struct nvkm_mmu *mmu, int *count, u8 *invalid) in gf100_mmu_kind()
88 struct nvkm_mmu **pmmu) in gf100_mmu_new()
H A Dtu102.c32 tu102_mmu_kind(struct nvkm_mmu *mmu, int *count, u8 *invalid) in tu102_mmu_kind()
56 struct nvkm_mmu **pmmu) in tu102_mmu_new()
H A Dnv41.c32 nv41_mmu_init(struct nvkm_mmu *mmu) in nv41_mmu_init()
51 struct nvkm_mmu **pmmu) in nv41_mmu_new()
H A Dgm200.c30 gm200_mmu_kind(struct nvkm_mmu *mmu, int *count, u8 *invalid) in gm200_mmu_kind()
94 struct nvkm_mmu **pmmu) in gm200_mmu_new()
H A Dvmmgm200.c144 struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, in gm200_vmm_new_()
172 gm200_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, in gm200_vmm_new()
181 gm200_vmm_new_fixed(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, in gm200_vmm_new_fixed()
H A Dnv44.c32 nv44_mmu_init(struct nvkm_mmu *mmu) in nv44_mmu_init()
66 struct nvkm_mmu **pmmu) in nv44_mmu_new()
H A Dmemnv04.c31 nv04_mem_map(struct nvkm_mmu *mmu, struct nvkm_memory *memory, void *argv, in nv04_mem_map()
51 nv04_mem_new(struct nvkm_mmu *mmu, int type, u8 page, u64 size, in nv04_mem_new()
H A Dvmmgm20b.c57 gm20b_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, in gm20b_vmm_new()
66 gm20b_vmm_new_fixed(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, in gm20b_vmm_new_fixed()
H A Dmemgf100.c34 gf100_mem_map(struct nvkm_mmu *mmu, struct nvkm_memory *memory, void *argv, in gf100_mem_map()
69 gf100_mem_new(struct nvkm_mmu *mmu, int type, u8 page, u64 size, in gf100_mem_new()
H A Dmemnv50.c34 nv50_mem_map(struct nvkm_mmu *mmu, struct nvkm_memory *memory, void *argv, in nv50_mem_map()
66 nv50_mem_new(struct nvkm_mmu *mmu, int type, u8 page, u64 size, in nv50_mem_new()
H A Dmem.c33 struct nvkm_mmu *mmu;
144 nvkm_mem_new_host(struct nvkm_mmu *mmu, int type, u8 page, u64 size, in nvkm_mem_new_host()
224 nvkm_mem_new_type(struct nvkm_mmu *mmu, int type, u8 page, u64 size, in nvkm_mem_new_type()
H A Dvmmnv04.c102 nv04_vmm_new_(const struct nvkm_vmm_func *func, struct nvkm_mmu *mmu, in nv04_vmm_new_()
121 nv04_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size, in nv04_vmm_new()
H A Dumem.c90 struct nvkm_mmu *mmu = umem->mmu; in nvkm_umem_map()
145 struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu; in nvkm_umem_new()
H A Dummu.h9 struct nvkm_mmu *mmu;
H A Dnv04.c39 struct nvkm_mmu **pmmu) in nv04_mmu_new()
H A Dgh100.c22 struct nvkm_mmu **pmmu) in gh100_mmu_new()
H A Dgv100.c41 struct nvkm_mmu **pmmu) in gv100_mmu_new()
H A Dg84.c39 struct nvkm_mmu **pmmu) in g84_mmu_new()
/linux/drivers/gpu/drm/nouveau/include/nvkm/subdev/
H A Dmmu.h28 struct nvkm_mmu *mmu;
116 struct nvkm_mmu { struct
154 int nv04_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **); argument
155 int nv41_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
156 int nv44_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
157 int nv50_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
158 int g84_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
159 int mcp77_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
160 int gf100_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
161 int gk104_mmu_new(struct nvkm_device *, enum nvkm_subdev_type, int inst, struct nvkm_mmu **);
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/
H A Dvmm.c161 r535_mmu_dtor(struct nvkm_mmu *mmu) in r535_mmu_dtor()
169 struct nvkm_mmu **pmmu) in r535_mmu_new()

12