Home
last modified time | relevance | path

Searched refs:gsp (Results 1 – 25 of 101) sorted by relevance

12345

/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/
H A Dgsp.c51 struct nvkm_gsp *gsp = container_of(work, typeof(*gsp), msgq.work); in r535_gsp_msgq_work() local
53 mutex_lock(&gsp->cmdq.mutex); in r535_gsp_msgq_work()
54 if (*gsp->msgq.rptr != *gsp->msgq.wptr) in r535_gsp_msgq_work()
55 r535_gsp_msg_recv(gsp, 0, 0); in r535_gsp_msgq_work()
56 mutex_unlock(&gsp->cmdq.mutex); in r535_gsp_msgq_work()
62 struct nvkm_gsp *gsp = container_of(inth, typeof(*gsp), subdev.inth); in r535_gsp_intr() local
63 struct nvkm_subdev *subdev = &gsp->subdev; in r535_gsp_intr()
64 u32 intr = nvkm_falcon_rd32(&gsp->falcon, 0x0008); in r535_gsp_intr()
65 u32 inte = nvkm_falcon_rd32(&gsp->falcon, gsp->falcon.func->addr2 + in r535_gsp_intr()
66 gsp->falcon.func->riscv_irqmask); in r535_gsp_intr()
[all …]
H A Drpc.c142 r535_gsp_msgq_wait(struct nvkm_gsp *gsp, u32 gsp_rpc_len, int *ptime) in r535_gsp_msgq_wait() argument
144 u32 size, rptr = *gsp->msgq.rptr; in r535_gsp_msgq_wait()
149 if (WARN_ON(!size || size >= gsp->msgq.cnt)) in r535_gsp_msgq_wait()
153 u32 wptr = *gsp->msgq.wptr; in r535_gsp_msgq_wait()
155 used = wptr + gsp->msgq.cnt - rptr; in r535_gsp_msgq_wait()
156 if (used >= gsp->msgq.cnt) in r535_gsp_msgq_wait()
157 used -= gsp->msgq.cnt; in r535_gsp_msgq_wait()
171 r535_gsp_msgq_get_entry(struct nvkm_gsp *gsp) in r535_gsp_msgq_get_entry() argument
173 u32 rptr = *gsp->msgq.rptr; in r535_gsp_msgq_get_entry()
176 return (void *)((u8 *)gsp->shm.msgq.ptr + GSP_PAGE_SIZE + in r535_gsp_msgq_get_entry()
[all …]
H A DKbuild5 nvkm-y += nvkm/subdev/gsp/rm/r535/rm.o
6 nvkm-y += nvkm/subdev/gsp/rm/r535/gsp.o
7 nvkm-y += nvkm/subdev/gsp/rm/r535/rpc.o
8 nvkm-y += nvkm/subdev/gsp/rm/r535/ctrl.o
9 nvkm-y += nvkm/subdev/gsp/rm/r535/alloc.o
10 nvkm-y += nvkm/subdev/gsp/rm/r535/client.o
11 nvkm-y += nvkm/subdev/gsp/rm/r535/device.o
13 nvkm-y += nvkm/subdev/gsp/rm/r535/bar.o
14 nvkm-y += nvkm/subdev/gsp/rm/r535/fbsr.o
15 nvkm-y += nvkm/subdev/gsp/rm/r535/vmm.o
[all …]
H A Dalloc.c31 struct nvkm_gsp *gsp = client->gsp; in r535_gsp_rpc_rm_free() local
34 nvkm_debug(&gsp->subdev, "cli:0x%08x obj:0x%08x free\n", in r535_gsp_rpc_rm_free()
37 rpc = nvkm_gsp_rpc_get(gsp, NV_VGPU_MSG_FUNCTION_FREE, sizeof(*rpc)); in r535_gsp_rpc_rm_free()
44 return nvkm_gsp_rpc_wr(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV); in r535_gsp_rpc_rm_free()
52 nvkm_gsp_rpc_done(object->client->gsp, rpc); in r535_gsp_rpc_rm_alloc_done()
59 struct nvkm_gsp *gsp = object->client->gsp; in r535_gsp_rpc_rm_alloc_push() local
62 rpc = nvkm_gsp_rpc_push(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV, sizeof(*rpc)); in r535_gsp_rpc_rm_alloc_push()
69 nvkm_error(&gsp->subdev, "RM_ALLOC: 0x%x\n", rpc->status); in r535_gsp_rpc_rm_alloc_push()
72 nvkm_gsp_rpc_done(gsp, rpc); in r535_gsp_rpc_rm_alloc_push()
82 struct nvkm_gsp *gsp = client->gsp; in r535_gsp_rpc_rm_alloc_get() local
[all …]
H A Dfbsr.c56 struct nvkm_gsp *gsp = client->gsp; in r535_fbsr_memlist() local
61 rpc = nvkm_gsp_rpc_get(gsp, NV_VGPU_MSG_FUNCTION_ALLOC_MEMORY, in r535_fbsr_memlist()
102 ret = nvkm_gsp_rpc_wr(gsp, rpc, NVKM_GSP_RPC_REPLY_POLL); in r535_fbsr_memlist()
116 struct nvkm_gsp *gsp = fbsr->client.gsp; in fbsr_send() local
125 ctrl = nvkm_gsp_rm_ctrl_get(&gsp->internal.device.subdevice, in fbsr_send()
140 ret = nvkm_gsp_rm_ctrl_wr(&gsp->internal.device.subdevice, ctrl); in fbsr_send()
154 struct nvkm_gsp *gsp = fbsr->client.gsp; in fbsr_init() local
163 ctrl = nvkm_gsp_rm_ctrl_get(&gsp->internal.device.subdevice, in fbsr_init()
174 ret = nvkm_gsp_rm_ctrl_wr(&gsp->internal.device.subdevice, ctrl); in fbsr_init()
204 r535_fbsr_resume(struct nvkm_gsp *gsp) in r535_fbsr_resume() argument
[all …]
H A Dctrl.c34 nvkm_gsp_rpc_done(object->client->gsp, rpc); in r535_gsp_rpc_rm_ctrl_done()
41 struct nvkm_gsp *gsp = object->client->gsp; in r535_gsp_rpc_rm_ctrl_push() local
44 rpc = nvkm_gsp_rpc_push(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV, repc); in r535_gsp_rpc_rm_ctrl_push()
53 nvkm_error(&gsp->subdev, "cli:0x%08x obj:0x%08x ctrl cmd:0x%08x failed: 0x%08x\n", in r535_gsp_rpc_rm_ctrl_push()
60 nvkm_gsp_rpc_done(gsp, rpc); in r535_gsp_rpc_rm_ctrl_push()
69 struct nvkm_gsp *gsp = client->gsp; in r535_gsp_rpc_rm_ctrl_get() local
72 nvkm_debug(&gsp->subdev, "cli:0x%08x obj:0x%08x ctrl cmd:0x%08x params_size:%d\n", in r535_gsp_rpc_rm_ctrl_get()
75 rpc = nvkm_gsp_rpc_get(gsp, NV_VGPU_MSG_FUNCTION_GSP_RM_CONTROL, in r535_gsp_rpc_rm_ctrl_get()
H A Dbar.c53 r535_bar_bar2_update_pde(struct nvkm_gsp *gsp, u8 page_shift, u64 pdbe) in r535_bar_bar2_update_pde() argument
57 rpc = nvkm_gsp_rpc_get(gsp, NV_VGPU_MSG_FUNCTION_UPDATE_BAR_PDE, sizeof(*rpc)); in r535_bar_bar2_update_pde()
65 return nvkm_gsp_rpc_wr(gsp, rpc, NVKM_GSP_RPC_REPLY_RECV); in r535_bar_bar2_update_pde()
72 struct nvkm_gsp *gsp = bar->subdev.device->gsp; in r535_bar_bar2_fini() local
77 WARN_ON(r535_bar_bar2_update_pde(gsp, vmm->func->page[0].shift, 0)); in r535_bar_bar2_fini()
85 struct nvkm_gsp *gsp = device->gsp; in r535_bar_bar2_init() local
97 WARN_ON(r535_bar_bar2_update_pde(gsp, vmm->func->page[0].shift, pdbe)); in r535_bar_bar2_init()
98 vmm->rm.bar2_pdb = gsp->bar.rm_bar2_pdb; in r535_bar_bar2_init()
131 struct nvkm_gsp *gsp = device->gsp; in r535_bar_bar1_init() local
136 ret = nvkm_ram_wrap(device, gsp->bar.rm_bar1_pdb, 0x1000, &pd3); in r535_bar_bar1_init()
H A Ddevice.c32 struct nvkm_gsp *gsp = client->gsp; in r535_gsp_event_dtor() local
34 mutex_lock(&gsp->client_id.mutex); in r535_gsp_event_dtor()
39 mutex_unlock(&gsp->client_id.mutex); in r535_gsp_event_dtor()
66 struct nvkm_gsp *gsp = client->gsp; in r535_gsp_device_event_ctor() local
95 mutex_lock(&gsp->client_id.mutex); in r535_gsp_device_event_ctor()
98 mutex_unlock(&gsp->client_id.mutex); in r535_gsp_device_event_ctor()
H A Dfifo.c44 struct nvkm_gsp *gsp = chan->rm.object.client->gsp; in r535_chan_doorbell_handle() local
46 return gsp->rm->gpu->fifo.chan.doorbell_handle(chan); in r535_chan_doorbell_handle()
80 struct nvkm_gsp *gsp = device->object.client->gsp; in r535_chan_alloc() local
81 struct nvkm_fifo *fifo = gsp->subdev.device->fifo; in r535_chan_alloc()
159 const struct nvkm_rm_api *rmapi = device->gsp->rm->api; in r535_chan_ramfc_write()
245 ret = nvkm_gsp_intr_nonstall(subdev->device->gsp, subdev->type, subdev->inst); in r535_engn_nonstall()
364 struct nvkm_gsp *gsp = priv; in r535_fifo_rc_triggered() local
369 nvkm_error(&gsp->subdev, "rc: engn:%08x chid:%d type:%d scope:%d part:%d\n", in r535_fifo_rc_triggered()
373 r535_fifo_rc_chid(gsp->subdev.device->fifo, msg->chid); in r535_fifo_rc_triggered()
430 struct nvkm_gsp *gsp = fifo->engine.subdev.device->gsp; in r535_fifo_ectx_size() local
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/
H A Dbase.c25 nvkm_gsp_intr_nonstall(struct nvkm_gsp *gsp, enum nvkm_subdev_type type, int inst) in nvkm_gsp_intr_nonstall() argument
27 for (int i = 0; i < gsp->intr_nr; i++) { in nvkm_gsp_intr_nonstall()
28 if (gsp->intr[i].type == type && gsp->intr[i].inst == inst) in nvkm_gsp_intr_nonstall()
29 return gsp->intr[i].nonstall; in nvkm_gsp_intr_nonstall()
36 nvkm_gsp_intr_stall(struct nvkm_gsp *gsp, enum nvkm_subdev_type type, int inst) in nvkm_gsp_intr_stall() argument
38 for (int i = 0; i < gsp->intr_nr; i++) { in nvkm_gsp_intr_stall()
39 if (gsp->intr[i].type == type && gsp->intr[i].inst == inst) { in nvkm_gsp_intr_stall()
40 if (gsp->intr[i].stall != ~0) in nvkm_gsp_intr_stall()
41 return gsp->intr[i].stall; in nvkm_gsp_intr_stall()
53 struct nvkm_gsp *gsp = nvkm_gsp(subdev); in nvkm_gsp_fini() local
[all …]
H A Dtu102.c34 tu102_gsp_fwsec_sb_ctor(struct nvkm_gsp *gsp) in tu102_gsp_fwsec_sb_ctor() argument
36 return nvkm_gsp_fwsec_sb_init(gsp); in tu102_gsp_fwsec_sb_ctor()
40 tu102_gsp_fwsec_sb_dtor(struct nvkm_gsp *gsp) in tu102_gsp_fwsec_sb_dtor() argument
42 nvkm_falcon_fw_dtor(&gsp->fws.falcon.sb); in tu102_gsp_fwsec_sb_dtor()
46 tu102_gsp_booter_unload(struct nvkm_gsp *gsp, u32 mbox0, u32 mbox1) in tu102_gsp_booter_unload() argument
48 struct nvkm_subdev *subdev = &gsp->subdev; in tu102_gsp_booter_unload()
59 ret = nvkm_falcon_fw_boot(&gsp->booter.unload, &gsp->subdev, true, &mbox0, &mbox1, 0, 0); in tu102_gsp_booter_unload()
71 tu102_gsp_booter_load(struct nvkm_gsp *gsp, u32 mbox0, u32 mbox1) in tu102_gsp_booter_load() argument
73 return nvkm_falcon_fw_boot(&gsp->booter.load, &gsp->subdev, true, &mbox0, &mbox1, 0, 0); in tu102_gsp_booter_load()
77 tu102_gsp_booter_ctor(struct nvkm_gsp *gsp, const char *name, const struct firmware *blob, in tu102_gsp_booter_ctor() argument
[all …]
H A Dgh100.c20 gh100_gsp_fini(struct nvkm_gsp *gsp, enum nvkm_suspend_state suspend) in gh100_gsp_fini() argument
22 struct nvkm_falcon *falcon = &gsp->falcon; in gh100_gsp_fini()
26 ret = r535_gsp_fini(gsp, suspend); in gh100_gsp_fini()
44 gh100_gsp_lockdown_released(struct nvkm_gsp *gsp, u32 *mbox0) in gh100_gsp_lockdown_released() argument
49 *mbox0 = nvkm_falcon_rd32(&gsp->falcon, NV_PFALCON_FALCON_MAILBOX0); in gh100_gsp_lockdown_released()
56 u32 mbox1 = nvkm_falcon_rd32(&gsp->falcon, NV_PFALCON_FALCON_MAILBOX1); in gh100_gsp_lockdown_released()
59 if ((((u64)mbox1 << 32) | *mbox0) != gsp->fmc.args.addr) in gh100_gsp_lockdown_released()
64 data = nvkm_falcon_rd32(&gsp->falcon, NV_PFALCON_FALCON_HWCFG2); in gh100_gsp_lockdown_released()
69 gh100_gsp_init(struct nvkm_gsp *gsp) in gh100_gsp_init() argument
71 struct nvkm_subdev *subdev = &gsp->subdev; in gh100_gsp_init()
[all …]
H A DKbuild2 nvkm-y += nvkm/subdev/gsp/base.o
3 nvkm-y += nvkm/subdev/gsp/fwsec.o
5 nvkm-y += nvkm/subdev/gsp/gv100.o
6 nvkm-y += nvkm/subdev/gsp/tu102.o
7 nvkm-y += nvkm/subdev/gsp/tu116.o
8 nvkm-y += nvkm/subdev/gsp/ga100.o
9 nvkm-y += nvkm/subdev/gsp/ga102.o
10 nvkm-y += nvkm/subdev/gsp/gh100.o
11 nvkm-y += nvkm/subdev/gsp/ad102.o
12 nvkm-y += nvkm/subdev/gsp/gb100.o
[all …]
H A Dfwsec.c89 nvkm_gsp_fwsec_patch(struct nvkm_gsp *gsp, struct nvkm_falcon_fw *fw, u32 if_offset, u32 init_cmd) in nvkm_gsp_fwsec_patch() argument
120 frtscmd->frts_region.addr = gsp->fb.wpr2.frts.addr >> 12; in nvkm_gsp_fwsec_patch()
121 frtscmd->frts_region.size = gsp->fb.wpr2.frts.size >> 12; in nvkm_gsp_fwsec_patch()
172 nvkm_gsp_fwsec_v2(struct nvkm_gsp *gsp, const char *name, in nvkm_gsp_fwsec_v2() argument
176 struct nvkm_subdev *subdev = &gsp->subdev; in nvkm_gsp_fwsec_v2()
183 ret = nvkm_falcon_fw_ctor(gsp->func->fwsec, name, subdev->device, true, in nvkm_gsp_fwsec_v2()
185 &gsp->falcon, fw); in nvkm_gsp_fwsec_v2()
219 return nvkm_gsp_fwsec_patch(gsp, fw, desc->InterfaceOffset, init_cmd); in nvkm_gsp_fwsec_v2()
223 nvkm_gsp_fwsec_v3(struct nvkm_gsp *gsp, const char *name, in nvkm_gsp_fwsec_v3() argument
227 struct nvkm_device *device = gsp->subdev.device; in nvkm_gsp_fwsec_v3()
[all …]
H A Dpriv.h11 int nvkm_gsp_fwsec_sb_init(struct nvkm_gsp *gsp);
100 static inline int nvkm_gsp_fwsec_sb_ctor(struct nvkm_gsp *gsp) in nvkm_gsp_fwsec_sb_ctor() argument
102 if (gsp->func->fwsec_sb.ctor) in nvkm_gsp_fwsec_sb_ctor()
103 return gsp->func->fwsec_sb.ctor(gsp); in nvkm_gsp_fwsec_sb_ctor()
107 static inline void nvkm_gsp_fwsec_sb_dtor(struct nvkm_gsp *gsp) in nvkm_gsp_fwsec_sb_dtor() argument
109 if (gsp->func->fwsec_sb.dtor) in nvkm_gsp_fwsec_sb_dtor()
110 gsp->func->fwsec_sb.dtor(gsp); in nvkm_gsp_fwsec_sb_dtor()
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/
H A Dgsp.c18 r570_gsp_sr_data_size(struct nvkm_gsp *gsp) in r570_gsp_sr_data_size() argument
20 GspFwWprMeta *meta = gsp->wpr_meta.data; in r570_gsp_sr_data_size()
27 r570_gsp_drop_post_nocat_record(struct nvkm_gsp *gsp) in r570_gsp_drop_post_nocat_record() argument
29 if (gsp->subdev.debug < NV_DBG_DEBUG) { in r570_gsp_drop_post_nocat_record()
30 r535_gsp_msg_ntfy_add(gsp, NV_VGPU_MSG_EVENT_GSP_POST_NOCAT_RECORD, NULL, NULL); in r570_gsp_drop_post_nocat_record()
31 r535_gsp_msg_ntfy_add(gsp, NV_VGPU_MSG_EVENT_GSP_LOCKDOWN_NOTICE, NULL, NULL); in r570_gsp_drop_post_nocat_record()
77 r570_gsp_get_static_info(struct nvkm_gsp *gsp) in r570_gsp_get_static_info() argument
84 rpc = nvkm_gsp_rpc_rd(gsp, NV_VGPU_MSG_FUNCTION_GET_GSP_STATIC_INFO, sizeof(*rpc)); in r570_gsp_get_static_info()
88 gsp->internal.client.object.client = &gsp->internal.client; in r570_gsp_get_static_info()
89 gsp->internal.client.object.parent = NULL; in r570_gsp_get_static_info()
[all …]
H A Dfbsr.c14 r570_fbsr_suspend_channels(struct nvkm_gsp *gsp, bool suspend) in r570_fbsr_suspend_channels() argument
18 ctrl = nvkm_gsp_rm_ctrl_get(&gsp->internal.device.subdevice, in r570_fbsr_suspend_channels()
26 return nvkm_gsp_rm_ctrl_wr(&gsp->internal.device.subdevice, ctrl); in r570_fbsr_suspend_channels()
30 r570_fbsr_resume(struct nvkm_gsp *gsp) in r570_fbsr_resume() argument
32 struct nvkm_device *device = gsp->subdev.device; in r570_fbsr_resume()
58 r570_fbsr_suspend_channels(device->gsp, false); in r570_fbsr_resume()
61 r535_fbsr_resume(gsp); in r570_fbsr_resume()
65 r570_fbsr_init(struct nvkm_gsp *gsp, struct sg_table *sgt, u64 size, bool runtime) in r570_fbsr_init() argument
71 ret = r535_fbsr_memlist(&gsp->internal.device, 0xcaf00003, NVKM_MEM_TARGET_HOST, in r570_fbsr_init()
76 ctrl = nvkm_gsp_rm_ctrl_get(&gsp->internal.device.subdevice, in r570_fbsr_init()
[all …]
H A DKbuild2 nvkm-y += nvkm/subdev/gsp/rm/r570/rm.o
3 nvkm-y += nvkm/subdev/gsp/rm/r570/gsp.o
4 nvkm-y += nvkm/subdev/gsp/rm/r570/client.o
5 nvkm-y += nvkm/subdev/gsp/rm/r570/fbsr.o
6 nvkm-y += nvkm/subdev/gsp/rm/r570/disp.o
7 nvkm-y += nvkm/subdev/gsp/rm/r570/fifo.o
8 nvkm-y += nvkm/subdev/gsp/rm/r570/gr.o
9 nvkm-y += nvkm/subdev/gsp/rm/r570/ofa.o
H A Dgr.c16 r570_gr_tpc_mask(struct nvkm_gsp *gsp, int gpc, u32 *pmask) in r570_gr_tpc_mask() argument
21 ctrl = nvkm_gsp_rm_ctrl_get(&gsp->internal.device.subdevice, in r570_gr_tpc_mask()
28 ret = nvkm_gsp_rm_ctrl_push(&gsp->internal.device.subdevice, &ctrl, sizeof(*ctrl)); in r570_gr_tpc_mask()
34 nvkm_gsp_rm_ctrl_done(&gsp->internal.device.subdevice, ctrl); in r570_gr_tpc_mask()
39 r570_gr_gpc_mask(struct nvkm_gsp *gsp, u32 *pmask) in r570_gr_gpc_mask() argument
43 ctrl = nvkm_gsp_rm_ctrl_rd(&gsp->internal.device.subdevice, in r570_gr_gpc_mask()
50 nvkm_gsp_rm_ctrl_done(&gsp->internal.device.subdevice, ctrl); in r570_gr_gpc_mask()
97 struct nvkm_gsp *gsp = device->gsp; in r570_gr_scrubber_init() local
98 struct nvkm_rm *rm = gsp->rm; in r570_gr_scrubber_init()
171 struct nvkm_gsp *gsp = subdev->device->gsp; in r570_gr_get_ctxbufs_info() local
[all …]
H A Dfifo.c25 struct nvkm_gsp *gsp = device->object.client->gsp; in r570_chan_alloc() local
26 struct nvkm_fifo *fifo = gsp->subdev.device->fifo; in r570_chan_alloc()
102 struct nvkm_gsp *gsp = priv; in r570_fifo_rc_triggered() local
107 nvkm_error(&gsp->subdev, "rc engn:%08x chid:%d gfid:%d level:%d type:%d scope:%d part:%d " in r570_fifo_rc_triggered()
113 r535_fifo_rc_chid(gsp->subdev.device->fifo, msg->chid); in r570_fifo_rc_triggered()
121 struct nvkm_gsp *gsp = fifo->engine.subdev.device->gsp; in r570_fifo_ectx_size() local
125 ctrl = nvkm_gsp_rm_ctrl_rd(&gsp->internal.device.subdevice, in r570_fifo_ectx_size()
143 nvkm_gsp_rm_ctrl_done(&gsp->internal.device.subdevice, ctrl); in r570_fifo_ectx_size()
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/
H A Dclient.c11 struct nvkm_gsp *gsp = client->gsp; in nvkm_gsp_client_dtor() local
13 if (!gsp) in nvkm_gsp_client_dtor()
19 mutex_lock(&gsp->client_id.mutex); in nvkm_gsp_client_dtor()
20 idr_remove(&gsp->client_id.idr, id); in nvkm_gsp_client_dtor()
21 mutex_unlock(&gsp->client_id.mutex); in nvkm_gsp_client_dtor()
23 client->gsp = NULL; in nvkm_gsp_client_dtor()
27 nvkm_gsp_client_ctor(struct nvkm_gsp *gsp, struct nvkm_gsp_client *client) in nvkm_gsp_client_ctor() argument
31 if (WARN_ON(!gsp->rm)) in nvkm_gsp_client_ctor()
34 mutex_lock(&gsp->client_id.mutex); in nvkm_gsp_client_ctor()
35 id = idr_alloc(&gsp->client_id.idr, client, 0, NVKM_RM_CLIENT_MASK + 1, GFP_KERNEL); in nvkm_gsp_client_ctor()
[all …]
H A DKbuild4 nvkm-y += nvkm/subdev/gsp/rm/client.o
5 nvkm-y += nvkm/subdev/gsp/rm/engine.o
6 nvkm-y += nvkm/subdev/gsp/rm/gr.o
7 nvkm-y += nvkm/subdev/gsp/rm/nvdec.o
8 nvkm-y += nvkm/subdev/gsp/rm/nvenc.o
10 nvkm-y += nvkm/subdev/gsp/rm/tu1xx.o
11 nvkm-y += nvkm/subdev/gsp/rm/ga100.o
12 nvkm-y += nvkm/subdev/gsp/rm/ga1xx.o
13 nvkm-y += nvkm/subdev/gsp/rm/ad10x.o
14 nvkm-y += nvkm/subdev/gsp/rm/gh100.o
[all …]
/linux/drivers/gpu/drm/nouveau/include/nvkm/subdev/
H A Dgsp.h207 struct nvkm_gsp *gsp; member
271 nvkm_gsp_rm(struct nvkm_gsp *gsp) in nvkm_gsp_rm() argument
273 return gsp && (gsp->fws.rm || gsp->fw.img); in nvkm_gsp_rm()
279 nvkm_gsp_rpc_get(struct nvkm_gsp *gsp, u32 fn, u32 argc) in nvkm_gsp_rpc_get() argument
281 return gsp->rm->api->rpc->get(gsp, fn, argc); in nvkm_gsp_rpc_get()
285 nvkm_gsp_rpc_push(struct nvkm_gsp *gsp, void *argv, in nvkm_gsp_rpc_push() argument
288 return gsp->rm->api->rpc->push(gsp, argv, policy, repc); in nvkm_gsp_rpc_push()
292 nvkm_gsp_rpc_rd(struct nvkm_gsp *gsp, u32 fn, u32 argc) in nvkm_gsp_rpc_rd() argument
294 void *argv = nvkm_gsp_rpc_get(gsp, fn, argc); in nvkm_gsp_rpc_rd()
299 return nvkm_gsp_rpc_push(gsp, argv, NVKM_GSP_RPC_REPLY_RECV, argc); in nvkm_gsp_rpc_rd()
[all …]
/linux/arch/powerpc/kvm/
H A Dtest-guest-state-buffer.c104 struct kvmppc_gs_parser gsp = { 0 }; in test_gs_parsing() local
115 KUNIT_EXPECT_GE(test, kvmppc_gse_parse(&gsp, gsb), 0); in test_gs_parsing()
117 gse = kvmppc_gsp_lookup(&gsp, KVMPPC_GSID_GPR(0)); in test_gs_parsing()
264 struct kvmppc_gs_parser gsp = { 0 }; in test1_refresh_info() local
269 rc = kvmppc_gse_parse(&gsp, gsb); in test1_refresh_info()
273 gse = kvmppc_gsp_lookup(&gsp, KVMPPC_GSID_GPR(0)); in test1_refresh_info()
277 gse = kvmppc_gsp_lookup(&gsp, KVMPPC_GSID_CR); in test1_refresh_info()
376 struct kvmppc_gs_parser gsp = { 0 }; in test_hostwide_refresh_info() local
381 rc = kvmppc_gse_parse(&gsp, gsb); in test_hostwide_refresh_info()
385 gse = kvmppc_gsp_lookup(&gsp, KVMPPC_GSID_L0_GUEST_HEAP); in test_hostwide_refresh_info()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dr535.c34 struct nvkm_gsp *gsp = fb->subdev.device->gsp; in r535_fb_ram_new() local
44 ram->size = gsp->fb.size; in r535_fb_ram_new()
48 for (int i = 0; i < gsp->fb.region_nr; i++) { in r535_fb_ram_new()
50 gsp->fb.region[i].addr >> NVKM_RAM_MM_SHIFT, in r535_fb_ram_new()
51 gsp->fb.region[i].size >> NVKM_RAM_MM_SHIFT, in r535_fb_ram_new()

12345