/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_ring_submission.c | 36 static void set_hwstam(struct intel_engine_cs *engine, u32 mask) in set_hwstam() argument 42 if (engine->class == RENDER_CLASS) { in set_hwstam() 43 if (GRAPHICS_VER(engine->i915) >= 6) in set_hwstam() 49 intel_engine_set_hwsp_writemask(engine, mask); in set_hwstam() 52 static void set_hws_pga(struct intel_engine_cs *engine, phys_addr_t phys) in set_hws_pga() argument 57 if (GRAPHICS_VER(engine->i915) >= 4) in set_hws_pga() 60 intel_uncore_write(engine->uncore, HWS_PGA, addr); in set_hws_pga() 63 static struct page *status_page(struct intel_engine_cs *engine) in status_page() argument 65 struct drm_i915_gem_object *obj = engine->status_page.vma->obj; in status_page() 71 static void ring_setup_phys_status_page(struct intel_engine_cs *engine) in ring_setup_phys_status_page() argument 77 set_hwsp(struct intel_engine_cs * engine,u32 offset) set_hwsp() argument 117 flush_cs_tlb(struct intel_engine_cs * engine) flush_cs_tlb() argument 138 ring_setup_status_page(struct intel_engine_cs * engine) ring_setup_status_page() argument 159 set_pp_dir(struct intel_engine_cs * engine) set_pp_dir() argument 176 stop_ring(struct intel_engine_cs * engine) stop_ring() argument 193 xcs_resume(struct intel_engine_cs * engine) xcs_resume() argument 304 sanitize_hwsp(struct intel_engine_cs * engine) sanitize_hwsp() argument 312 xcs_sanitize(struct intel_engine_cs * engine) xcs_sanitize() argument 339 reset_prepare(struct intel_engine_cs * engine) reset_prepare() argument 387 reset_rewind(struct intel_engine_cs * engine,bool stalled) reset_rewind() argument 454 reset_finish(struct intel_engine_cs * engine) reset_finish() argument 458 reset_cancel(struct intel_engine_cs * engine) reset_cancel() argument 561 alloc_context_vma(struct intel_engine_cs * engine) alloc_context_vma() argument 605 struct intel_engine_cs *engine = ce->engine; ring_context_alloc() local 645 struct intel_engine_cs *engine; ring_context_revoke() local 663 struct intel_engine_cs *engine = NULL; ring_context_cancel_request() local 696 const struct intel_engine_cs * const engine = rq->engine; load_pd_dir() local 730 struct intel_engine_cs *engine = rq->engine; mi_set_context() local 922 struct intel_engine_cs *engine = rq->engine; clear_residuals() local 953 struct intel_engine_cs *engine = rq->engine; switch_context() local 1081 i9xx_set_default_submission(struct intel_engine_cs * engine) i9xx_set_default_submission() argument 1086 gen6_bsd_set_default_submission(struct intel_engine_cs * engine) gen6_bsd_set_default_submission() argument 1091 ring_release(struct intel_engine_cs * engine) ring_release() argument 1112 irq_handler(struct intel_engine_cs * engine,u16 iir) irq_handler() argument 1117 setup_irq(struct intel_engine_cs * engine) setup_irq() argument 1154 setup_common(struct intel_engine_cs * engine) setup_common() argument 1198 setup_rcs(struct intel_engine_cs * engine) setup_rcs() argument 1227 setup_vcs(struct intel_engine_cs * engine) setup_vcs() argument 1251 setup_bcs(struct intel_engine_cs * engine) setup_bcs() argument 1264 setup_vecs(struct intel_engine_cs * engine) setup_vecs() argument 1278 gen7_ctx_switch_bb_setup(struct intel_engine_cs * const engine,struct i915_vma * const vma) gen7_ctx_switch_bb_setup() argument 1284 gen7_ctx_switch_bb_init(struct intel_engine_cs * engine,struct i915_gem_ww_ctx * ww,struct i915_vma * vma) gen7_ctx_switch_bb_init() argument 1310 gen7_ctx_vma(struct intel_engine_cs * engine) gen7_ctx_vma() argument 1348 intel_ring_submission_setup(struct intel_engine_cs * engine) intel_ring_submission_setup() argument [all...] |
H A D | intel_engine_heartbeat.c | 17 * While the engine is active, we send a periodic pulse along the engine 19 * is stuck, and we fail to preempt it, we declare the engine hung and 23 static bool next_heartbeat(struct intel_engine_cs *engine) in next_heartbeat() argument 28 delay = READ_ONCE(engine->props.heartbeat_interval_ms); in next_heartbeat() 30 rq = engine->heartbeat.systole; in next_heartbeat() 42 delay == engine->defaults.heartbeat_interval_ms) { in next_heartbeat() 50 longer = READ_ONCE(engine->props.preempt_timeout_ms) * 2; in next_heartbeat() 51 longer = intel_clamp_heartbeat_interval_ms(engine, longer); in next_heartbeat() 62 mod_delayed_work(system_highpri_wq, &engine in next_heartbeat() 79 idle_pulse(struct intel_engine_cs * engine,struct i915_request * rq) idle_pulse() argument 97 show_heartbeat(const struct i915_request * rq,struct intel_engine_cs * engine) show_heartbeat() argument 117 reset_engine(struct intel_engine_cs * engine,struct i915_request * rq) reset_engine() argument 139 struct intel_engine_cs *engine = heartbeat() local 237 intel_engine_unpark_heartbeat(struct intel_engine_cs * engine) intel_engine_unpark_heartbeat() argument 245 intel_engine_park_heartbeat(struct intel_engine_cs * engine) intel_engine_park_heartbeat() argument 253 struct intel_engine_cs *engine; intel_gt_unpark_heartbeats() local 263 struct intel_engine_cs *engine; intel_gt_park_heartbeats() local 270 intel_engine_init_heartbeat(struct intel_engine_cs * engine) intel_engine_init_heartbeat() argument 275 __intel_engine_pulse(struct intel_engine_cs * engine) __intel_engine_pulse() argument 300 set_heartbeat(struct intel_engine_cs * engine,unsigned long delay) set_heartbeat() argument 314 intel_engine_set_heartbeat(struct intel_engine_cs * engine,unsigned long delay) intel_engine_set_heartbeat() argument 358 intel_engine_pulse(struct intel_engine_cs * engine) intel_engine_pulse() argument 380 intel_engine_flush_barriers(struct intel_engine_cs * engine) intel_engine_flush_barriers() argument [all...] |
H A D | mock_engine.c | 60 static struct intel_ring *mock_ring(struct intel_engine_cs *engine) in mock_ring() argument 75 ring->vma = create_ring_vma(engine->gt->ggtt, PAGE_SIZE); in mock_ring() 93 static struct i915_request *first_request(struct mock_engine *engine) in first_request() argument 95 return list_first_entry_or_null(&engine->hw_queue, in first_request() 106 intel_engine_signal_breadcrumbs(request->engine); in advance() 111 struct mock_engine *engine = timer_container_of(engine, t, hw_delay); in hw_delay_complete() local 115 spin_lock_irqsave(&engine->hw_lock, flags); in hw_delay_complete() 118 request = first_request(engine); in hw_delay_complete() 126 while ((request = first_request(engine))) { in hw_delay_complete() 238 struct mock_engine *engine = mock_submit_request() local 264 struct intel_engine_cs *engine, *locked; mock_remove_from_engine() local 284 mock_reset_prepare(struct intel_engine_cs * engine) mock_reset_prepare() argument 288 mock_reset_rewind(struct intel_engine_cs * engine,bool stalled) mock_reset_rewind() argument 293 mock_reset_cancel(struct intel_engine_cs * engine) mock_reset_cancel() argument 321 mock_reset_finish(struct intel_engine_cs * engine) mock_reset_finish() argument 325 mock_engine_release(struct intel_engine_cs * engine) mock_engine_release() argument 345 struct mock_engine *engine; mock_engine() local 393 mock_engine_init(struct intel_engine_cs * engine) mock_engine_init() argument 429 mock_engine_flush(struct intel_engine_cs * engine) mock_engine_flush() argument 443 mock_engine_reset(struct intel_engine_cs * engine) mock_engine_reset() argument [all...] |
H A D | selftest_engine_heartbeat.c | 14 static void reset_heartbeat(struct intel_engine_cs *engine) in reset_heartbeat() argument 16 intel_engine_set_heartbeat(engine, in reset_heartbeat() 17 engine->defaults.heartbeat_interval_ms); in reset_heartbeat() 37 static int engine_sync_barrier(struct intel_engine_cs *engine) in engine_sync_barrier() argument 39 return timeline_sync(engine->kernel_context->timeline); in engine_sync_barrier() 90 static int __live_idle_pulse(struct intel_engine_cs *engine, in __live_idle_pulse() argument 96 GEM_BUG_ON(!intel_engine_pm_is_awake(engine)); in __live_idle_pulse() 106 err = i915_active_acquire_preallocate_barrier(&p->active, engine); in __live_idle_pulse() 116 GEM_BUG_ON(llist_empty(&engine->barrier_tasks)); in __live_idle_pulse() 118 err = fn(engine); in __live_idle_pulse() 157 struct intel_engine_cs *engine; live_idle_flush() local 177 struct intel_engine_cs *engine; live_idle_pulse() local 196 __live_heartbeat_off(struct intel_engine_cs * engine) __live_heartbeat_off() argument 241 struct intel_engine_cs *engine; live_heartbeat_off() local 283 st_engine_heartbeat_disable(struct intel_engine_cs * engine) st_engine_heartbeat_disable() argument 291 st_engine_heartbeat_enable(struct intel_engine_cs * engine) st_engine_heartbeat_enable() argument 299 st_engine_heartbeat_disable_no_pm(struct intel_engine_cs * engine) st_engine_heartbeat_disable_no_pm() argument 315 st_engine_heartbeat_enable_no_pm(struct intel_engine_cs * engine) st_engine_heartbeat_enable_no_pm() argument [all...] |
H A D | intel_engine_pm.c | 20 static void intel_gsc_idle_msg_enable(struct intel_engine_cs *engine) in intel_gsc_idle_msg_enable() argument 22 struct drm_i915_private *i915 = engine->i915; in intel_gsc_idle_msg_enable() 24 if (MEDIA_VER(i915) >= 13 && engine->id == GSC0) { in intel_gsc_idle_msg_enable() 25 intel_uncore_write(engine->gt->uncore, in intel_gsc_idle_msg_enable() 29 intel_uncore_write(engine->gt->uncore, in intel_gsc_idle_msg_enable() 42 int type = intel_gt_coherent_map_type(ce->engine->gt, obj, true); in dbg_poison_ce() 60 struct intel_engine_cs *engine = in __engine_unpark() local 61 container_of(wf, typeof(*engine), wakeref); in __engine_unpark() 64 ENGINE_TRACE(engine, "\n"); in __engine_unpark() 66 engine in __engine_unpark() 111 __queue_and_release_pm(struct i915_request * rq,struct intel_timeline * tl,struct intel_engine_cs * engine) __queue_and_release_pm() argument 151 switch_to_kernel_context(struct intel_engine_cs * engine) switch_to_kernel_context() argument 245 call_idle_barriers(struct intel_engine_cs * engine) call_idle_barriers() argument 260 struct intel_engine_cs *engine = __engine_park() local 295 intel_engine_init__pm(struct intel_engine_cs * engine) intel_engine_init__pm() argument 311 intel_engine_reset_pinned_contexts(struct intel_engine_cs * engine) intel_engine_reset_pinned_contexts() argument [all...] |
H A D | intel_engine_pm.h | 17 intel_engine_pm_is_awake(const struct intel_engine_cs *engine) in intel_engine_pm_is_awake() argument 19 return intel_wakeref_is_active(&engine->wakeref); in intel_engine_pm_is_awake() 22 static inline void __intel_engine_pm_get(struct intel_engine_cs *engine) in __intel_engine_pm_get() argument 24 __intel_wakeref_get(&engine->wakeref); in __intel_engine_pm_get() 27 static inline void intel_engine_pm_get(struct intel_engine_cs *engine) in intel_engine_pm_get() argument 29 intel_wakeref_get(&engine->wakeref); in intel_engine_pm_get() 32 static inline bool intel_engine_pm_get_if_awake(struct intel_engine_cs *engine) in intel_engine_pm_get_if_awake() argument 34 return intel_wakeref_get_if_active(&engine->wakeref); in intel_engine_pm_get_if_awake() 37 static inline void intel_engine_pm_might_get(struct intel_engine_cs *engine) in intel_engine_pm_might_get() argument 39 if (!intel_engine_is_virtual(engine)) { in intel_engine_pm_might_get() 52 intel_engine_pm_put(struct intel_engine_cs * engine) intel_engine_pm_put() argument 57 intel_engine_pm_put_async(struct intel_engine_cs * engine) intel_engine_pm_put_async() argument 62 intel_engine_pm_put_delay(struct intel_engine_cs * engine,unsigned long delay) intel_engine_pm_put_delay() argument 68 intel_engine_pm_flush(struct intel_engine_cs * engine) intel_engine_pm_flush() argument 73 intel_engine_pm_might_put(struct intel_engine_cs * engine) intel_engine_pm_might_put() argument 89 intel_engine_create_kernel_request(struct intel_engine_cs * engine) intel_engine_create_kernel_request() argument [all...] |
H A D | intel_execlists_submission.c | 24 * shouldn't we just need a set of those per engine command streamer? This is 26 * rings, the engine cs shifts to a new "ring buffer" with every context 41 * Now that ringbuffers belong per-context (and not per-engine, like before) 42 * and that contexts are uniquely tied to a given engine (and not reusable, 45 * - One ringbuffer per-engine inside each context. 46 * - One backing object per-engine inside each context. 50 * more complex, because we don't know at creation time which engine is going 55 * gets populated for a given engine once we receive an execbuffer. If later 57 * engine, we allocate/populate a new ringbuffer and context backing object and 74 * for the appropriate engine 201 to_virtual_engine(struct intel_engine_cs * engine) to_virtual_engine() argument 238 ring_set_paused(const struct intel_engine_cs * engine,int state) ring_set_paused() argument 297 need_preempt(const struct intel_engine_cs * engine,const struct i915_request * rq) need_preempt() argument 367 __unwind_incomplete_requests(struct intel_engine_cs * engine) __unwind_incomplete_requests() argument 423 reset_active(struct i915_request * rq,struct intel_engine_cs * engine) reset_active() argument 468 struct intel_engine_cs * const engine = rq->engine; __execlists_schedule_in() local 540 struct intel_engine_cs *engine = rq->engine; resubmit_virtual_request() local 554 struct intel_engine_cs *engine = rq->engine; kick_siblings() local 583 struct intel_engine_cs * const engine = rq->engine; __execlists_schedule_out() local 751 const struct intel_engine_cs *engine = trace_ports() local 764 reset_in_progress(const struct intel_engine_cs * engine) reset_in_progress() argument 773 struct intel_engine_cs *engine = assert_pending_valid() local 899 execlists_submit_ports(struct intel_engine_cs * engine) execlists_submit_ports() argument 989 virtual_matches(const struct virtual_engine * ve,const struct i915_request * rq,const struct intel_engine_cs * engine) virtual_matches() argument 1016 first_virtual_engine(struct intel_engine_cs * engine) first_virtual_engine() argument 1041 virtual_xfer_context(struct virtual_engine * ve,struct intel_engine_cs * engine) virtual_xfer_context() argument 1114 defer_active(struct intel_engine_cs * engine) defer_active() argument 1145 needs_timeslice(const struct intel_engine_cs * engine,const struct i915_request * rq) needs_timeslice() argument 1181 timeslice_expired(struct intel_engine_cs * engine,const struct i915_request * rq) timeslice_expired() argument 1194 timeslice(const struct intel_engine_cs * engine) timeslice() argument 1199 start_timeslice(struct intel_engine_cs * engine) start_timeslice() argument 1230 active_preempt_timeout(struct intel_engine_cs * engine,const struct i915_request * rq) active_preempt_timeout() argument 1246 set_preempt_timeout(struct intel_engine_cs * engine,const struct i915_request * rq) set_preempt_timeout() argument 1264 execlists_dequeue(struct intel_engine_cs * engine) execlists_dequeue() argument 1611 execlists_dequeue_irq(struct intel_engine_cs * engine) execlists_dequeue_irq() argument 1743 wa_csb_read(const struct intel_engine_cs * engine,u64 * const csb) wa_csb_read() argument 1776 csb_read(const struct intel_engine_cs * engine,u64 * const csb) csb_read() argument 1808 process_csb(struct intel_engine_cs * engine,struct i915_request ** inactive) process_csb() argument 2090 execlists_hold(struct intel_engine_cs * engine,struct i915_request * rq) execlists_hold() argument 2188 execlists_unhold(struct intel_engine_cs * engine,struct i915_request * rq) execlists_unhold() argument 2218 struct intel_engine_cs *engine = cap->rq->engine; execlists_capture_work() local 2246 capture_regs(struct intel_engine_cs * engine) capture_regs() argument 2281 active_context(struct intel_engine_cs * engine,u32 ccid) active_context() argument 2314 active_ccid(struct intel_engine_cs * engine) active_ccid() argument 2319 execlists_capture(struct intel_engine_cs * engine) execlists_capture() argument 2380 execlists_reset(struct intel_engine_cs * engine,const char * msg) execlists_reset() argument 2404 preempt_timeout(const struct intel_engine_cs * const engine) preempt_timeout() argument 2425 struct intel_engine_cs * const engine = sched_engine->private_data; execlists_submission_tasklet() local 2480 execlists_irq_handler(struct intel_engine_cs * engine,u16 iir) execlists_irq_handler() argument 2521 struct intel_engine_cs *engine = __execlists_kick() local 2541 queue_request(struct intel_engine_cs * engine,struct i915_request * rq) queue_request() argument 2551 submit_queue(struct intel_engine_cs * engine,const struct i915_request * rq) submit_queue() argument 2563 ancestor_on_hold(const struct intel_engine_cs * engine,const struct i915_request * rq) ancestor_on_hold() argument 2572 struct intel_engine_cs *engine = request->engine; execlists_submit_request() local 2598 __execlists_context_pre_pin(struct intel_context * ce,struct intel_engine_cs * engine,struct i915_gem_ww_ctx * ww,void ** vaddr) __execlists_context_pre_pin() argument 2636 struct intel_engine_cs *engine = NULL; execlists_context_cancel_request() local 2707 const struct intel_engine_cs * const engine = rq->engine; emit_pdps() local 2798 reset_csb_pointers(struct intel_engine_cs * engine) reset_csb_pointers() argument 2840 sanitize_hwsp(struct intel_engine_cs * engine) sanitize_hwsp() argument 2848 execlists_sanitize(struct intel_engine_cs * engine) execlists_sanitize() argument 2879 enable_error_interrupt(struct intel_engine_cs * engine) enable_error_interrupt() argument 2922 enable_execlists(struct intel_engine_cs * engine) enable_execlists() argument 2946 execlists_resume(struct intel_engine_cs * engine) execlists_resume() argument 2959 execlists_reset_prepare(struct intel_engine_cs * engine) execlists_reset_prepare() argument 3002 reset_csb(struct intel_engine_cs * engine,struct i915_request ** inactive) reset_csb() argument 3018 execlists_reset_active(struct intel_engine_cs * engine,bool stalled) execlists_reset_active() argument 3095 execlists_reset_csb(struct intel_engine_cs * engine,bool stalled) execlists_reset_csb() argument 3111 execlists_reset_rewind(struct intel_engine_cs * engine,bool stalled) execlists_reset_rewind() argument 3132 struct intel_engine_cs * const engine = sched_engine->private_data; nop_submission_tasklet() local 3138 execlists_reset_cancel(struct intel_engine_cs * engine) execlists_reset_cancel() argument 3226 execlists_reset_finish(struct intel_engine_cs * engine) execlists_reset_finish() argument 3250 gen8_logical_ring_enable_irq(struct intel_engine_cs * engine) gen8_logical_ring_enable_irq() argument 3257 gen8_logical_ring_disable_irq(struct intel_engine_cs * engine) gen8_logical_ring_disable_irq() argument 3262 execlists_park(struct intel_engine_cs * engine) execlists_park() argument 3279 struct intel_engine_cs *engine, *locked; remove_from_engine() local 3307 can_preempt(struct intel_engine_cs * engine) can_preempt() argument 3314 struct intel_engine_cs *engine = rq->engine; kick_execlists() local 3362 execlists_set_default_submission(struct intel_engine_cs * engine) execlists_set_default_submission() argument 3370 execlists_shutdown(struct intel_engine_cs * engine) execlists_shutdown() argument 3378 execlists_release(struct intel_engine_cs * engine) execlists_release() argument 3388 __execlists_engine_busyness(struct intel_engine_cs * engine,ktime_t * now) __execlists_engine_busyness() argument 3405 execlists_engine_busyness(struct intel_engine_cs * engine,ktime_t * now) execlists_engine_busyness() argument 3421 logical_ring_default_vfuncs(struct intel_engine_cs * engine) logical_ring_default_vfuncs() argument 3487 logical_ring_default_irqs(struct intel_engine_cs * engine) logical_ring_default_irqs() argument 3509 rcs_submission_override(struct intel_engine_cs * engine) rcs_submission_override() argument 3527 intel_execlists_submission_setup(struct intel_engine_cs * engine) intel_execlists_submission_setup() argument 3746 virtual_get_sibling(struct intel_engine_cs * engine,unsigned int sibling) virtual_get_sibling() argument 4067 intel_execlists_show_requests(struct intel_engine_cs * engine,struct drm_printer * m,void (* show_request)(struct drm_printer * m,const struct i915_request * rq,const char * prefix,int indent),unsigned int max) intel_execlists_show_requests() argument 4152 intel_execlists_dump_active_requests(struct intel_engine_cs * engine,struct i915_request * hung_rq,struct drm_printer * m) intel_execlists_dump_active_requests() argument [all...] |
H A D | intel_engine_user.c | 39 void intel_engine_add_user(struct intel_engine_cs *engine) in intel_engine_add_user() argument 41 llist_add(&engine->uabi_llist, &engine->i915->uabi_engines_llist); in intel_engine_add_user() 87 struct intel_engine_cs *engine = in sort_engines() local 88 container_of(pos, typeof(*engine), uabi_llist); in sort_engines() 89 list_add(&engine->uabi_list, engines); in sort_engines() 97 u8 engine; in set_scheduler_caps() member 106 struct intel_engine_cs *engine; in set_scheduler_caps() local 111 for_each_uabi_engine(engine, i915) { /* all engines must agree! */ in set_scheduler_caps() 114 if (engine in set_scheduler_caps() 182 add_legacy_ring(struct legacy_ring * ring,struct intel_engine_cs * engine) add_legacy_ring() argument 195 engine_rename(struct intel_engine_cs * engine,const char * name,u16 instance) engine_rename() argument 217 struct intel_engine_cs *engine = intel_engines_driver_register() local 262 struct intel_engine_cs *engine; intel_engines_driver_register() local 317 struct intel_engine_cs *engine; intel_engines_has_context_isolation() local [all...] |
H A D | selftest_context.c | 75 static int __live_context_size(struct intel_engine_cs *engine) in __live_context_size() argument 82 ce = intel_context_create(engine); in __live_context_size() 91 intel_gt_coherent_map_type(engine->gt, in __live_context_size() 112 vaddr += engine->context_size - I915_GTT_PAGE_SIZE; in __live_context_size() 127 rq = intel_engine_create_kernel_request(engine); in __live_context_size() 137 pr_err("%s context overwrote trailing red-zone!", engine->name); in __live_context_size() 151 struct intel_engine_cs *engine; in live_context_size() local 160 for_each_engine(engine, gt, id) { in live_context_size() 163 if (!engine->context_size) in live_context_size() 166 intel_engine_pm_get(engine); in live_context_size() 195 __live_active_context(struct intel_engine_cs * engine) __live_active_context() argument 310 struct intel_engine_cs *engine; live_active_context() local 355 __live_remote_context(struct intel_engine_cs * engine) __live_remote_context() argument 423 struct intel_engine_cs *engine; live_remote_context() local [all...] |
H A D | selftest_engine_pm.c | 76 struct intel_engine_cs *engine = ce->engine; in __measure_timestamps() local 77 u32 *sema = memset32(engine->status_page.addr + 1000, 0, 5); in __measure_timestamps() 78 u32 offset = i915_ggtt_offset(engine->status_page.vma); in __measure_timestamps() 96 cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4000); in __measure_timestamps() 97 cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4004); in __measure_timestamps() 102 cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4016); in __measure_timestamps() 103 cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4012); in __measure_timestamps() 108 intel_engine_flush_submission(engine); in __measure_timestamps() 132 engine in __measure_timestamps() 139 __live_engine_timestamps(struct intel_engine_cs * engine) __live_engine_timestamps() argument 195 struct intel_engine_cs *engine; live_engine_timestamps() local 219 __spin_until_busier(struct intel_engine_cs * engine,ktime_t busyness) __spin_until_busier() argument 247 struct intel_engine_cs *engine; live_engine_busy_stats() local 352 struct intel_engine_cs *engine; live_engine_pm() local [all...] |
H A D | intel_lrc.c | 42 * [6:0]: Register offset, without considering the engine base. 49 const struct intel_engine_cs *engine, in set_offsets() argument 60 const u32 base = engine->mmio_base; in set_offsets() 78 if (GRAPHICS_VER(engine->i915) >= 11) in set_offsets() 101 if (GRAPHICS_VER(engine->i915) >= 11) in set_offsets() 641 static const u8 *reg_offsets(const struct intel_engine_cs *engine) in reg_offsets() argument 647 * physical engines for virtual engine. in reg_offsets() 649 GEM_BUG_ON(GRAPHICS_VER(engine->i915) >= 12 && in reg_offsets() 650 !intel_engine_has_relative_mmio(engine)); in reg_offsets() 652 if (engine in reg_offsets() 677 lrc_ring_mi_mode(const struct intel_engine_cs * engine) lrc_ring_mi_mode() argument 691 lrc_ring_bb_offset(const struct intel_engine_cs * engine) lrc_ring_bb_offset() argument 706 lrc_ring_gpr0(const struct intel_engine_cs * engine) lrc_ring_gpr0() argument 720 lrc_ring_wa_bb_per_ctx(const struct intel_engine_cs * engine) lrc_ring_wa_bb_per_ctx() argument 730 lrc_ring_indirect_ptr(const struct intel_engine_cs * engine) lrc_ring_indirect_ptr() argument 741 lrc_ring_indirect_offset(const struct intel_engine_cs * engine) lrc_ring_indirect_offset() argument 752 lrc_ring_cmd_buf_cctl(const struct intel_engine_cs * engine) lrc_ring_cmd_buf_cctl() argument 771 lrc_ring_indirect_offset_default(const struct intel_engine_cs * engine) lrc_ring_indirect_offset_default() argument 789 lrc_setup_bb_per_ctx(u32 * regs,const struct intel_engine_cs * engine,u32 ctx_bb_ggtt_addr) lrc_setup_bb_per_ctx() argument 801 lrc_setup_indirect_ctx(u32 * regs,const struct intel_engine_cs * engine,u32 ctx_bb_ggtt_addr,u32 size) lrc_setup_indirect_ctx() argument 841 init_common_regs(u32 * const regs,const struct intel_context * ce,const struct intel_engine_cs * engine,bool inhibit) init_common_regs() argument 867 init_wa_bb_regs(u32 * const regs,const struct intel_engine_cs * engine) init_wa_bb_regs() argument 911 __reset_stop_ring(u32 * regs,const struct intel_engine_cs * engine) __reset_stop_ring() argument 924 __lrc_init_regs(u32 * regs,const struct intel_context * ce,const struct intel_engine_cs * engine,bool inhibit) __lrc_init_regs() argument 952 lrc_init_regs(const struct intel_context * ce,const struct intel_engine_cs * engine,bool inhibit) lrc_init_regs() argument 959 lrc_reset_regs(const struct intel_context * ce,const struct intel_engine_cs * engine) lrc_reset_regs() argument 965 set_redzone(void * vaddr,const struct intel_engine_cs * engine) set_redzone() argument 976 check_redzone(const void * vaddr,const struct intel_engine_cs * engine) check_redzone() argument 1015 lrc_init_state(struct intel_context * ce,struct intel_engine_cs * engine,void * state) lrc_init_state() argument 1072 __lrc_alloc_state(struct intel_context * ce,struct intel_engine_cs * engine) __lrc_alloc_state() argument 1120 pinned_timeline(struct intel_context * ce,struct intel_engine_cs * engine) pinned_timeline() argument 1127 lrc_alloc(struct intel_context * ce,struct intel_engine_cs * engine) lrc_alloc() argument 1192 lrc_pre_pin(struct intel_context * ce,struct intel_engine_cs * engine,struct i915_gem_ww_ctx * ww,void ** vaddr) lrc_pre_pin() argument 1210 lrc_pin(struct intel_context * ce,struct intel_engine_cs * engine,void * vaddr) lrc_pin() argument 1447 setup_per_ctx_bb(const struct intel_context * ce,const struct intel_engine_cs * engine,u32 * (* emit)(const struct intel_context *,u32 *)) setup_per_ctx_bb() argument 1466 setup_indirect_ctx_bb(const struct intel_context * ce,const struct intel_engine_cs * engine,u32 * (* emit)(const struct intel_context *,u32 *)) setup_indirect_ctx_bb() argument 1536 lrc_update_regs(const struct intel_context * ce,const struct intel_engine_cs * engine,u32 head) lrc_update_regs() argument 1575 lrc_update_offsets(struct intel_context * ce,struct intel_engine_cs * engine) lrc_update_offsets() argument 1581 lrc_check_regs(const struct intel_context * ce,const struct intel_engine_cs * engine,const char * when) lrc_check_regs() argument 1637 gen8_emit_flush_coherentl3_wa(struct intel_engine_cs * engine,u32 * batch) gen8_emit_flush_coherentl3_wa() argument 1679 gen8_init_indirectctx_bb(struct intel_engine_cs * engine,u32 * batch) gen8_init_indirectctx_bb() argument 1731 gen9_init_indirectctx_bb(struct intel_engine_cs * engine,u32 * batch) gen9_init_indirectctx_bb() argument 1805 lrc_create_wa_ctx(struct intel_engine_cs * engine) lrc_create_wa_ctx() argument 1829 lrc_fini_wa_ctx(struct intel_engine_cs * engine) lrc_fini_wa_ctx() argument 1836 lrc_init_wa_ctx(struct intel_engine_cs * engine) lrc_init_wa_ctx() argument [all...] |
H A D | selftest_hangcheck.c | 104 hang_create_request(struct hang *h, struct intel_engine_cs *engine) in hang_create_request() argument 157 rq = igt_request_alloc(h->ctx, engine); in hang_create_request() 226 intel_gt_chipset_flush(engine->gt); in hang_create_request() 228 if (rq->engine->emit_init_breadcrumb) { in hang_create_request() 229 err = rq->engine->emit_init_breadcrumb(rq); in hang_create_request() 238 err = rq->engine->emit_bb_start(rq, i915_vma_offset(vma), PAGE_SIZE, flags); in hang_create_request() 288 struct intel_engine_cs *engine; in igt_hang_sanitycheck() local 299 for_each_engine(engine, gt, id) { in igt_hang_sanitycheck() 303 if (!intel_engine_can_store_dword(engine)) in igt_hang_sanitycheck() 306 rq = hang_create_request(&h, engine); in igt_hang_sanitycheck() 343 wait_for_idle(struct intel_engine_cs * engine) wait_for_idle() argument 352 struct intel_engine_cs *engine; igt_reset_nop() local 428 struct intel_engine_cs *engine; igt_reset_nop_engine() local 534 force_reset_timeout(struct intel_engine_cs * engine) force_reset_timeout() argument 540 cancel_reset_timeout(struct intel_engine_cs * engine) cancel_reset_timeout() argument 548 struct intel_engine_cs *engine; igt_reset_fail_engine() local 687 struct intel_engine_cs *engine; __igt_reset_engine() local 856 struct intel_engine_cs *engine; global() member 895 struct intel_engine_cs *engine = arg->engine; active_engine() local 967 struct intel_engine_cs *engine, *other; __igt_reset_engines() local 1299 struct intel_engine_cs *engine; igt_reset_wait() local 1431 struct intel_engine_cs *engine; __igt_reset_evict_vma() local 1621 struct intel_engine_cs *engine; wait_for_others() local 1639 struct intel_engine_cs *engine; igt_reset_queue() local 1811 struct intel_engine_cs *engine; igt_handle_error() local 1876 __igt_atomic_reset_engine(struct intel_engine_cs * engine,const struct igt_atomic_section * p,const char * mode) __igt_atomic_reset_engine() argument 1909 igt_atomic_reset_engine(struct intel_engine_cs * engine,const struct igt_atomic_section * p) igt_atomic_reset_engine() argument 1982 struct intel_engine_cs *engine; igt_reset_engines_atomic() local [all...] |
H A D | selftest_workarounds.c | 34 } engine[I915_NUM_ENGINES]; member 64 struct intel_engine_cs *engine; in reference_lists_init() local 73 for_each_engine(engine, gt, id) { in reference_lists_init() 74 struct i915_wa_list *wal = &lists->engine[id].wa_list; in reference_lists_init() 76 wa_init_start(wal, gt, "REF", engine->name); in reference_lists_init() 77 engine_init_workarounds(engine, wal); in reference_lists_init() 80 __intel_engine_init_ctx_wa(engine, in reference_lists_init() 81 &lists->engine[id].ctx_wa_list, in reference_lists_init() 89 struct intel_engine_cs *engine; in reference_lists_fini() local 92 for_each_engine(engine, g in reference_lists_fini() 101 struct intel_engine_cs *engine = ce->engine; read_nonprivs() local 178 get_whitelist_reg(const struct intel_engine_cs * engine,unsigned int i) get_whitelist_reg() argument 188 print_results(const struct intel_engine_cs * engine,const u32 * results) print_results() argument 203 struct intel_engine_cs *engine = ce->engine; check_whitelist() local 251 do_device_reset(struct intel_engine_cs * engine) do_device_reset() argument 257 do_engine_reset(struct intel_engine_cs * engine) do_engine_reset() argument 262 do_guc_reset(struct intel_engine_cs * engine) do_guc_reset() argument 269 switch_to_scratch_context(struct intel_engine_cs * engine,struct igt_spinner * spin,struct i915_request ** rq) switch_to_scratch_context() argument 297 check_whitelist_across_reset(struct intel_engine_cs * engine,int (* reset)(struct intel_engine_cs *),const char * name) check_whitelist_across_reset() argument 418 wo_register(struct intel_engine_cs * engine,u32 reg) wo_register() argument 436 timestamp(const struct intel_engine_cs * engine,u32 reg) timestamp() argument 459 whitelist_writable_count(struct intel_engine_cs * engine) whitelist_writable_count() argument 502 struct intel_engine_cs *engine = ce->engine; check_dirty_whitelist() local 759 struct intel_engine_cs *engine; live_dirty_whitelist() local 790 struct intel_engine_cs *engine; live_reset_whitelist() local 845 struct intel_engine_cs *engine = ce->engine; read_whitelisted_registers() local 888 struct intel_engine_cs *engine = ce->engine; scrub_whitelisted_registers() local 984 result_eq(struct intel_engine_cs * engine,u32 a,u32 b,i915_reg_t reg) result_eq() argument 1006 result_neq(struct intel_engine_cs * engine,u32 a,u32 b,i915_reg_t reg) result_neq() argument 1019 check_whitelisted_registers(struct intel_engine_cs * engine,struct i915_vma * A,struct i915_vma * B,bool (* fn)(struct intel_engine_cs * engine,u32 a,u32 b,i915_reg_t reg)) check_whitelisted_registers() argument 1022 check_whitelisted_registers(struct intel_engine_cs * engine,struct i915_vma * A,struct i915_vma * B,bool (* fn)(struct intel_engine_cs * engine,u32 a,u32 b,i915_reg_t reg)) check_whitelisted_registers() argument 1063 struct intel_engine_cs *engine; live_isolated_whitelist() local 1169 struct intel_engine_cs *engine; verify_wa_lists() local 1239 struct intel_engine_cs *engine; live_engine_reset_workarounds() local [all...] |
H A D | intel_reset.c | 276 * On some platforms, e.g. Jasperlake, we see that the engine register in gen6_hw_domain_reset() 283 * engine state is still in flux and not ready for resuming. in gen6_hw_domain_reset() 313 * As we have observed that the engine state is still volatile in gen6_hw_domain_reset() 325 struct intel_engine_cs *engine; in __gen6_reset_engines() local 334 for_each_engine_masked(engine, gt, engine_mask, tmp) { in __gen6_reset_engines() 335 hw_mask |= engine->reset_domain; in __gen6_reset_engines() 356 static struct intel_engine_cs *find_sfc_paired_vecs_engine(struct intel_engine_cs *engine) in find_sfc_paired_vecs_engine() argument 360 GEM_BUG_ON(engine->class != VIDEO_DECODE_CLASS); in find_sfc_paired_vecs_engine() 362 vecs_id = _VECS((engine->instance) / 2); in find_sfc_paired_vecs_engine() 364 return engine in find_sfc_paired_vecs_engine() 377 get_sfc_forced_lock_data(struct intel_engine_cs * engine,struct sfc_lock_data * sfc_lock) get_sfc_forced_lock_data() argument 411 gen11_lock_sfc(struct intel_engine_cs * engine,u32 * reset_mask,u32 * unlock_mask) gen11_lock_sfc() argument 503 gen11_unlock_sfc(struct intel_engine_cs * engine) gen11_unlock_sfc() argument 526 struct intel_engine_cs *engine; __gen11_reset_engines() local 563 gen8_engine_reset_prepare(struct intel_engine_cs * engine) gen8_engine_reset_prepare() argument 604 gen8_engine_reset_cancel(struct intel_engine_cs * engine) gen8_engine_reset_cancel() argument 615 struct intel_engine_cs *engine; gen8_reset_engines() local 831 reset_prepare_engine(struct intel_engine_cs * engine) reset_prepare_engine() argument 878 struct intel_engine_cs *engine; reset_prepare() local 910 struct intel_engine_cs *engine; gt_reset() local 934 reset_finish_engine(struct intel_engine_cs * engine) reset_finish_engine() argument 945 struct intel_engine_cs *engine; reset_finish() local 972 struct intel_engine_cs *engine; __intel_gt_set_wedged() local 1038 struct intel_engine_cs *engine; intel_gt_set_wedged() local 1167 struct intel_engine_cs *engine; resume() local 1325 intel_gt_reset_engine(struct intel_engine_cs * engine) intel_gt_reset_engine() argument 1330 __intel_engine_reset_bh(struct intel_engine_cs * engine,const char * msg) __intel_engine_reset_bh() argument 1392 intel_engine_reset(struct intel_engine_cs * engine,const char * msg) intel_engine_reset() argument 1473 struct intel_engine_cs *engine; intel_gt_handle_error() local [all...] |
H A D | sysfs_engines.c | 16 struct intel_engine_cs *engine; member 21 return container_of(kobj, struct kobj_engine, base)->engine; in kobj_to_engine() 81 __caps_show(struct intel_engine_cs *engine, in __caps_show() argument 88 switch (engine->class) { in __caps_show() 123 struct intel_engine_cs *engine = kobj_to_engine(kobj); in caps_show() local 125 return __caps_show(engine, engine->uabi_capabilities, buf, true); in caps_show() 144 struct intel_engine_cs *engine = kobj_to_engine(kobj); in max_spin_store() local 169 clamped = intel_clamp_max_busywait_duration_ns(engine, duration); in max_spin_store() 173 WRITE_ONCE(engine in max_spin_store() 181 struct intel_engine_cs *engine = kobj_to_engine(kobj); max_spin_show() local 192 struct intel_engine_cs *engine = kobj_to_engine(kobj); max_spin_default() local 204 struct intel_engine_cs *engine = kobj_to_engine(kobj); timeslice_store() local 235 struct intel_engine_cs *engine = kobj_to_engine(kobj); timeslice_show() local 246 struct intel_engine_cs *engine = kobj_to_engine(kobj); timeslice_default() local 258 struct intel_engine_cs *engine = kobj_to_engine(kobj); stop_store() local 286 struct intel_engine_cs *engine = kobj_to_engine(kobj); stop_show() local 297 struct intel_engine_cs *engine = kobj_to_engine(kobj); stop_default() local 309 struct intel_engine_cs *engine = kobj_to_engine(kobj); preempt_timeout_store() local 342 struct intel_engine_cs *engine = kobj_to_engine(kobj); preempt_timeout_show() local 354 struct intel_engine_cs *engine = kobj_to_engine(kobj); preempt_timeout_default() local 366 struct intel_engine_cs *engine = kobj_to_engine(kobj); heartbeat_store() local 398 struct intel_engine_cs *engine = kobj_to_engine(kobj); heartbeat_show() local 409 struct intel_engine_cs *engine = kobj_to_engine(kobj); heartbeat_default() local 428 kobj_engine(struct kobject * dir,struct intel_engine_cs * engine) kobj_engine() argument 502 struct intel_engine_cs *engine; intel_engines_add_sysfs() local [all...] |
/linux/crypto/ |
H A D | crypto_engine.c | 3 * Handle async block request by crypto hardware engine. 12 #include <crypto/internal/engine.h> 33 * @engine: the hardware engine 37 static void crypto_finalize_request(struct crypto_engine *engine, in crypto_finalize_request() argument 47 if (!engine->retry_support) { in crypto_finalize_request() 48 spin_lock_irqsave(&engine->queue_lock, flags); in crypto_finalize_request() 49 if (engine->cur_req == req) { in crypto_finalize_request() 50 engine->cur_req = NULL; in crypto_finalize_request() 52 spin_unlock_irqrestore(&engine in crypto_finalize_request() 70 crypto_pump_requests(struct crypto_engine * engine,bool in_kthread) crypto_pump_requests() argument 176 struct crypto_engine *engine = crypto_pump_work() local 188 crypto_transfer_request(struct crypto_engine * engine,struct crypto_async_request * req,bool need_pump) crypto_transfer_request() argument 217 crypto_transfer_request_to_engine(struct crypto_engine * engine,struct crypto_async_request * req) crypto_transfer_request_to_engine() argument 229 crypto_transfer_aead_request_to_engine(struct crypto_engine * engine,struct aead_request * req) crypto_transfer_aead_request_to_engine() argument 242 crypto_transfer_akcipher_request_to_engine(struct crypto_engine * engine,struct akcipher_request * req) crypto_transfer_akcipher_request_to_engine() argument 255 crypto_transfer_hash_request_to_engine(struct crypto_engine * engine,struct ahash_request * req) crypto_transfer_hash_request_to_engine() argument 268 crypto_transfer_kpp_request_to_engine(struct crypto_engine * engine,struct kpp_request * req) crypto_transfer_kpp_request_to_engine() argument 281 crypto_transfer_skcipher_request_to_engine(struct crypto_engine * engine,struct skcipher_request * req) crypto_transfer_skcipher_request_to_engine() argument 295 crypto_finalize_aead_request(struct crypto_engine * engine,struct aead_request * req,int err) crypto_finalize_aead_request() argument 309 crypto_finalize_akcipher_request(struct crypto_engine * engine,struct akcipher_request * req,int err) crypto_finalize_akcipher_request() argument 323 crypto_finalize_hash_request(struct crypto_engine * engine,struct ahash_request * req,int err) crypto_finalize_hash_request() argument 336 crypto_finalize_kpp_request(struct crypto_engine * engine,struct kpp_request * req,int err) crypto_finalize_kpp_request() argument 350 crypto_finalize_skcipher_request(struct crypto_engine * engine,struct skcipher_request * req,int err) crypto_finalize_skcipher_request() argument 363 crypto_engine_start(struct crypto_engine * engine) crypto_engine_start() argument 389 crypto_engine_stop(struct crypto_engine * engine) crypto_engine_stop() argument 437 struct crypto_engine *engine; crypto_engine_alloc_init_and_set() local 495 crypto_engine_exit(struct crypto_engine * engine) crypto_engine_exit() argument [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | engine.c | 24 #include <core/engine.h> 31 nvkm_engine_chsw_load(struct nvkm_engine *engine) in nvkm_engine_chsw_load() argument 33 if (engine->func->chsw_load) in nvkm_engine_chsw_load() 34 return engine->func->chsw_load(engine); in nvkm_engine_chsw_load() 39 nvkm_engine_reset(struct nvkm_engine *engine) in nvkm_engine_reset() argument 41 if (engine->func->reset) in nvkm_engine_reset() 42 return engine->func->reset(engine); in nvkm_engine_reset() 44 nvkm_subdev_fini(&engine in nvkm_engine_reset() 51 struct nvkm_engine *engine = *pengine; nvkm_engine_unref() local 60 nvkm_engine_ref(struct nvkm_engine * engine) nvkm_engine_ref() argument 74 nvkm_engine_tile(struct nvkm_engine * engine,int region) nvkm_engine_tile() argument 84 struct nvkm_engine *engine = nvkm_engine(subdev); nvkm_engine_intr() local 92 struct nvkm_engine *engine = nvkm_engine(subdev); nvkm_engine_info() local 103 struct nvkm_engine *engine = nvkm_engine(subdev); nvkm_engine_fini() local 112 struct nvkm_engine *engine = nvkm_engine(subdev); nvkm_engine_init() local 127 struct nvkm_engine *engine = nvkm_engine(subdev); nvkm_engine_oneinit() local 138 struct nvkm_engine *engine = nvkm_engine(subdev); nvkm_engine_preinit() local 147 struct nvkm_engine *engine = nvkm_engine(subdev); nvkm_engine_dtor() local 166 nvkm_engine_ctor(const struct nvkm_engine_func * func,struct nvkm_device * device,enum nvkm_subdev_type type,int inst,bool enable,struct nvkm_engine * engine) nvkm_engine_ctor() argument [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
H A D | Kbuild | 2 nvkm-y += nvkm/engine/gr/base.o 3 nvkm-y += nvkm/engine/gr/nv04.o 4 nvkm-y += nvkm/engine/gr/nv10.o 5 nvkm-y += nvkm/engine/gr/nv15.o 6 nvkm-y += nvkm/engine/gr/nv17.o 7 nvkm-y += nvkm/engine/gr/nv20.o 8 nvkm-y += nvkm/engine/gr/nv25.o 9 nvkm-y += nvkm/engine/gr/nv2a.o 10 nvkm-y += nvkm/engine/gr/nv30.o 11 nvkm-y += nvkm/engine/g [all...] |
/linux/drivers/gpu/drm/sun4i/ |
H A D | sunxi_engine.h | 27 * This callback allows to prepare our engine for an atomic 34 void (*atomic_begin)(struct sunxi_engine *engine, 51 int (*atomic_check)(struct sunxi_engine *engine, 63 void (*commit)(struct sunxi_engine *engine, 71 * the layers supported by that engine. 81 struct sunxi_engine *engine); 87 * engine. This is useful only for the composite output. 91 void (*apply_color_correction)(struct sunxi_engine *engine); 97 * engine. This is useful only for the composite output. 101 void (*disable_color_correction)(struct sunxi_engine *engine); 154 sunxi_engine_commit(struct sunxi_engine * engine,struct drm_crtc * crtc,struct drm_atomic_state * state) sunxi_engine_commit() argument 168 sunxi_engine_layers_init(struct drm_device * drm,struct sunxi_engine * engine) sunxi_engine_layers_init() argument 185 sunxi_engine_apply_color_correction(struct sunxi_engine * engine) sunxi_engine_apply_color_correction() argument 198 sunxi_engine_disable_color_correction(struct sunxi_engine * engine) sunxi_engine_disable_color_correction() argument 212 sunxi_engine_mode_set(struct sunxi_engine * engine,const struct drm_display_mode * mode) sunxi_engine_mode_set() argument [all...] |
/linux/drivers/video/fbdev/via/ |
H A D | accel.c | 13 static int viafb_set_bpp(void __iomem *engine, u8 bpp) in viafb_set_bpp() argument 19 gemode = readl(engine + VIA_REG_GEMODE) & 0xfffffcfc; in viafb_set_bpp() 34 writel(gemode, engine + VIA_REG_GEMODE); in viafb_set_bpp() 39 static int hw_bitblt_1(void __iomem *engine, u8 op, u32 width, u32 height, in hw_bitblt_1() argument 79 ret = viafb_set_bpp(engine, dst_bpp); in hw_bitblt_1() 91 writel(tmp, engine + 0x08); in hw_bitblt_1() 100 writel(tmp, engine + 0x0C); in hw_bitblt_1() 108 writel(tmp, engine + 0x10); in hw_bitblt_1() 111 writel(fg_color, engine + 0x18); in hw_bitblt_1() 114 writel(bg_color, engine in hw_bitblt_1() 172 hw_bitblt_2(void __iomem * engine,u8 op,u32 width,u32 height,u8 dst_bpp,u32 dst_addr,u32 dst_pitch,u32 dst_x,u32 dst_y,u32 * src_mem,u32 src_addr,u32 src_pitch,u32 src_x,u32 src_y,u32 fg_color,u32 bg_color,u8 fill_rop) hw_bitblt_2() argument 307 void __iomem *engine; viafb_setup_engine() local 369 void __iomem *engine = viapar->shared->vdev->engine_mmio; viafb_reset_engine() local 508 void __iomem *engine = viapar->shared->vdev->engine_mmio; viafb_wait_engine_idle() local [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
H A D | Kbuild | 2 nvkm-y += nvkm/engine/disp/base.o 3 nvkm-y += nvkm/engine/disp/chan.o 4 nvkm-y += nvkm/engine/disp/conn.o 5 nvkm-y += nvkm/engine/disp/dp.o 6 nvkm-y += nvkm/engine/disp/hdmi.o 7 nvkm-y += nvkm/engine/disp/head.o 8 nvkm-y += nvkm/engine/disp/ior.o 9 nvkm-y += nvkm/engine/disp/outp.o 10 nvkm-y += nvkm/engine/disp/vga.o 12 nvkm-y += nvkm/engine/dis [all...] |
/linux/drivers/crypto/marvell/cesa/ |
H A D | cesa.c | 5 * driver supports the TDMA engine on platforms on which it is available. 38 mv_cesa_dequeue_req_locked(struct mv_cesa_engine *engine, in mv_cesa_dequeue_req_locked() argument 43 *backlog = crypto_get_backlog(&engine->queue); in mv_cesa_dequeue_req_locked() 44 req = crypto_dequeue_request(&engine->queue); in mv_cesa_dequeue_req_locked() 52 static void mv_cesa_rearm_engine(struct mv_cesa_engine *engine) in mv_cesa_rearm_engine() argument 58 spin_lock_bh(&engine->lock); in mv_cesa_rearm_engine() 59 if (!engine->req) { in mv_cesa_rearm_engine() 60 req = mv_cesa_dequeue_req_locked(engine, &backlog); in mv_cesa_rearm_engine() 61 engine->req = req; in mv_cesa_rearm_engine() 63 spin_unlock_bh(&engine in mv_cesa_rearm_engine() 75 mv_cesa_std_process(struct mv_cesa_engine * engine,u32 status) mv_cesa_std_process() argument 95 mv_cesa_int_process(struct mv_cesa_engine * engine,u32 status) mv_cesa_int_process() argument 115 struct mv_cesa_engine *engine = priv; mv_cesa_int() local 173 struct mv_cesa_engine *engine = creq->engine; mv_cesa_queue_req() local 315 mv_cesa_conf_mbus_windows(struct mv_cesa_engine * engine,const struct mbus_dram_target_info * dram) mv_cesa_conf_mbus_windows() argument 377 struct mv_cesa_engine *engine = &cesa->engines[idx]; mv_cesa_get_sram() local 409 struct mv_cesa_engine *engine = &cesa->engines[idx]; mv_cesa_put_sram() local 477 struct mv_cesa_engine *engine = &cesa->engines[i]; mv_cesa_probe() local [all...] |
/linux/drivers/gpu/drm/i915/selftests/ |
H A D | intel_scheduler_helpers.c | 21 struct intel_engine_cs *engine; in intel_selftest_find_any_engine() local 24 for_each_engine(engine, gt, id) in intel_selftest_find_any_engine() 25 return engine; in intel_selftest_find_any_engine() 27 pr_err("No valid engine found!\n"); in intel_selftest_find_any_engine() 31 int intel_selftest_modify_policy(struct intel_engine_cs *engine, in intel_selftest_modify_policy() argument 37 saved->reset = engine->i915->params.reset; in intel_selftest_modify_policy() 38 saved->flags = engine->flags; in intel_selftest_modify_policy() 39 saved->timeslice = engine->props.timeslice_duration_ms; in intel_selftest_modify_policy() 40 saved->preempt_timeout = engine->props.preempt_timeout_ms; in intel_selftest_modify_policy() 46 * together with engine rese in intel_selftest_modify_policy() 77 intel_selftest_restore_policy(struct intel_engine_cs * engine,struct intel_selftest_saved_policy * saved) intel_selftest_restore_policy() argument [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | Kbuild | 2 nvkm-y += nvkm/engine/fifo/base.o 3 nvkm-y += nvkm/engine/fifo/cgrp.o 4 nvkm-y += nvkm/engine/fifo/chan.o 5 nvkm-y += nvkm/engine/fifo/chid.o 6 nvkm-y += nvkm/engine/fifo/runl.o 7 nvkm-y += nvkm/engine/fifo/runq.o 9 nvkm-y += nvkm/engine/fifo/nv04.o 10 nvkm-y += nvkm/engine/fifo/nv10.o 11 nvkm-y += nvkm/engine/fifo/nv17.o 12 nvkm-y += nvkm/engine/fif [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/engine/ |
H A D | Kbuild | 2 nvkm-y += nvkm/engine/falcon.o 3 nvkm-y += nvkm/engine/xtensa.o 5 include $(src)/nvkm/engine/bsp/Kbuild 6 include $(src)/nvkm/engine/ce/Kbuild 7 include $(src)/nvkm/engine/cipher/Kbuild 8 include $(src)/nvkm/engine/device/Kbuild 9 include $(src)/nvkm/engine/disp/Kbuild 10 include $(src)/nvkm/engine/dma/Kbuild 11 include $(src)/nvkm/engine/fifo/Kbuild 12 include $(src)/nvkm/engine/g [all...] |