/linux/tools/perf/util/ |
H A D | intel-pt.c | 35 #include "intel-pt.h" 43 #include "intel-pt-decoder/intel-pt-log.h" 44 #include "intel-pt-decoder/intel-pt-decoder.h" 45 #include "intel-pt-decoder/intel-pt-insn-decoder.h" 46 #include "intel-pt-decoder/intel-pt-pkt-decoder.h" 183 struct intel_pt *pt; member 273 intel_pt_dump_event(struct intel_pt * pt,unsigned char * buf,size_t len) intel_pt_dump_event() argument 293 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_dump_sample() local 300 intel_pt_log_events(struct intel_pt * pt,u64 tm) intel_pt_log_events() argument 355 struct intel_pt *pt = ptq->pt; intel_pt_findnew_vmcs_info() local 363 intel_pt_free_vmcs_info(struct intel_pt * pt) intel_pt_free_vmcs_info() argument 377 intel_pt_do_fix_overlap(struct intel_pt * pt,struct auxtrace_buffer * a,struct auxtrace_buffer * b) intel_pt_do_fix_overlap() argument 947 intel_pt_match_pgd_ip(struct intel_pt * pt,uint64_t ip,uint64_t offset,const char * filename) intel_pt_match_pgd_ip() argument 1025 intel_pt_get_config(struct intel_pt * pt,struct perf_event_attr * attr,u64 * config) intel_pt_get_config() argument 1037 intel_pt_exclude_kernel(struct intel_pt * pt) intel_pt_exclude_kernel() argument 1049 intel_pt_return_compression(struct intel_pt * pt) intel_pt_return_compression() argument 1065 intel_pt_branch_enable(struct intel_pt * pt) intel_pt_branch_enable() argument 1079 intel_pt_disabled_tnt(struct intel_pt * pt) intel_pt_disabled_tnt() argument 1092 intel_pt_mtc_period(struct intel_pt * pt) intel_pt_mtc_period() argument 1111 intel_pt_timeless_decoding(struct intel_pt * pt) intel_pt_timeless_decoding() argument 1133 intel_pt_tracing_kernel(struct intel_pt * pt) intel_pt_tracing_kernel() argument 1145 intel_pt_have_tsc(struct intel_pt * pt) intel_pt_have_tsc() argument 1165 intel_pt_have_mtc(struct intel_pt * pt) intel_pt_have_mtc() argument 1178 intel_pt_sampling_mode(struct intel_pt * pt) intel_pt_sampling_mode() argument 1190 intel_pt_ctl(struct intel_pt * pt) intel_pt_ctl() argument 1202 intel_pt_ns_to_ticks(const struct intel_pt * pt,u64 ns) intel_pt_ns_to_ticks() argument 1212 intel_pt_alloc_chain(struct intel_pt * pt) intel_pt_alloc_chain() argument 1221 intel_pt_callchain_init(struct intel_pt * pt) intel_pt_callchain_init() argument 1237 intel_pt_add_callchain(struct intel_pt * pt,struct perf_sample * sample) intel_pt_add_callchain() argument 1259 intel_pt_br_stack_init(struct intel_pt * pt) intel_pt_br_stack_init() argument 1275 intel_pt_add_br_stack(struct intel_pt * pt,struct perf_sample * sample) intel_pt_add_br_stack() argument 1293 intel_pt_alloc_queue(struct intel_pt * pt,unsigned int queue_nr) intel_pt_alloc_queue() argument 1416 intel_pt_first_timestamp(struct intel_pt * pt,u64 timestamp) intel_pt_first_timestamp() argument 1480 intel_pt_set_pid_tid_cpu(struct intel_pt * pt,struct auxtrace_queue * queue) intel_pt_set_pid_tid_cpu() argument 1511 struct intel_pt *pt = ptq->pt; intel_pt_sample_flags() local 1558 intel_pt_setup_time_range(struct intel_pt * pt,struct intel_pt_queue * ptq) intel_pt_setup_time_range() argument 1575 intel_pt_setup_queue(struct intel_pt * pt,struct auxtrace_queue * queue,unsigned int queue_nr) intel_pt_setup_queue() argument 1657 intel_pt_setup_queues(struct intel_pt * pt) intel_pt_setup_queues() argument 1670 intel_pt_skip_event(struct intel_pt * pt) intel_pt_skip_event() argument 1681 intel_pt_skip_cbr_event(struct intel_pt * pt) intel_pt_skip_cbr_event() argument 1712 intel_pt_prep_b_sample(struct intel_pt * pt,struct intel_pt_queue * ptq,union perf_event * event,struct perf_sample * sample) intel_pt_prep_b_sample() argument 1738 intel_pt_opt_inject(struct intel_pt * pt,union perf_event * event,struct perf_sample * sample,u64 type) intel_pt_opt_inject() argument 1748 intel_pt_deliver_synth_event(struct intel_pt * pt,union perf_event * event,struct perf_sample * sample,u64 type) intel_pt_deliver_synth_event() argument 1767 struct intel_pt *pt = ptq->pt; intel_pt_synth_branch_sample() local 1819 intel_pt_prep_sample(struct intel_pt * pt,struct intel_pt_queue * ptq,union perf_event * event,struct perf_sample * sample) intel_pt_prep_sample() argument 1842 struct intel_pt *pt = ptq->pt; intel_pt_synth_instruction_sample() local 1878 struct intel_pt *pt = ptq->pt; intel_pt_synth_cycle_sample() local 1909 struct intel_pt *pt = ptq->pt; intel_pt_synth_transaction_sample() local 1929 intel_pt_prep_p_sample(struct intel_pt * pt,struct intel_pt_queue * ptq,union perf_event * event,struct perf_sample * sample) intel_pt_prep_p_sample() argument 1946 struct intel_pt *pt = ptq->pt; intel_pt_synth_ptwrite_sample() local 1972 struct intel_pt *pt = ptq->pt; intel_pt_synth_cbr_sample() local 2006 struct intel_pt *pt = ptq->pt; intel_pt_synth_psb_sample() local 2036 struct intel_pt *pt = ptq->pt; intel_pt_synth_mwait_sample() local 2065 struct intel_pt *pt = ptq->pt; intel_pt_synth_pwre_sample() local 2094 struct intel_pt *pt = ptq->pt; intel_pt_synth_exstop_sample() local 2123 struct intel_pt *pt = ptq->pt; intel_pt_synth_pwrx_sample() local 2421 struct intel_pt *pt = ptq->pt; intel_pt_do_synth_pebs_sample() local 2567 struct intel_pt *pt = ptq->pt; intel_pt_synth_single_pebs_sample() local 2579 struct intel_pt *pt = ptq->pt; intel_pt_synth_pebs_sample() local 2607 struct intel_pt *pt = ptq->pt; intel_pt_synth_events_sample() local 2649 struct intel_pt *pt = ptq->pt; intel_pt_synth_iflag_chg_sample() local 2684 intel_pt_synth_error(struct intel_pt * pt,int code,int cpu,pid_t pid,pid_t tid,u64 ip,u64 timestamp,pid_t machine_pid,int vcpu) intel_pt_synth_error() argument 2730 struct intel_pt *pt = ptq->pt; intel_ptq_synth_error() local 2750 intel_pt_next_tid(struct intel_pt * pt,struct intel_pt_queue * ptq) intel_pt_next_tid() argument 2773 struct intel_pt *pt = ptq->pt; intel_pt_is_switch_ip() local 2787 struct intel_pt *pt = ptq->pt; intel_pt_sample() local 2961 intel_pt_switch_ip(struct intel_pt * pt,u64 * ptss_ip) intel_pt_switch_ip() argument 3013 intel_pt_enable_sync_switch(struct intel_pt * pt) intel_pt_enable_sync_switch() argument 3031 intel_pt_disable_sync_switch(struct intel_pt * pt) intel_pt_disable_sync_switch() argument 3054 struct intel_pt *pt = ptq->pt; intel_pt_next_time() local 3123 struct intel_pt *pt = ptq->pt; intel_pt_run_decoder() local 3204 intel_pt_update_queues(struct intel_pt * pt) intel_pt_update_queues() argument 3213 intel_pt_process_queues(struct intel_pt * pt,u64 timestamp) intel_pt_process_queues() argument 3268 intel_pt_process_timeless_queues(struct intel_pt * pt,pid_t tid,u64 time_) intel_pt_process_timeless_queues() argument 3316 intel_pt_process_timeless_sample(struct intel_pt * pt,struct perf_sample * sample) intel_pt_process_timeless_sample() argument 3338 intel_pt_lost(struct intel_pt * pt,struct perf_sample * sample) intel_pt_lost() argument 3345 intel_pt_cpu_to_ptq(struct intel_pt * pt,int cpu) intel_pt_cpu_to_ptq() argument 3373 intel_pt_sync_switch(struct intel_pt * pt,int cpu,pid_t tid,u64 timestamp) intel_pt_sync_switch() argument 3419 intel_pt_process_switch(struct intel_pt * pt,struct perf_sample * sample) intel_pt_process_switch() argument 3444 intel_pt_context_switch_in(struct intel_pt * pt,struct perf_sample * sample) intel_pt_context_switch_in() argument 3482 intel_pt_guest_context_switch(struct intel_pt * pt,union perf_event * event,struct perf_sample * sample) intel_pt_guest_context_switch() argument 3509 intel_pt_context_switch(struct intel_pt * pt,union perf_event * event,struct perf_sample * sample) intel_pt_context_switch() argument 3547 intel_pt_process_itrace_start(struct intel_pt * pt,union perf_event * event,struct perf_sample * sample) intel_pt_process_itrace_start() argument 3568 intel_pt_data_src_fmt(struct intel_pt * pt,struct evsel * evsel) intel_pt_data_src_fmt() argument 3607 intel_pt_process_aux_output_hw_id(struct intel_pt * pt,union perf_event * event,struct perf_sample * sample) intel_pt_process_aux_output_hw_id() argument 3644 intel_pt_text_poke(struct intel_pt * pt,union perf_event * event) intel_pt_text_poke() argument 3704 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_process_event() local 3788 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_flush() local 3811 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_free_events() local 3826 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_free() local 3845 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_evsel_is_auxtrace() local 3855 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_process_auxtrace_event() local 3894 struct intel_pt *pt = container_of(session->auxtrace, struct intel_pt, intel_pt_queue_data() local 3943 intel_pt_evsel(struct intel_pt * pt,struct evlist * evlist) intel_pt_evsel() argument 3956 intel_pt_synth_events(struct intel_pt * pt,struct perf_session * session) intel_pt_synth_events() argument 4162 intel_pt_setup_pebs_events(struct intel_pt * pt) intel_pt_setup_pebs_events() argument 4211 struct intel_pt *pt = data; intel_pt_perf_config() local 4223 intel_pt_tsc_start(u64 ns,struct intel_pt * pt) intel_pt_tsc_start() argument 4243 intel_pt_tsc_end(u64 ns,struct intel_pt * pt) intel_pt_tsc_end() argument 4262 intel_pt_setup_time_ranges(struct intel_pt * pt,struct itrace_synth_opts * opts) intel_pt_setup_time_ranges() argument 4301 intel_pt_parse_vm_tm_corr_arg(struct intel_pt * pt,char ** args) intel_pt_parse_vm_tm_corr_arg() argument 4341 intel_pt_parse_vm_tm_corr_args(struct intel_pt * pt) intel_pt_parse_vm_tm_corr_args() argument 4415 struct intel_pt *pt; intel_pt_process_auxtrace_info() local [all...] |
/linux/drivers/dma/amd/ptdma/ |
H A D | ptdma-pci.c | 38 struct pt_device *pt; in pt_alloc_struct() local 40 pt = devm_kzalloc(dev, sizeof(*pt), GFP_KERNEL); in pt_alloc_struct() 42 if (!pt) in pt_alloc_struct() 44 pt->dev = dev; in pt_alloc_struct() 46 INIT_LIST_HEAD(&pt->cmd); in pt_alloc_struct() 48 return pt; in pt_alloc_struct() 51 static int pt_get_msix_irqs(struct pt_device *pt) in pt_get_msix_irqs() argument 53 struct pt_msix *pt_msix = pt->pt_msix; in pt_get_msix_irqs() 54 struct device *dev = pt in pt_get_msix_irqs() 71 pt_get_msi_irq(struct pt_device * pt) pt_get_msi_irq() argument 86 pt_get_irqs(struct pt_device * pt) pt_get_irqs() argument 107 pt_free_irqs(struct pt_device * pt) pt_free_irqs() argument 123 struct pt_device *pt; pt_pci_probe() local 208 struct pt_device *pt = dev_get_drvdata(dev); pt_pci_remove() local [all...] |
H A D | ptdma-dev.c | 104 struct pt_device *pt = container_of(cmd_q, struct pt_device, cmd_q); in pt_core_perform_passthru() local 117 pt_core_enable_queue_interrupts(pt); in pt_core_perform_passthru() 119 pt_core_disable_queue_interrupts(pt); in pt_core_perform_passthru() 128 struct pt_cmd_queue *cmd_q = &cmd->pt->cmd_q; in pt_do_cmd_complete() 137 pt_log_error(cmd_q->pt, cmd_q->cmd_error); in pt_do_cmd_complete() 144 void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q) in pt_check_status_trans() argument 160 pt_do_cmd_complete((ulong)&pt->tdata); in pt_check_status_trans() 166 struct pt_device *pt = data; in pt_core_irq_handler() local 167 struct pt_cmd_queue *cmd_q = &pt->cmd_q; in pt_core_irq_handler() 169 pt_core_disable_queue_interrupts(pt); in pt_core_irq_handler() 176 pt_core_init(struct pt_device * pt) pt_core_init() argument 277 pt_core_destroy(struct pt_device * pt) pt_core_destroy() argument [all...] |
H A D | ptdma-dmaengine.c | 60 ae4_log_error(cmd_q->pt, cmd_q->cmd_error); in ae4_check_status_error() 92 struct pt_device *pt = desc->pt; in pt_do_cleanup() local 94 kmem_cache_free(pt->dma_desc_cache, desc); in pt_do_cleanup() 97 static struct pt_cmd_queue *pt_get_cmd_queue(struct pt_device *pt, struct pt_dma_chan *chan) in pt_get_cmd_queue() argument 103 if (pt->ver == AE4_DMA_VERSION) { in pt_get_cmd_queue() 104 ae4 = container_of(pt, struct ae4_device, pt); in pt_get_cmd_queue() 108 cmd_q = &pt->cmd_q; in pt_get_cmd_queue() 164 struct pt_device *pt; in pt_dma_start_desc() local 201 struct pt_device *pt; pt_handle_active_desc() local 268 struct pt_device *pt; pt_cmd_callback() local 388 struct pt_device *pt = chan->pt; pt_create_desc() local 452 struct pt_device *pt; pt_issue_pending() local 475 pt_check_status_trans_ae4(struct pt_device * pt,struct pt_cmd_queue * cmd_q) pt_check_status_trans_ae4() argument 489 struct pt_device *pt = chan->pt; pt_tx_status() local 505 struct pt_device *pt = chan->pt; pt_pause() local 521 struct pt_device *pt = chan->pt; pt_resume() local 541 struct pt_device *pt = chan->pt; pt_terminate_all() local 562 pt_dmaengine_register(struct pt_device * pt) pt_dmaengine_register() argument 652 pt_dmaengine_unregister(struct pt_device * pt) pt_dmaengine_unregister() argument [all...] |
H A D | ptdma-debugfs.c | 26 struct pt_device *pt = s->private; in pt_debugfs_info_show() local 30 seq_printf(s, "Device name: %s\n", dev_name(pt->dev)); in pt_debugfs_info_show() 32 if (pt->ver == AE4_DMA_VERSION) { in pt_debugfs_info_show() 33 ae4 = container_of(pt, struct ae4_device, pt); in pt_debugfs_info_show() 38 seq_printf(s, " # Cmds: %d\n", pt->cmd_count); in pt_debugfs_info_show() 41 regval = ioread32(pt->io_regs + CMD_PT_VERSION); in pt_debugfs_info_show() 57 struct pt_device *pt = s->private; in pt_debugfs_stats_show() local 59 seq_printf(s, "Total Interrupts Handled: %ld\n", pt->total_interrupts); in pt_debugfs_stats_show() 67 struct pt_device *pt; in pt_debugfs_queue_show() local 101 ptdma_debugfs_setup(struct pt_device * pt) ptdma_debugfs_setup() argument [all...] |
/linux/sound/i2c/other/ |
H A D | pt2258.c | 32 int snd_pt2258_reset(struct snd_pt2258 *pt) in snd_pt2258_reset() argument 39 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() 40 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) in snd_pt2258_reset() 42 snd_i2c_unlock(pt->i2c_bus); in snd_pt2258_reset() 45 pt->mute = 1; in snd_pt2258_reset() 47 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() 48 if (snd_i2c_sendbytes(pt->i2c_dev, bytes, 1) != 1) in snd_pt2258_reset() 50 snd_i2c_unlock(pt->i2c_bus); in snd_pt2258_reset() 54 pt->volume[i] = 0; in snd_pt2258_reset() 57 snd_i2c_lock(pt in snd_pt2258_reset() 83 struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol); pt2258_stereo_volume_get() local 95 struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol); pt2258_stereo_volume_put() local 136 struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol); pt2258_switch_get() local 145 struct snd_pt2258 *pt = snd_kcontrol_chip(kcontrol); pt2258_switch_put() local 170 snd_pt2258_build_controls(struct snd_pt2258 * pt) snd_pt2258_build_controls() argument [all...] |
/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_ppgtt.c | 18 struct i915_page_table *pt; in alloc_pt() local 20 pt = kmalloc(sizeof(*pt), I915_GFP_ALLOW_FAIL); in alloc_pt() 21 if (unlikely(!pt)) in alloc_pt() 24 pt->base = vm->alloc_pt_dma(vm, sz); in alloc_pt() 25 if (IS_ERR(pt->base)) { in alloc_pt() 26 kfree(pt); in alloc_pt() 30 pt->is_compact = false; in alloc_pt() 31 atomic_set(&pt->used, 0); in alloc_pt() 32 return pt; in alloc_pt() 71 free_px(struct i915_address_space * vm,struct i915_page_table * pt,int lvl) free_px() argument 127 release_pd_entry(struct i915_page_directory * const pd,const unsigned short idx,struct i915_page_table * const pt,const struct drm_i915_gem_object * const scratch) release_pd_entry() argument 244 struct i915_page_table *pt; i915_vm_alloc_pt_stash() local 279 struct i915_page_table *pt; i915_vm_map_pt_stash() local 296 struct i915_page_table *pt; i915_vm_free_pt_stash() local [all...] |
H A D | gen6_ppgtt.c | 18 /* Write pde (index) from the page directory @pd to the page table @pt */ 21 const struct i915_page_table *pt) in gen6_write_pde() argument 23 dma_addr_t addr = pt ? px_dma(pt) : px_dma(ppgtt->base.vm.scratch[1]); in gen6_write_pde() 85 struct i915_page_table * const pt = in gen6_ppgtt_clear_range() local 92 GEM_BUG_ON(count > atomic_read(&pt->used)); in gen6_ppgtt_clear_range() 93 if (!atomic_sub_return(count, &pt->used)) in gen6_ppgtt_clear_range() 103 vaddr = px_vaddr(pt); in gen6_ppgtt_clear_range() 153 struct i915_page_table *pt; in gen6_flush_pd() local 161 gen6_for_each_pde(pt, p in gen6_flush_pd() 178 struct i915_page_table *pt; gen6_alloc_va_range() local 259 struct i915_page_table *pt; gen6_ppgtt_free_pd() local 301 struct i915_page_table *pt; pd_vma_unbind() local [all...] |
/linux/tools/testing/selftests/bpf/progs/ |
H A D | kfunc_call_fail.c | 41 struct prog_test_ref_kfunc *pt; in kfunc_call_test_get_mem_fail_rdonly() local 46 pt = bpf_kfunc_call_test_acquire(&s); in kfunc_call_test_get_mem_fail_rdonly() 47 if (pt) { in kfunc_call_test_get_mem_fail_rdonly() 48 p = bpf_kfunc_call_test_get_rdonly_mem(pt, 2 * sizeof(int)); in kfunc_call_test_get_mem_fail_rdonly() 54 bpf_kfunc_call_test_release(pt); in kfunc_call_test_get_mem_fail_rdonly() 62 struct prog_test_ref_kfunc *pt; in kfunc_call_test_get_mem_fail_use_after_free() local 67 pt = bpf_kfunc_call_test_acquire(&s); in kfunc_call_test_get_mem_fail_use_after_free() 68 if (pt) { in kfunc_call_test_get_mem_fail_use_after_free() 69 p = bpf_kfunc_call_test_get_rdwr_mem(pt, 2 * sizeof(int)); in kfunc_call_test_get_mem_fail_use_after_free() 77 bpf_kfunc_call_test_release(pt); in kfunc_call_test_get_mem_fail_use_after_free() 88 struct prog_test_ref_kfunc *pt; kfunc_call_test_get_mem_fail_oob() local 111 struct prog_test_ref_kfunc *pt; kfunc_call_test_get_mem_fail_not_const() local 132 struct prog_test_ref_kfunc *pt; kfunc_call_test_mem_acquire_fail() local [all...] |
/linux/arch/sparc/lib/ |
H A D | Memcpy_utils.S | 20 ba,pt %xcc, __restore_asi 24 ba,pt %xcc, __restore_asi 28 ba,pt %xcc, __restore_asi 32 ba,pt %xcc, __restore_asi 36 ba,pt %xcc, __restore_asi 40 ba,pt %xcc, __restore_asi 44 ba,pt %xcc, __restore_asi 48 ba,pt %xcc, __restore_asi 52 ba,pt %xcc, __restore_asi 57 ba,pt [all...] |
/linux/drivers/gpu/drm/i915/selftests/ |
H A D | scatterlist.c | 41 static noinline int expect_pfn_sg(struct pfn_table *pt, in expect_pfn_sg() argument 50 pfn = pt->start; in expect_pfn_sg() 51 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { in expect_pfn_sg() 53 unsigned int npages = npages_fn(n, pt->st.nents, rnd); in expect_pfn_sg() 72 if (pfn != pt->end) { in expect_pfn_sg() 74 __func__, who, pt->end, pfn); in expect_pfn_sg() 81 static noinline int expect_pfn_sg_page_iter(struct pfn_table *pt, in expect_pfn_sg_page_iter() argument 88 pfn = pt->start; in expect_pfn_sg_page_iter() 89 for_each_sg_page(pt in expect_pfn_sg_page_iter() 112 expect_pfn_sgtiter(struct pfn_table * pt,const char * who,unsigned long timeout) expect_pfn_sgtiter() argument 142 expect_pfn_sgtable(struct pfn_table * pt,npages_fn_t npages_fn,struct rnd_state * rnd,const char * who,unsigned long timeout) expect_pfn_sgtable() argument 214 alloc_table(struct pfn_table * pt,unsigned long count,unsigned long max,npages_fn_t npages_fn,struct rnd_state * rnd,int alloc_error) alloc_table() argument 293 struct pfn_table pt; igt_sg_alloc() local 329 struct pfn_table pt; igt_sg_trim() local [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmnv44.c | 27 nv44_vmm_pgt_fill(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv44_vmm_pgt_fill() argument 33 tmp[0] = nvkm_ro32(pt->memory, pteo + 0x0); in nv44_vmm_pgt_fill() 34 tmp[1] = nvkm_ro32(pt->memory, pteo + 0x4); in nv44_vmm_pgt_fill() 35 tmp[2] = nvkm_ro32(pt->memory, pteo + 0x8); in nv44_vmm_pgt_fill() 36 tmp[3] = nvkm_ro32(pt->memory, pteo + 0xc); in nv44_vmm_pgt_fill() 66 VMM_WO032(pt, vmm, pteo + 0x0, tmp[0]); in nv44_vmm_pgt_fill() 67 VMM_WO032(pt, vmm, pteo + 0x4, tmp[1]); in nv44_vmm_pgt_fill() 68 VMM_WO032(pt, vmm, pteo + 0x8, tmp[2]); in nv44_vmm_pgt_fill() 69 VMM_WO032(pt, vmm, pteo + 0xc, tmp[3] | 0x40000000); in nv44_vmm_pgt_fill() 73 nv44_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv44_vmm_pgt_pte() argument 105 nv44_vmm_pgt_sgl(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) nv44_vmm_pgt_sgl() argument 112 nv44_vmm_pgt_dma(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) nv44_vmm_pgt_dma() argument 148 nv44_vmm_pgt_unmap(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes) nv44_vmm_pgt_unmap() argument [all...] |
H A D | vmmgp100.c | 35 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in gp100_vmm_pfn_unmap() argument 40 nvkm_kmap(pt->memory); in gp100_vmm_pfn_unmap() 42 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0); in gp100_vmm_pfn_unmap() 43 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4); in gp100_vmm_pfn_unmap() 51 nvkm_done(pt->memory); in gp100_vmm_pfn_unmap() 56 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes) in gp100_vmm_pfn_clear() argument 59 nvkm_kmap(pt->memory); in gp100_vmm_pfn_clear() 61 u32 datalo = nvkm_ro32(pt in gp100_vmm_pfn_clear() 75 gp100_vmm_pgt_pfn(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) gp100_vmm_pgt_pfn() argument 115 gp100_vmm_pgt_pte(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map,u64 addr) gp100_vmm_pgt_pte() argument 129 gp100_vmm_pgt_sgl(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) gp100_vmm_pgt_sgl() argument 136 gp100_vmm_pgt_dma(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) gp100_vmm_pgt_dma() argument 155 gp100_vmm_pgt_mem(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) gp100_vmm_pgt_mem() argument 163 gp100_vmm_pgt_sparse(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes) gp100_vmm_pgt_sparse() argument 183 gp100_vmm_lpt_invalid(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes) gp100_vmm_lpt_invalid() argument 198 gp100_vmm_pd0_pte(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map,u64 addr) gp100_vmm_pd0_pte() argument 212 gp100_vmm_pd0_mem(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) gp100_vmm_pd0_mem() argument 219 gp100_vmm_pde(struct nvkm_mmu_pt * pt,u64 * data) gp100_vmm_pde() argument 254 gp100_vmm_pd0_sparse(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 pdei,u32 pdes) gp100_vmm_pd0_sparse() argument 262 gp100_vmm_pd0_unmap(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 pdei,u32 pdes) gp100_vmm_pd0_unmap() argument 269 gp100_vmm_pd0_pfn_unmap(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes) gp100_vmm_pd0_pfn_unmap() argument 291 gp100_vmm_pd0_pfn_clear(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes) gp100_vmm_pd0_pfn_clear() argument 312 gp100_vmm_pd0_pfn(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) gp100_vmm_pd0_pfn() argument [all...] |
H A D | vmmgh100.c | 13 gh100_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes, in gh100_vmm_pgt_pte() argument 19 VMM_WO064(pt, vmm, ptei++ * NV_MMU_VER3_PTE__SIZE, data); in gh100_vmm_pgt_pte() 25 gh100_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes, in gh100_vmm_pgt_sgl() argument 28 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, gh100_vmm_pgt_pte); in gh100_vmm_pgt_sgl() 32 gh100_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes, in gh100_vmm_pgt_dma() argument 38 nvkm_kmap(pt->memory); in gh100_vmm_pgt_dma() 42 VMM_WO064(pt, vmm, ptei++ * NV_MMU_VER3_PTE__SIZE, data); in gh100_vmm_pgt_dma() 44 nvkm_done(pt->memory); in gh100_vmm_pgt_dma() 48 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, gh100_vmm_pgt_pte); in gh100_vmm_pgt_dma() 52 gh100_vmm_pgt_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, u3 argument 60 gh100_vmm_pgt_sparse(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes) gh100_vmm_pgt_sparse() argument 78 gh100_vmm_lpt_invalid(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes) gh100_vmm_lpt_invalid() argument 94 gh100_vmm_pd0_pte(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map,u64 addr) gh100_vmm_pd0_pte() argument 106 gh100_vmm_pd0_mem(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 ptei,u32 ptes,struct nvkm_vmm_map * map) gh100_vmm_pd0_mem() argument 113 gh100_vmm_pde(struct nvkm_mmu_pt * pt,u64 * data) gh100_vmm_pde() argument 156 gh100_vmm_pd0_sparse(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 pdei,u32 pdes) gh100_vmm_pd0_sparse() argument 165 gh100_vmm_pd0_unmap(struct nvkm_vmm * vmm,struct nvkm_mmu_pt * pt,u32 pdei,u32 pdes) gh100_vmm_pd0_unmap() argument [all...] |
H A D | vmmnv41.c | 27 nv41_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_pte() argument 32 VMM_WO032(pt, vmm, ptei++ * 4, data); in nv41_vmm_pgt_pte() 38 nv41_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_sgl() argument 41 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte); in nv41_vmm_pgt_sgl() 45 nv41_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt, in nv41_vmm_pgt_dma() argument 49 nvkm_kmap(pt->memory); in nv41_vmm_pgt_dma() 52 VMM_WO032(pt, vmm, ptei++ * 4, data); in nv41_vmm_pgt_dma() 54 nvkm_done(pt->memory); in nv41_vmm_pgt_dma() 56 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, nv41_vmm_pgt_pte); in nv41_vmm_pgt_dma() 62 struct nvkm_mmu_pt *pt, u3 in nv41_vmm_pgt_unmap() argument [all...] |
/linux/drivers/hv/ |
H A D | mshv_eventfd.c | 223 struct mshv_partition *pt = rp->rsmplr_partn; in mshv_irqfd_resampler_shutdown() local 225 mutex_lock(&pt->irqfds_resampler_lock); in mshv_irqfd_resampler_shutdown() 228 synchronize_srcu(&pt->pt_irq_srcu); in mshv_irqfd_resampler_shutdown() 232 mshv_unregister_irq_ack_notifier(pt, &rp->rsmplr_notifier); in mshv_irqfd_resampler_shutdown() 236 mutex_unlock(&pt->irqfds_resampler_lock); in mshv_irqfd_resampler_shutdown() 297 struct mshv_partition *pt = irqfd->irqfd_partn; in mshv_irqfd_wakeup() local 304 idx = srcu_read_lock(&pt->pt_irq_srcu); in mshv_irqfd_wakeup() 314 srcu_read_unlock(&pt->pt_irq_srcu, idx); in mshv_irqfd_wakeup() 323 spin_lock_irqsave(&pt->pt_irqfds_lock, flags); in mshv_irqfd_wakeup() 337 spin_unlock_irqrestore(&pt in mshv_irqfd_wakeup() 344 mshv_irqfd_update(struct mshv_partition * pt,struct mshv_irqfd * irqfd) mshv_irqfd_update() argument 354 mshv_irqfd_routing_update(struct mshv_partition * pt) mshv_irqfd_routing_update() argument 382 mshv_irqfd_assign(struct mshv_partition * pt,struct mshv_user_irqfd * args) mshv_irqfd_assign() argument 527 mshv_irqfd_deassign(struct mshv_partition * pt,struct mshv_user_irqfd * args) mshv_irqfd_deassign() argument 558 mshv_set_unset_irqfd(struct mshv_partition * pt,struct mshv_user_irqfd * args) mshv_set_unset_irqfd() argument 574 mshv_irqfd_release(struct mshv_partition * pt) mshv_irqfd_release() argument 640 ioeventfd_check_collision(struct mshv_partition * pt,struct mshv_ioeventfd * p) ioeventfd_check_collision() argument 656 mshv_assign_ioeventfd(struct mshv_partition * pt,struct mshv_user_ioeventfd * args) mshv_assign_ioeventfd() argument 748 mshv_deassign_ioeventfd(struct mshv_partition * pt,struct mshv_user_ioeventfd * args) mshv_deassign_ioeventfd() argument 789 mshv_set_unset_ioeventfd(struct mshv_partition * pt,struct mshv_user_ioeventfd * args) mshv_set_unset_ioeventfd() argument 807 mshv_eventfd_init(struct mshv_partition * pt) mshv_eventfd_init() argument 818 mshv_eventfd_release(struct mshv_partition * pt) mshv_eventfd_release() argument [all...] |
/linux/drivers/dma-buf/ |
H A D | sw_sync.c | 154 struct sync_pt *pt = dma_fence_to_sync_pt(fence); in timeline_fence_release() local 159 if (!list_empty(&pt->link)) { in timeline_fence_release() 160 list_del(&pt->link); in timeline_fence_release() 161 rb_erase(&pt->node, &parent->pt_tree); in timeline_fence_release() 178 struct sync_pt *pt = dma_fence_to_sync_pt(fence); in timeline_fence_set_deadline() local 183 if (ktime_before(deadline, pt->deadline)) in timeline_fence_set_deadline() 184 pt->deadline = deadline; in timeline_fence_set_deadline() 186 pt->deadline = deadline; in timeline_fence_set_deadline() 211 struct sync_pt *pt, *next; in sync_timeline_signal() local 219 list_for_each_entry_safe(pt, nex in sync_timeline_signal() 252 struct sync_pt *pt; sync_pt_create() local 328 struct sync_pt *pt, *next; sw_sync_debugfs_release() local 348 struct sync_pt *pt; sw_sync_ioctl_create_fence() local 411 struct sync_pt *pt; sw_sync_ioctl_get_deadline() local [all...] |
/linux/drivers/gpu/drm/gma500/ |
H A D | mmu.c | 224 static void psb_mmu_free_pt(struct psb_mmu_pt *pt) in psb_mmu_free_pt() argument 226 __free_page(pt->p); in psb_mmu_free_pt() 227 kfree(pt); in psb_mmu_free_pt() 235 struct psb_mmu_pt *pt; in psb_mmu_free_pagedir() local 248 pt = pd->tables[i]; in psb_mmu_free_pagedir() 249 if (pt) in psb_mmu_free_pagedir() 250 psb_mmu_free_pt(pt); in psb_mmu_free_pagedir() 263 struct psb_mmu_pt *pt = kmalloc(sizeof(*pt), GFP_KERNEL); in psb_mmu_alloc_pt() local 272 if (!pt) in psb_mmu_alloc_pt() 311 struct psb_mmu_pt *pt; psb_mmu_pt_alloc_map_lock() local 351 struct psb_mmu_pt *pt; psb_mmu_pt_map_lock() local 364 psb_mmu_pt_unmap_unlock(struct psb_mmu_pt * pt) psb_mmu_pt_unmap_unlock() argument 387 psb_mmu_set_pte(struct psb_mmu_pt * pt,unsigned long addr,uint32_t pte) psb_mmu_set_pte() argument 393 psb_mmu_invalidate_pte(struct psb_mmu_pt * pt,unsigned long addr) psb_mmu_invalidate_pte() argument 482 struct psb_mmu_pt *pt; psb_mmu_flush_ptes() local 529 struct psb_mmu_pt *pt; psb_mmu_remove_pfn_sequence() local 569 struct psb_mmu_pt *pt; psb_mmu_remove_pages() local 625 struct psb_mmu_pt *pt; psb_mmu_insert_pfn_sequence() local 672 struct psb_mmu_pt *pt; psb_mmu_insert_pages() local [all...] |
/linux/lib/ |
H A D | nlattr.c | 119 void nla_get_range_unsigned(const struct nla_policy *pt, in nla_get_range_unsigned() argument 122 WARN_ON_ONCE(pt->validation_type != NLA_VALIDATE_RANGE_PTR && in nla_get_range_unsigned() 123 (pt->min < 0 || pt->max < 0)); in nla_get_range_unsigned() 127 switch (pt->type) { in nla_get_range_unsigned() 150 switch (pt->validation_type) { in nla_get_range_unsigned() 153 range->min = pt->min; in nla_get_range_unsigned() 154 range->max = pt->max; in nla_get_range_unsigned() 157 *range = *pt->range; in nla_get_range_unsigned() 160 range->min = pt in nla_get_range_unsigned() 170 nla_validate_range_unsigned(const struct nla_policy * pt,const struct nlattr * nla,struct netlink_ext_ack * extack,unsigned int validate) nla_validate_range_unsigned() argument 242 nla_get_range_signed(const struct nla_policy * pt,struct netlink_range_validation_signed * range) nla_get_range_signed() argument 287 nla_validate_int_range_signed(const struct nla_policy * pt,const struct nlattr * nla,struct netlink_ext_ack * extack) nla_validate_int_range_signed() argument 325 nla_validate_int_range(const struct nla_policy * pt,const struct nlattr * nla,struct netlink_ext_ack * extack,unsigned int validate) nla_validate_int_range() argument 353 nla_validate_mask(const struct nla_policy * pt,const struct nlattr * nla,struct netlink_ext_ack * extack) nla_validate_mask() argument 398 const struct nla_policy *pt; validate_nla() local [all...] |
/linux/arch/alpha/kernel/ |
H A D | process.c | 286 dump_elf_thread(elf_greg_t *dest, struct pt_regs *pt, struct thread_info *ti) in dump_elf_thread() argument 289 struct switch_stack * sw = ((struct switch_stack *) pt) - 1; in dump_elf_thread() 291 dest[ 0] = pt->r0; in dump_elf_thread() 292 dest[ 1] = pt->r1; in dump_elf_thread() 293 dest[ 2] = pt->r2; in dump_elf_thread() 294 dest[ 3] = pt->r3; in dump_elf_thread() 295 dest[ 4] = pt->r4; in dump_elf_thread() 296 dest[ 5] = pt->r5; in dump_elf_thread() 297 dest[ 6] = pt->r6; in dump_elf_thread() 298 dest[ 7] = pt in dump_elf_thread() [all...] |
/linux/arch/x86/events/intel/ |
H A D | pt.c | 30 #include "pt.h" 32 static DEFINE_PER_CPU(struct pt, pt_ctx); 39 * attribute group of pt pmu device 108 PMU_FORMAT_ATTR(pt, "config:0" ); 420 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_start() local 427 if (READ_ONCE(pt->vmx_on)) in pt_config_start() 428 perf_aux_output_flag(&pt->handle, PERF_AUX_FLAG_PARTIAL); in pt_config_start() 466 struct pt *pt in pt_config_filters() local 506 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_config() local 564 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_config_stop() local 647 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_config_buffer() local 870 pt_update_head(struct pt * pt) pt_update_head() argument 924 pt_handle_status(struct pt * pt) pt_handle_status() argument 985 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_read_offset() local 1506 struct pt *pt = this_cpu_ptr(&pt_ctx); intel_pt_interrupt() local 1561 struct pt *pt = this_cpu_ptr(&pt_ctx); intel_pt_handle_vmx() local 1602 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_event_start() local 1650 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_event_stop() local 1708 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_event_snapshot_aux() local 1764 struct pt *pt = this_cpu_ptr(&pt_ctx); pt_event_add() local 1819 struct pt *pt = this_cpu_ptr(&pt_ctx); cpu_emergency_stop_pt() local [all...] |
/linux/net/core/ |
H A D | net-procfs.c | 175 struct packet_type *pt = NULL; in ptype_get_idx() local 182 list_for_each_entry_rcu(pt, ptype_list, list) { in ptype_get_idx() 184 return pt; in ptype_get_idx() 189 list_for_each_entry_rcu(pt, &seq_file_net(seq)->ptype_all, list) { in ptype_get_idx() 191 return pt; in ptype_get_idx() 195 list_for_each_entry_rcu(pt, &seq_file_net(seq)->ptype_specific, list) { in ptype_get_idx() 197 return pt; in ptype_get_idx() 202 list_for_each_entry_rcu(pt, &ptype_base[t], list) { in ptype_get_idx() 204 return pt; in ptype_get_idx() 222 struct packet_type *pt; in ptype_seq_next() local 281 struct packet_type *pt = v; ptype_seq_show() local [all...] |
/linux/arch/x86/kernel/ |
H A D | vm86_32.c | 79 #define AL(regs) (((unsigned char *)&((regs)->pt.ax))[0]) 80 #define AH(regs) (((unsigned char *)&((regs)->pt.ax))[1]) 81 #define IP(regs) (*(unsigned short *)&((regs)->pt.ip)) 82 #define SP(regs) (*(unsigned short *)&((regs)->pt.sp)) 111 set_flags(regs->pt.flags, VEFLAGS, X86_EFLAGS_VIF | vm86->veflags_mask); in save_v86_state() 119 unsafe_put_user(regs->pt.bx, &user->regs.ebx, Efault_end); in save_v86_state() 120 unsafe_put_user(regs->pt.cx, &user->regs.ecx, Efault_end); in save_v86_state() 121 unsafe_put_user(regs->pt.dx, &user->regs.edx, Efault_end); in save_v86_state() 122 unsafe_put_user(regs->pt.si, &user->regs.esi, Efault_end); in save_v86_state() 123 unsafe_put_user(regs->pt in save_v86_state() [all...] |
/linux/drivers/dma/amd/ae4dma/ |
H A D | ae4dma-dev.c | 53 struct pt_device *pt; in ae4_core_irq_handler() local 57 pt = cmd_q->pt; in ae4_core_irq_handler() 59 pt->total_interrupts++; in ae4_core_irq_handler() 91 struct pt_device *pt = &ae4->pt; in ae4_core_init() local 93 struct device *dev = pt->dev; in ae4_core_init() 97 writel(max_hw_q, pt->io_regs); in ae4_core_init() 105 cmd_q->pt = pt; in ae4_core_init() [all...] |
/linux/drivers/media/common/saa7146/ |
H A D | saa7146_core.c | 166 void *saa7146_vmalloc_build_pgtable(struct pci_dev *pci, long length, struct saa7146_pgtable *pt) in saa7146_vmalloc_build_pgtable() argument 175 if (!(pt->slist = vmalloc_to_sg(mem, pages))) in saa7146_vmalloc_build_pgtable() 178 if (saa7146_pgtable_alloc(pci, pt)) in saa7146_vmalloc_build_pgtable() 181 pt->nents = pages; in saa7146_vmalloc_build_pgtable() 182 slen = dma_map_sg(&pci->dev, pt->slist, pt->nents, DMA_FROM_DEVICE); in saa7146_vmalloc_build_pgtable() 186 if (0 != saa7146_pgtable_build_single(pci, pt, pt->slist, slen)) in saa7146_vmalloc_build_pgtable() 192 dma_unmap_sg(&pci->dev, pt->slist, pt in saa7146_vmalloc_build_pgtable() 204 saa7146_vfree_destroy_pgtable(struct pci_dev * pci,void * mem,struct saa7146_pgtable * pt) saa7146_vfree_destroy_pgtable() argument 213 saa7146_pgtable_free(struct pci_dev * pci,struct saa7146_pgtable * pt) saa7146_pgtable_free() argument 221 saa7146_pgtable_alloc(struct pci_dev * pci,struct saa7146_pgtable * pt) saa7146_pgtable_alloc() argument 237 saa7146_pgtable_build_single(struct pci_dev * pci,struct saa7146_pgtable * pt,struct scatterlist * list,int sglen) saa7146_pgtable_build_single() argument [all...] |