Home
last modified time | relevance | path

Searched refs:event (Results 1 – 25 of 3322) sorted by relevance

12345678910>>...133

/linux/tools/testing/selftests/powerpc/pmu/event_code_tests/
H A Dgeneric_events_valid_test.c19 struct event event; in generic_events_valid_test() local
36 event_init_opts(&event, PERF_COUNT_HW_CPU_CYCLES, PERF_TYPE_HARDWARE, "event"); in generic_events_valid_test()
37 FAIL_IF(event_open(&event)); in generic_events_valid_test()
38 event_close(&event); in generic_events_valid_test()
40 event_init_opts(&event, PERF_COUNT_HW_INSTRUCTIONS, in generic_events_valid_test()
42 FAIL_IF(event_open(&event)); in generic_events_valid_test()
43 event_close(&event); in generic_events_valid_test()
45 event_init_opts(&event, PERF_COUNT_HW_CACHE_REFERENCES, in generic_events_valid_test()
47 FAIL_IF(event_open(&event)); in generic_events_valid_test()
48 event_close(&event); in generic_events_valid_test()
[all …]
H A Dreserved_bits_mmcra_sample_elig_mode_test.c23 struct event event; in reserved_bits_mmcra_sample_elig_mode() local
37 event_init(&event, 0x50401e0); in reserved_bits_mmcra_sample_elig_mode()
38 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode()
40 event_init(&event, 0x90401e0); in reserved_bits_mmcra_sample_elig_mode()
41 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode()
43 event_init(&event, 0xD0401e0); in reserved_bits_mmcra_sample_elig_mode()
44 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode()
46 event_init(&event, 0x190401e0); in reserved_bits_mmcra_sample_elig_mode()
47 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode()
49 event_init(&event, 0x1D0401e0); in reserved_bits_mmcra_sample_elig_mode()
[all …]
/linux/tools/testing/selftests/powerpc/pmu/ebb/
H A Devent_attributes_test.c17 struct event event, leader; in event_attributes() local
21 event_init(&event, 0x1001e); in event_attributes()
22 event_leader_ebb_init(&event); in event_attributes()
24 FAIL_IF(event_open(&event)); in event_attributes()
25 event_close(&event); in event_attributes()
28 event_init(&event, 0x001e); /* CYCLES - no PMC specified */ in event_attributes()
29 event_leader_ebb_init(&event); in event_attributes()
31 FAIL_IF(event_open(&event) == 0); in event_attributes()
34 event_init(&event, 0x2001e); in event_attributes()
35 event_leader_ebb_init(&event); in event_attributes()
[all …]
H A Dinstruction_count_test.c25 static int do_count_loop(struct event *event, uint64_t instructions, in do_count_loop() argument
45 event->result.value = ebb_state.stats.pmc_count[4-1]; in do_count_loop()
47 difference = event->result.value - expected; in do_count_loop()
48 percentage = (double)difference / event->result.value * 100; in do_count_loop()
53 printf("Actual %llu\n", event->result.value); in do_count_loop()
63 if (difference / event->result.value) in do_count_loop()
70 static uint64_t determine_overhead(struct event *event) in determine_overhead() argument
75 do_count_loop(event, 0, 0, false); in determine_overhead()
76 overhead = event->result.value; in determine_overhead()
79 do_count_loop(event, 0, 0, false); in determine_overhead()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/core/
H A Devent.c26 nvkm_event_put(struct nvkm_event *event, u32 types, int index) in nvkm_event_put() argument
28 assert_spin_locked(&event->refs_lock); in nvkm_event_put()
30 nvkm_trace(event->subdev, "event: decr %08x on %d\n", types, index); in nvkm_event_put()
34 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put()
35 nvkm_trace(event->subdev, "event: blocking %d on %d\n", type, index); in nvkm_event_put()
36 if (event->func->fini) in nvkm_event_put()
37 event->func->fini(event, 1 << type, index); in nvkm_event_put()
43 nvkm_event_get(struct nvkm_event *event, u32 types, int index) in nvkm_event_get() argument
45 assert_spin_locked(&event->refs_lock); in nvkm_event_get()
47 nvkm_trace(event->subdev, "event: incr %08x on %d\n", types, index); in nvkm_event_get()
[all …]
/linux/kernel/events/
H A Dcore.c225 static bool is_kernel_event(struct perf_event *event) in is_kernel_event() argument
227 return READ_ONCE(event->owner) == TASK_TOMBSTONE; in is_kernel_event()
261 struct perf_event *event; member
269 struct perf_event *event = efs->event; in event_function() local
270 struct perf_event_context *ctx = event->ctx; in event_function()
305 efs->func(event, cpuctx, ctx, efs->data); in event_function()
312 static void event_function_call(struct perf_event *event, event_f func, void *data) in event_function_call() argument
314 struct perf_event_context *ctx = event->ctx; in event_function_call()
318 .event = event, in event_function_call()
323 if (!event->parent) { in event_function_call()
[all …]
/linux/net/sctp/
H A Dulpevent.c31 static void sctp_ulpevent_receive_data(struct sctp_ulpevent *event,
33 static void sctp_ulpevent_release_data(struct sctp_ulpevent *event);
34 static void sctp_ulpevent_release_frag_data(struct sctp_ulpevent *event);
38 static void sctp_ulpevent_init(struct sctp_ulpevent *event, in sctp_ulpevent_init() argument
42 memset(event, 0, sizeof(struct sctp_ulpevent)); in sctp_ulpevent_init()
43 event->msg_flags = msg_flags; in sctp_ulpevent_init()
44 event->rmem_len = len; in sctp_ulpevent_init()
51 struct sctp_ulpevent *event; in sctp_ulpevent_new() local
58 event = sctp_skb2event(skb); in sctp_ulpevent_new()
59 sctp_ulpevent_init(event, msg_flags, skb->truesize); in sctp_ulpevent_new()
[all …]
/linux/arch/powerpc/perf/
H A Dcore-fsl-emb.c23 struct perf_event *event[MAX_HWEVENTS]; member
162 static void fsl_emb_pmu_read(struct perf_event *event) in fsl_emb_pmu_read() argument
166 if (event->hw.state & PERF_HES_STOPPED) in fsl_emb_pmu_read()
175 prev = local64_read(&event->hw.prev_count); in fsl_emb_pmu_read()
177 val = read_pmc(event->hw.idx); in fsl_emb_pmu_read()
178 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in fsl_emb_pmu_read()
182 local64_add(delta, &event->count); in fsl_emb_pmu_read()
183 local64_sub(delta, &event->hw.period_left); in fsl_emb_pmu_read()
255 struct perf_event *event; in collect_events() local
263 for_each_sibling_event(event, group) { in collect_events()
[all …]
H A Dcore-book3s.c39 struct perf_event *event[MAX_HWEVENTS]; member
109 static inline void perf_get_data_addr(struct perf_event *event, struct pt_regs *regs, u64 *addrp) {… in perf_get_data_addr() argument
124 static bool is_ebb_event(struct perf_event *event) { return false; } in is_ebb_event() argument
125 static int ebb_event_check(struct perf_event *event) { return 0; } in ebb_event_check() argument
126 static void ebb_event_add(struct perf_event *event) { } in ebb_event_add() argument
133 static inline void power_pmu_bhrb_enable(struct perf_event *event) {} in power_pmu_bhrb_enable() argument
134 static inline void power_pmu_bhrb_disable(struct perf_event *event) {} in power_pmu_bhrb_disable() argument
139 static inline void power_pmu_bhrb_read(struct perf_event *event, struct cpu_hw_events *cpuhw) {} in power_pmu_bhrb_read() argument
205 static inline void perf_get_data_addr(struct perf_event *event, struct pt_regs *regs, u64 *addrp) in perf_get_data_addr() argument
230 if (is_kernel_addr(mfspr(SPRN_SDAR)) && event->attr.exclude_kernel) in perf_get_data_addr()
[all …]
/linux/Documentation/admin-guide/perf/
H A Dmrvl-odyssey-ddr-pmu.rst11 Additionally DSS also supports two fixed performance event counters, one
24 mrvl_ddr_pmu_<>/ddr_act_bypass_access/ [Kernel PMU event]
25 mrvl_ddr_pmu_<>/ddr_bsm_alloc/ [Kernel PMU event]
26 mrvl_ddr_pmu_<>/ddr_bsm_starvation/ [Kernel PMU event]
27 mrvl_ddr_pmu_<>/ddr_cam_active_access/ [Kernel PMU event]
28 mrvl_ddr_pmu_<>/ddr_cam_mwr/ [Kernel PMU event]
29 mrvl_ddr_pmu_<>/ddr_cam_rd_active_access/ [Kernel PMU event]
30 mrvl_ddr_pmu_<>/ddr_cam_rd_or_wr_access/ [Kernel PMU event]
31 mrvl_ddr_pmu_<>/ddr_cam_read/ [Kernel PMU event]
32 mrvl_ddr_pmu_<>/ddr_cam_wr_access/ [Kernel PMU event]
[all …]
/linux/tools/perf/util/
H A Devent.c155 struct perf_record_stat_config *event) in perf_event__read_stat_config() argument
159 for (i = 0; i < event->nr; i++) { in perf_event__read_stat_config()
161 switch (event->data[i].tag) { in perf_event__read_stat_config()
164 config->__val = event->data[i].val; \ in perf_event__read_stat_config()
174 event->data[i].tag); in perf_event__read_stat_config()
179 size_t perf_event__fprintf_comm(union perf_event *event, FILE *fp) in perf_event__fprintf_comm() argument
183 if (event->header.misc & PERF_RECORD_MISC_COMM_EXEC) in perf_event__fprintf_comm()
188 return fprintf(fp, "%s: %s:%d/%d\n", s, event->comm.comm, event->comm.pid, event->comm.tid); in perf_event__fprintf_comm()
191 size_t perf_event__fprintf_namespaces(union perf_event *event, FILE *fp) in perf_event__fprintf_namespaces() argument
197 ns_link_info = event->namespaces.link_info; in perf_event__fprintf_namespaces()
[all …]
H A Dsession.c45 union perf_event *event,
129 struct ordered_event *event) in ordered_events__deliver_event() argument
134 return perf_session__deliver_event(session, event->event, in ordered_events__deliver_event()
135 session->tool, event->file_offset, in ordered_events__deliver_event()
136 event->file_path); in ordered_events__deliver_event()
268 static void swap_sample_id_all(union perf_event *event, void *data) in swap_sample_id_all() argument
270 void *end = (void *) event + event->header.size; in swap_sample_id_all()
277 static void perf_event__all64_swap(union perf_event *event, in perf_event__all64_swap() argument
280 struct perf_event_header *hdr = &event->header; in perf_event__all64_swap()
281 mem_bswap_64(hdr + 1, event->header.size - sizeof(*hdr)); in perf_event__all64_swap()
[all …]
/linux/drivers/perf/
H A Driscv_pmu.c21 static bool riscv_perf_user_access(struct perf_event *event) in riscv_perf_user_access() argument
23 return ((event->attr.type == PERF_TYPE_HARDWARE) || in riscv_perf_user_access()
24 (event->attr.type == PERF_TYPE_HW_CACHE) || in riscv_perf_user_access()
25 (event->attr.type == PERF_TYPE_RAW)) && in riscv_perf_user_access()
26 !!(event->hw.flags & PERF_EVENT_FLAG_USER_READ_CNT) && in riscv_perf_user_access()
27 (event->hw.idx != -1); in riscv_perf_user_access()
30 void arch_perf_update_userpage(struct perf_event *event, in arch_perf_update_userpage() argument
40 userpg->cap_user_rdpmc = riscv_perf_user_access(event); in arch_perf_update_userpage()
48 userpg->pmc_width = to_riscv_pmu(event->pmu)->ctr_get_width(event->hw.idx) + 1; in arch_perf_update_userpage()
145 u64 riscv_pmu_ctr_get_width_mask(struct perf_event *event) in riscv_pmu_ctr_get_width_mask() argument
[all …]
H A Dqcom_l3_pmu.c136 static inline u32 get_event_type(struct perf_event *event) in get_event_type() argument
138 return (event->attr.config) & L3_EVTYPE_MASK; in get_event_type()
141 static inline bool event_uses_long_counter(struct perf_event *event) in event_uses_long_counter() argument
143 return !!(event->attr.config & BIT_ULL(L3_EVENT_LC_BIT)); in event_uses_long_counter()
146 static inline int event_num_counters(struct perf_event *event) in event_num_counters() argument
148 return event_uses_long_counter(event) ? 2 : 1; in event_num_counters()
175 void (*start)(struct perf_event *event);
177 void (*stop)(struct perf_event *event, int flags);
179 void (*update)(struct perf_event *event);
193 static void qcom_l3_cache__64bit_counter_start(struct perf_event *event) in qcom_l3_cache__64bit_counter_start() argument
[all …]
H A Darm_pmu.c108 static inline u64 arm_pmu_event_max_period(struct perf_event *event) in arm_pmu_event_max_period() argument
110 if (event->hw.flags & ARMPMU_EVT_64BIT) in arm_pmu_event_max_period()
112 else if (event->hw.flags & ARMPMU_EVT_63BIT) in arm_pmu_event_max_period()
114 else if (event->hw.flags & ARMPMU_EVT_47BIT) in arm_pmu_event_max_period()
174 armpmu_map_event(struct perf_event *event, in armpmu_map_event() argument
182 u64 config = event->attr.config; in armpmu_map_event()
183 int type = event->attr.type; in armpmu_map_event()
185 if (type == event->pmu->type) in armpmu_map_event()
200 int armpmu_event_set_period(struct perf_event *event) in armpmu_event_set_period() argument
202 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
[all …]
/linux/drivers/md/
H A Ddm-uevent.c39 static void dm_uevent_free(struct dm_uevent *event) in dm_uevent_free() argument
41 kmem_cache_free(_dm_event_cache, event); in dm_uevent_free()
46 struct dm_uevent *event; in dm_uevent_alloc() local
48 event = kmem_cache_zalloc(_dm_event_cache, GFP_ATOMIC); in dm_uevent_alloc()
49 if (!event) in dm_uevent_alloc()
52 INIT_LIST_HEAD(&event->elist); in dm_uevent_alloc()
53 event->md = md; in dm_uevent_alloc()
55 return event; in dm_uevent_alloc()
65 struct dm_uevent *event; in dm_build_path_uevent() local
67 event = dm_uevent_alloc(md); in dm_build_path_uevent()
[all …]
/linux/drivers/firmware/
H A Darm_sdei.c77 struct sdei_event *event; member
82 #define CROSSCALL_INIT(arg, event) \ argument
84 arg.event = event; \
90 struct sdei_event *event) in sdei_do_local_call() argument
94 CROSSCALL_INIT(arg, event); in sdei_do_local_call()
101 struct sdei_event *event) in sdei_do_cross_call() argument
105 CROSSCALL_INIT(arg, event); in sdei_do_cross_call()
191 static int sdei_api_event_get_info(u32 event, u32 info, u64 *result) in sdei_api_event_get_info() argument
193 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_GET_INFO, event, info, 0, in sdei_api_event_get_info()
203 struct sdei_event *event; in sdei_event_create() local
[all …]
/linux/arch/s390/kernel/
H A Dperf_pai_crypto.c38 struct perf_event *event; /* Perf event for sampling */ member
95 static void paicrypt_event_destroy_cpu(struct perf_event *event, int cpu) in paicrypt_event_destroy_cpu() argument
102 "refcnt %u\n", __func__, event->attr.config, in paicrypt_event_destroy_cpu()
103 event->cpu, cpump->active_events, in paicrypt_event_destroy_cpu()
111 static void paicrypt_event_destroy(struct perf_event *event) in paicrypt_event_destroy() argument
116 free_page(PAI_SAVE_AREA(event)); in paicrypt_event_destroy()
117 if (event->cpu == -1) { in paicrypt_event_destroy()
118 struct cpumask *mask = PAI_CPU_MASK(event); in paicrypt_event_destroy()
121 paicrypt_event_destroy_cpu(event, cpu); in paicrypt_event_destroy()
124 paicrypt_event_destroy_cpu(event, event->cpu); in paicrypt_event_destroy()
[all …]
H A Dperf_pai_ext.c51 struct perf_event *event; /* Perf event for sampling */ member
119 static void paiext_event_destroy_cpu(struct perf_event *event, int cpu) in paiext_event_destroy_cpu() argument
131 static void paiext_event_destroy(struct perf_event *event) in paiext_event_destroy() argument
135 free_page(PAI_SAVE_AREA(event)); in paiext_event_destroy()
136 if (event->cpu == -1) { in paiext_event_destroy()
137 struct cpumask *mask = PAI_CPU_MASK(event); in paiext_event_destroy()
140 paiext_event_destroy_cpu(event, cpu); in paiext_event_destroy()
143 paiext_event_destroy_cpu(event, event->cpu); in paiext_event_destroy()
146 event->cpu); in paiext_event_destroy()
163 static int paiext_alloc_cpu(struct perf_event *event, int cpu) in paiext_alloc_cpu() argument
[all …]
/linux/Documentation/ABI/testing/
H A Dsysfs-bus-event_source-devices-dfl_fme13 event = "config:0-11" - event ID
14 evtype = "config:12-15" - event type
15 portid = "config:16-23" - event source
19 fab_mmio_read = "event=0x06,evtype=0x02,portid=0xff"
21 It shows this fab_mmio_read is a fabric type (0x02) event with
22 0x06 local event id for overall monitoring (portid=0xff).
37 a single performance monitoring event supported by this fme pmu.
38 The name of the file is the name of the event.
45 clock = "event=0x00,evtype=0x00,portid=0xff"
49 cache_read_hit = "event=0x00,evtype=0x01,portid=0xff"
[all …]
/linux/arch/powerpc/kernel/
H A Deeh_event.c42 struct eeh_event *event; in eeh_event_handler() local
50 event = NULL; in eeh_event_handler()
52 event = list_entry(eeh_eventlist.next, in eeh_event_handler()
54 list_del(&event->list); in eeh_event_handler()
57 if (!event) in eeh_event_handler()
61 if (event->pe) in eeh_event_handler()
62 eeh_handle_normal_event(event->pe); in eeh_event_handler()
66 kfree(event); in eeh_event_handler()
105 struct eeh_event *event; in __eeh_send_failure_event() local
107 event = kzalloc(sizeof(*event), GFP_ATOMIC); in __eeh_send_failure_event()
[all …]
/linux/scripts/dtc/
H A Dyamltree.c24 #define yaml_emitter_emit_or_die(emitter, event) ( \ argument
26 if (!yaml_emitter_emit(emitter, event)) \
35 yaml_event_t event; in yaml_propval_int() local
49 yaml_sequence_start_event_initialize(&event, NULL, in yaml_propval_int()
51 yaml_emitter_emit_or_die(emitter, &event); in yaml_propval_int()
82 yaml_scalar_event_initialize(&event, NULL, in yaml_propval_int()
86 yaml_scalar_event_initialize(&event, NULL, in yaml_propval_int()
89 yaml_emitter_emit_or_die(emitter, &event); in yaml_propval_int()
92 yaml_sequence_end_event_initialize(&event); in yaml_propval_int()
93 yaml_emitter_emit_or_die(emitter, &event); in yaml_propval_int()
[all …]
/linux/arch/sh/kernel/
H A Dperf_event.c63 static void hw_perf_event_destroy(struct perf_event *event) in hw_perf_event_destroy() argument
100 static int __hw_perf_event_init(struct perf_event *event) in __hw_perf_event_init() argument
102 struct perf_event_attr *attr = &event->attr; in __hw_perf_event_init()
103 struct hw_perf_event *hwc = &event->hw; in __hw_perf_event_init()
131 event->destroy = hw_perf_event_destroy; in __hw_perf_event_init()
158 static void sh_perf_event_update(struct perf_event *event, in sh_perf_event_update() argument
196 local64_add(delta, &event->count); in sh_perf_event_update()
199 static void sh_pmu_stop(struct perf_event *event, int flags) in sh_pmu_stop() argument
202 struct hw_perf_event *hwc = &event->hw; in sh_pmu_stop()
205 if (!(event->hw.state & PERF_HES_STOPPED)) { in sh_pmu_stop()
[all …]
/linux/fs/notify/fanotify/
H A Dfanotify.h267 static inline void fanotify_init_event(struct fanotify_event *event, in fanotify_init_event() argument
270 fsnotify_init_event(&event->fse); in fanotify_init_event()
271 INIT_HLIST_NODE(&event->merge_list); in fanotify_init_event()
272 event->hash = hash; in fanotify_init_event()
273 event->mask = mask; in fanotify_init_event()
274 event->pid = NULL; in fanotify_init_event()
292 FANOTIFY_FE(struct fanotify_event *event) in FANOTIFY_FE() argument
294 return container_of(event, struct fanotify_fid_event, fae); in FANOTIFY_FE()
304 FANOTIFY_NE(struct fanotify_event *event) in FANOTIFY_NE() argument
306 return container_of(event, struct fanotify_name_event, fae); in FANOTIFY_NE()
[all …]
/linux/drivers/perf/hisilicon/
H A Dhisi_uncore_pmu.c97 static bool hisi_validate_event_group(struct perf_event *event) in hisi_validate_event_group() argument
99 struct perf_event *sibling, *leader = event->group_leader; in hisi_validate_event_group()
100 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_validate_event_group()
109 if (leader->pmu != event->pmu) in hisi_validate_event_group()
113 if (leader != event) in hisi_validate_event_group()
117 for_each_sibling_event(sibling, event->group_leader) { in hisi_validate_event_group()
120 if (sibling->pmu != event->pmu) in hisi_validate_event_group()
130 int hisi_uncore_pmu_get_event_idx(struct perf_event *event) in hisi_uncore_pmu_get_event_idx() argument
132 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_uncore_pmu_get_event_idx()
155 struct perf_event *event; in hisi_uncore_pmu_isr() local
[all …]

12345678910>>...133