| /linux/tools/testing/selftests/powerpc/pmu/event_code_tests/ |
| H A D | generic_events_valid_test.c | 19 struct event event; in generic_events_valid_test() local 36 event_init_opts(&event, PERF_COUNT_HW_CPU_CYCLES, PERF_TYPE_HARDWARE, "event"); in generic_events_valid_test() 37 FAIL_IF(event_open(&event)); in generic_events_valid_test() 38 event_close(&event); in generic_events_valid_test() 40 event_init_opts(&event, PERF_COUNT_HW_INSTRUCTIONS, in generic_events_valid_test() 42 FAIL_IF(event_open(&event)); in generic_events_valid_test() 43 event_close(&event); in generic_events_valid_test() 45 event_init_opts(&event, PERF_COUNT_HW_CACHE_REFERENCES, in generic_events_valid_test() 47 FAIL_IF(event_open(&event)); in generic_events_valid_test() 48 event_close(&event); in generic_events_valid_test() [all …]
|
| H A D | reserved_bits_mmcra_sample_elig_mode_test.c | 23 struct event event; in reserved_bits_mmcra_sample_elig_mode() local 37 event_init(&event, 0x50401e0); in reserved_bits_mmcra_sample_elig_mode() 38 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode() 40 event_init(&event, 0x90401e0); in reserved_bits_mmcra_sample_elig_mode() 41 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode() 43 event_init(&event, 0xD0401e0); in reserved_bits_mmcra_sample_elig_mode() 44 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode() 46 event_init(&event, 0x190401e0); in reserved_bits_mmcra_sample_elig_mode() 47 FAIL_IF(!event_open(&event)); in reserved_bits_mmcra_sample_elig_mode() 49 event_init(&event, 0x1D0401e0); in reserved_bits_mmcra_sample_elig_mode() [all …]
|
| /linux/tools/testing/selftests/powerpc/pmu/ebb/ |
| H A D | event_attributes_test.c | 17 struct event event, leader; in event_attributes() local 21 event_init(&event, 0x1001e); in event_attributes() 22 event_leader_ebb_init(&event); in event_attributes() 24 FAIL_IF(event_open(&event)); in event_attributes() 25 event_close(&event); in event_attributes() 28 event_init(&event, 0x001e); /* CYCLES - no PMC specified */ in event_attributes() 29 event_leader_ebb_init(&event); in event_attributes() 31 FAIL_IF(event_open(&event) == 0); in event_attributes() 34 event_init(&event, 0x2001e); in event_attributes() 35 event_leader_ebb_init(&event); in event_attributes() [all …]
|
| H A D | instruction_count_test.c | 25 static int do_count_loop(struct event *event, uint64_t instructions, in do_count_loop() argument 45 event->result.value = ebb_state.stats.pmc_count[4-1]; in do_count_loop() 47 difference = event->result.value - expected; in do_count_loop() 48 percentage = (double)difference / event->result.value * 100; in do_count_loop() 53 printf("Actual %llu\n", event->result.value); in do_count_loop() 63 if (difference / event->result.value) in do_count_loop() 70 static uint64_t determine_overhead(struct event *event) in determine_overhead() argument 75 do_count_loop(event, 0, 0, false); in determine_overhead() 76 overhead = event->result.value; in determine_overhead() 79 do_count_loop(event, 0, 0, false); in determine_overhead() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/core/ |
| H A D | event.c | 26 nvkm_event_put(struct nvkm_event *event, u32 types, int index) in nvkm_event_put() argument 28 assert_spin_locked(&event->refs_lock); in nvkm_event_put() 30 nvkm_trace(event->subdev, "event: decr %08x on %d\n", types, index); in nvkm_event_put() 34 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put() 35 nvkm_trace(event->subdev, "event: blocking %d on %d\n", type, index); in nvkm_event_put() 36 if (event->func->fini) in nvkm_event_put() 37 event->func->fini(event, 1 << type, index); in nvkm_event_put() 43 nvkm_event_get(struct nvkm_event *event, u32 types, int index) in nvkm_event_get() argument 45 assert_spin_locked(&event->refs_lock); in nvkm_event_get() 47 nvkm_trace(event->subdev, "event: incr %08x on %d\n", types, index); in nvkm_event_get() [all …]
|
| /linux/kernel/events/ |
| H A D | core.c | 225 static bool is_kernel_event(struct perf_event *event) in is_kernel_event() argument 227 return READ_ONCE(event->owner) == TASK_TOMBSTONE; in is_kernel_event() 261 struct perf_event *event; member 269 struct perf_event *event = efs->event; in event_function() local 270 struct perf_event_context *ctx = event->ctx; in event_function() 305 efs->func(event, cpuctx, ctx, efs->data); in event_function() 312 static void event_function_call(struct perf_event *event, event_f func, void *data) in event_function_call() argument 314 struct perf_event_context *ctx = event->ctx; in event_function_call() 318 .event = event, in event_function_call() 323 if (!event->parent) { in event_function_call() [all …]
|
| /linux/net/sctp/ |
| H A D | ulpevent.c | 31 static void sctp_ulpevent_receive_data(struct sctp_ulpevent *event, 33 static void sctp_ulpevent_release_data(struct sctp_ulpevent *event); 34 static void sctp_ulpevent_release_frag_data(struct sctp_ulpevent *event); 38 static void sctp_ulpevent_init(struct sctp_ulpevent *event, in sctp_ulpevent_init() argument 42 memset(event, 0, sizeof(struct sctp_ulpevent)); in sctp_ulpevent_init() 43 event->msg_flags = msg_flags; in sctp_ulpevent_init() 44 event->rmem_len = len; in sctp_ulpevent_init() 51 struct sctp_ulpevent *event; in sctp_ulpevent_new() local 58 event = sctp_skb2event(skb); in sctp_ulpevent_new() 59 sctp_ulpevent_init(event, msg_flags, skb->truesize); in sctp_ulpevent_new() [all …]
|
| /linux/arch/powerpc/perf/ |
| H A D | core-fsl-emb.c | 23 struct perf_event *event[MAX_HWEVENTS]; member 162 static void fsl_emb_pmu_read(struct perf_event *event) in fsl_emb_pmu_read() argument 166 if (event->hw.state & PERF_HES_STOPPED) in fsl_emb_pmu_read() 175 prev = local64_read(&event->hw.prev_count); in fsl_emb_pmu_read() 177 val = read_pmc(event->hw.idx); in fsl_emb_pmu_read() 178 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in fsl_emb_pmu_read() 182 local64_add(delta, &event->count); in fsl_emb_pmu_read() 183 local64_sub(delta, &event->hw.period_left); in fsl_emb_pmu_read() 255 struct perf_event *event; in collect_events() local 263 for_each_sibling_event(event, group) { in collect_events() [all …]
|
| H A D | core-book3s.c | 39 struct perf_event *event[MAX_HWEVENTS]; member 109 static inline void perf_get_data_addr(struct perf_event *event, struct pt_regs *regs, u64 *addrp) {… in perf_get_data_addr() argument 124 static bool is_ebb_event(struct perf_event *event) { return false; } in is_ebb_event() argument 125 static int ebb_event_check(struct perf_event *event) { return 0; } in ebb_event_check() argument 126 static void ebb_event_add(struct perf_event *event) { } in ebb_event_add() argument 133 static inline void power_pmu_bhrb_enable(struct perf_event *event) {} in power_pmu_bhrb_enable() argument 134 static inline void power_pmu_bhrb_disable(struct perf_event *event) {} in power_pmu_bhrb_disable() argument 139 static inline void power_pmu_bhrb_read(struct perf_event *event, struct cpu_hw_events *cpuhw) {} in power_pmu_bhrb_read() argument 205 static inline void perf_get_data_addr(struct perf_event *event, struct pt_regs *regs, u64 *addrp) in perf_get_data_addr() argument 230 if (is_kernel_addr(mfspr(SPRN_SDAR)) && event->attr.exclude_kernel) in perf_get_data_addr() [all …]
|
| /linux/Documentation/admin-guide/perf/ |
| H A D | mrvl-odyssey-ddr-pmu.rst | 11 Additionally DSS also supports two fixed performance event counters, one 24 mrvl_ddr_pmu_<>/ddr_act_bypass_access/ [Kernel PMU event] 25 mrvl_ddr_pmu_<>/ddr_bsm_alloc/ [Kernel PMU event] 26 mrvl_ddr_pmu_<>/ddr_bsm_starvation/ [Kernel PMU event] 27 mrvl_ddr_pmu_<>/ddr_cam_active_access/ [Kernel PMU event] 28 mrvl_ddr_pmu_<>/ddr_cam_mwr/ [Kernel PMU event] 29 mrvl_ddr_pmu_<>/ddr_cam_rd_active_access/ [Kernel PMU event] 30 mrvl_ddr_pmu_<>/ddr_cam_rd_or_wr_access/ [Kernel PMU event] 31 mrvl_ddr_pmu_<>/ddr_cam_read/ [Kernel PMU event] 32 mrvl_ddr_pmu_<>/ddr_cam_wr_access/ [Kernel PMU event] [all …]
|
| /linux/tools/perf/util/ |
| H A D | event.c | 155 struct perf_record_stat_config *event) in perf_event__read_stat_config() argument 159 for (i = 0; i < event->nr; i++) { in perf_event__read_stat_config() 161 switch (event->data[i].tag) { in perf_event__read_stat_config() 164 config->__val = event->data[i].val; \ in perf_event__read_stat_config() 174 event->data[i].tag); in perf_event__read_stat_config() 179 size_t perf_event__fprintf_comm(union perf_event *event, FILE *fp) in perf_event__fprintf_comm() argument 183 if (event->header.misc & PERF_RECORD_MISC_COMM_EXEC) in perf_event__fprintf_comm() 188 return fprintf(fp, "%s: %s:%d/%d\n", s, event->comm.comm, event->comm.pid, event->comm.tid); in perf_event__fprintf_comm() 191 size_t perf_event__fprintf_namespaces(union perf_event *event, FILE *fp) in perf_event__fprintf_namespaces() argument 197 ns_link_info = event->namespaces.link_info; in perf_event__fprintf_namespaces() [all …]
|
| H A D | session.c | 45 union perf_event *event, 129 struct ordered_event *event) in ordered_events__deliver_event() argument 134 return perf_session__deliver_event(session, event->event, in ordered_events__deliver_event() 135 session->tool, event->file_offset, in ordered_events__deliver_event() 136 event->file_path); in ordered_events__deliver_event() 268 static void swap_sample_id_all(union perf_event *event, void *data) in swap_sample_id_all() argument 270 void *end = (void *) event + event->header.size; in swap_sample_id_all() 277 static void perf_event__all64_swap(union perf_event *event, in perf_event__all64_swap() argument 280 struct perf_event_header *hdr = &event->header; in perf_event__all64_swap() 281 mem_bswap_64(hdr + 1, event->header.size - sizeof(*hdr)); in perf_event__all64_swap() [all …]
|
| /linux/drivers/perf/ |
| H A D | riscv_pmu.c | 21 static bool riscv_perf_user_access(struct perf_event *event) in riscv_perf_user_access() argument 23 return ((event->attr.type == PERF_TYPE_HARDWARE) || in riscv_perf_user_access() 24 (event->attr.type == PERF_TYPE_HW_CACHE) || in riscv_perf_user_access() 25 (event->attr.type == PERF_TYPE_RAW)) && in riscv_perf_user_access() 26 !!(event->hw.flags & PERF_EVENT_FLAG_USER_READ_CNT) && in riscv_perf_user_access() 27 (event->hw.idx != -1); in riscv_perf_user_access() 30 void arch_perf_update_userpage(struct perf_event *event, in arch_perf_update_userpage() argument 40 userpg->cap_user_rdpmc = riscv_perf_user_access(event); in arch_perf_update_userpage() 48 userpg->pmc_width = to_riscv_pmu(event->pmu)->ctr_get_width(event->hw.idx) + 1; in arch_perf_update_userpage() 145 u64 riscv_pmu_ctr_get_width_mask(struct perf_event *event) in riscv_pmu_ctr_get_width_mask() argument [all …]
|
| H A D | qcom_l3_pmu.c | 136 static inline u32 get_event_type(struct perf_event *event) in get_event_type() argument 138 return (event->attr.config) & L3_EVTYPE_MASK; in get_event_type() 141 static inline bool event_uses_long_counter(struct perf_event *event) in event_uses_long_counter() argument 143 return !!(event->attr.config & BIT_ULL(L3_EVENT_LC_BIT)); in event_uses_long_counter() 146 static inline int event_num_counters(struct perf_event *event) in event_num_counters() argument 148 return event_uses_long_counter(event) ? 2 : 1; in event_num_counters() 175 void (*start)(struct perf_event *event); 177 void (*stop)(struct perf_event *event, int flags); 179 void (*update)(struct perf_event *event); 193 static void qcom_l3_cache__64bit_counter_start(struct perf_event *event) in qcom_l3_cache__64bit_counter_start() argument [all …]
|
| H A D | arm_pmu.c | 108 static inline u64 arm_pmu_event_max_period(struct perf_event *event) in arm_pmu_event_max_period() argument 110 if (event->hw.flags & ARMPMU_EVT_64BIT) in arm_pmu_event_max_period() 112 else if (event->hw.flags & ARMPMU_EVT_63BIT) in arm_pmu_event_max_period() 114 else if (event->hw.flags & ARMPMU_EVT_47BIT) in arm_pmu_event_max_period() 174 armpmu_map_event(struct perf_event *event, in armpmu_map_event() argument 182 u64 config = event->attr.config; in armpmu_map_event() 183 int type = event->attr.type; in armpmu_map_event() 185 if (type == event->pmu->type) in armpmu_map_event() 200 int armpmu_event_set_period(struct perf_event *event) in armpmu_event_set_period() argument 202 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period() [all …]
|
| /linux/drivers/md/ |
| H A D | dm-uevent.c | 39 static void dm_uevent_free(struct dm_uevent *event) in dm_uevent_free() argument 41 kmem_cache_free(_dm_event_cache, event); in dm_uevent_free() 46 struct dm_uevent *event; in dm_uevent_alloc() local 48 event = kmem_cache_zalloc(_dm_event_cache, GFP_ATOMIC); in dm_uevent_alloc() 49 if (!event) in dm_uevent_alloc() 52 INIT_LIST_HEAD(&event->elist); in dm_uevent_alloc() 53 event->md = md; in dm_uevent_alloc() 55 return event; in dm_uevent_alloc() 65 struct dm_uevent *event; in dm_build_path_uevent() local 67 event = dm_uevent_alloc(md); in dm_build_path_uevent() [all …]
|
| /linux/drivers/firmware/ |
| H A D | arm_sdei.c | 77 struct sdei_event *event; member 82 #define CROSSCALL_INIT(arg, event) \ argument 84 arg.event = event; \ 90 struct sdei_event *event) in sdei_do_local_call() argument 94 CROSSCALL_INIT(arg, event); in sdei_do_local_call() 101 struct sdei_event *event) in sdei_do_cross_call() argument 105 CROSSCALL_INIT(arg, event); in sdei_do_cross_call() 191 static int sdei_api_event_get_info(u32 event, u32 info, u64 *result) in sdei_api_event_get_info() argument 193 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_GET_INFO, event, info, 0, in sdei_api_event_get_info() 203 struct sdei_event *event; in sdei_event_create() local [all …]
|
| /linux/arch/s390/kernel/ |
| H A D | perf_pai_crypto.c | 38 struct perf_event *event; /* Perf event for sampling */ member 95 static void paicrypt_event_destroy_cpu(struct perf_event *event, int cpu) in paicrypt_event_destroy_cpu() argument 102 "refcnt %u\n", __func__, event->attr.config, in paicrypt_event_destroy_cpu() 103 event->cpu, cpump->active_events, in paicrypt_event_destroy_cpu() 111 static void paicrypt_event_destroy(struct perf_event *event) in paicrypt_event_destroy() argument 116 free_page(PAI_SAVE_AREA(event)); in paicrypt_event_destroy() 117 if (event->cpu == -1) { in paicrypt_event_destroy() 118 struct cpumask *mask = PAI_CPU_MASK(event); in paicrypt_event_destroy() 121 paicrypt_event_destroy_cpu(event, cpu); in paicrypt_event_destroy() 124 paicrypt_event_destroy_cpu(event, event->cpu); in paicrypt_event_destroy() [all …]
|
| H A D | perf_pai_ext.c | 51 struct perf_event *event; /* Perf event for sampling */ member 119 static void paiext_event_destroy_cpu(struct perf_event *event, int cpu) in paiext_event_destroy_cpu() argument 131 static void paiext_event_destroy(struct perf_event *event) in paiext_event_destroy() argument 135 free_page(PAI_SAVE_AREA(event)); in paiext_event_destroy() 136 if (event->cpu == -1) { in paiext_event_destroy() 137 struct cpumask *mask = PAI_CPU_MASK(event); in paiext_event_destroy() 140 paiext_event_destroy_cpu(event, cpu); in paiext_event_destroy() 143 paiext_event_destroy_cpu(event, event->cpu); in paiext_event_destroy() 146 event->cpu); in paiext_event_destroy() 163 static int paiext_alloc_cpu(struct perf_event *event, int cpu) in paiext_alloc_cpu() argument [all …]
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-bus-event_source-devices-dfl_fme | 13 event = "config:0-11" - event ID 14 evtype = "config:12-15" - event type 15 portid = "config:16-23" - event source 19 fab_mmio_read = "event=0x06,evtype=0x02,portid=0xff" 21 It shows this fab_mmio_read is a fabric type (0x02) event with 22 0x06 local event id for overall monitoring (portid=0xff). 37 a single performance monitoring event supported by this fme pmu. 38 The name of the file is the name of the event. 45 clock = "event=0x00,evtype=0x00,portid=0xff" 49 cache_read_hit = "event=0x00,evtype=0x01,portid=0xff" [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | eeh_event.c | 42 struct eeh_event *event; in eeh_event_handler() local 50 event = NULL; in eeh_event_handler() 52 event = list_entry(eeh_eventlist.next, in eeh_event_handler() 54 list_del(&event->list); in eeh_event_handler() 57 if (!event) in eeh_event_handler() 61 if (event->pe) in eeh_event_handler() 62 eeh_handle_normal_event(event->pe); in eeh_event_handler() 66 kfree(event); in eeh_event_handler() 105 struct eeh_event *event; in __eeh_send_failure_event() local 107 event = kzalloc(sizeof(*event), GFP_ATOMIC); in __eeh_send_failure_event() [all …]
|
| /linux/scripts/dtc/ |
| H A D | yamltree.c | 24 #define yaml_emitter_emit_or_die(emitter, event) ( \ argument 26 if (!yaml_emitter_emit(emitter, event)) \ 35 yaml_event_t event; in yaml_propval_int() local 49 yaml_sequence_start_event_initialize(&event, NULL, in yaml_propval_int() 51 yaml_emitter_emit_or_die(emitter, &event); in yaml_propval_int() 82 yaml_scalar_event_initialize(&event, NULL, in yaml_propval_int() 86 yaml_scalar_event_initialize(&event, NULL, in yaml_propval_int() 89 yaml_emitter_emit_or_die(emitter, &event); in yaml_propval_int() 92 yaml_sequence_end_event_initialize(&event); in yaml_propval_int() 93 yaml_emitter_emit_or_die(emitter, &event); in yaml_propval_int() [all …]
|
| /linux/arch/sh/kernel/ |
| H A D | perf_event.c | 63 static void hw_perf_event_destroy(struct perf_event *event) in hw_perf_event_destroy() argument 100 static int __hw_perf_event_init(struct perf_event *event) in __hw_perf_event_init() argument 102 struct perf_event_attr *attr = &event->attr; in __hw_perf_event_init() 103 struct hw_perf_event *hwc = &event->hw; in __hw_perf_event_init() 131 event->destroy = hw_perf_event_destroy; in __hw_perf_event_init() 158 static void sh_perf_event_update(struct perf_event *event, in sh_perf_event_update() argument 196 local64_add(delta, &event->count); in sh_perf_event_update() 199 static void sh_pmu_stop(struct perf_event *event, int flags) in sh_pmu_stop() argument 202 struct hw_perf_event *hwc = &event->hw; in sh_pmu_stop() 205 if (!(event->hw.state & PERF_HES_STOPPED)) { in sh_pmu_stop() [all …]
|
| /linux/fs/notify/fanotify/ |
| H A D | fanotify.h | 267 static inline void fanotify_init_event(struct fanotify_event *event, in fanotify_init_event() argument 270 fsnotify_init_event(&event->fse); in fanotify_init_event() 271 INIT_HLIST_NODE(&event->merge_list); in fanotify_init_event() 272 event->hash = hash; in fanotify_init_event() 273 event->mask = mask; in fanotify_init_event() 274 event->pid = NULL; in fanotify_init_event() 292 FANOTIFY_FE(struct fanotify_event *event) in FANOTIFY_FE() argument 294 return container_of(event, struct fanotify_fid_event, fae); in FANOTIFY_FE() 304 FANOTIFY_NE(struct fanotify_event *event) in FANOTIFY_NE() argument 306 return container_of(event, struct fanotify_name_event, fae); in FANOTIFY_NE() [all …]
|
| /linux/drivers/perf/hisilicon/ |
| H A D | hisi_uncore_pmu.c | 97 static bool hisi_validate_event_group(struct perf_event *event) in hisi_validate_event_group() argument 99 struct perf_event *sibling, *leader = event->group_leader; in hisi_validate_event_group() 100 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_validate_event_group() 109 if (leader->pmu != event->pmu) in hisi_validate_event_group() 113 if (leader != event) in hisi_validate_event_group() 117 for_each_sibling_event(sibling, event->group_leader) { in hisi_validate_event_group() 120 if (sibling->pmu != event->pmu) in hisi_validate_event_group() 130 int hisi_uncore_pmu_get_event_idx(struct perf_event *event) in hisi_uncore_pmu_get_event_idx() argument 132 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_uncore_pmu_get_event_idx() 155 struct perf_event *event; in hisi_uncore_pmu_isr() local [all …]
|