/linux/include/asm-generic/ |
H A D | local64.h | 31 #define local64_set(l,i) local_set((&(l)->a),(i)) macro 61 #define __local64_inc(l) local64_set((l), local64_read(l) + 1) 62 #define __local64_dec(l) local64_set((l), local64_read(l) - 1) 63 #define __local64_add(i,l) local64_set((l), local64_read(l) + (i)) 64 #define __local64_sub(i,l) local64_set((l), local64_read(l) - (i)) 78 #define local64_set(l,i) atomic64_set((&(l)->a),(i)) macro 100 #define __local64_inc(l) local64_set((l), local64_read(l) + 1) 101 #define __local64_dec(l) local64_set((l), local64_read(l) - 1) 102 #define __local64_add(i,l) local64_set((l), local64_read(l) + (i)) 103 #define __local64_sub(i,l) local64_set(( [all...] |
/linux/arch/arc/kernel/ |
H A D | perf_event.c | 291 local64_set(&hwc->prev_count, new_raw_count); in arc_perf_event_update() 337 local64_set(&hwc->period_left, hwc->sample_period); in arc_pmu_event_init() 417 local64_set(&hwc->period_left, left); in arc_pmu_event_set_period() 423 local64_set(&hwc->period_left, left); in arc_pmu_event_set_period() 432 local64_set(&hwc->prev_count, value); in arc_pmu_event_set_period() 550 local64_set(&hwc->prev_count, 0); in arc_pmu_add()
|
/linux/arch/loongarch/kernel/ |
H A D | perf_event.c | 310 local64_set(&hwc->period_left, left); in loongarch_pmu_event_set_period() 316 local64_set(&hwc->period_left, left); in loongarch_pmu_event_set_period() 323 local64_set(&hwc->period_left, left); in loongarch_pmu_event_set_period() 326 local64_set(&hwc->prev_count, loongarch_pmu.overflow - left); in loongarch_pmu_event_set_period() 797 local64_set(&hwc->period_left, hwc->sample_period); in __hw_perf_event_init()
|
/linux/arch/powerpc/perf/ |
H A D | core-fsl-emb.c | 314 local64_set(&event->hw.prev_count, val); in fsl_emb_pmu_add() 547 local64_set(&event->hw.period_left, event->hw.last_period); in fsl_emb_pmu_event_init() 626 local64_set(&event->hw.prev_count, val); in record_and_restart() 627 local64_set(&event->hw.period_left, left); in record_and_restart()
|
H A D | kvm-hv-pmu.c | 189 local64_set(&event->hw.prev_count, 0); in kvmppc_pmu_event_init() 190 local64_set(&event->count, 0); in kvmppc_pmu_event_init()
|
H A D | vpa-pmu.c | 132 local64_set(&event->hw.prev_count, data); in vpa_pmu_add()
|
H A D | 8xx-pmu.c | 115 local64_set(&event->hw.prev_count, val); in mpc8xx_pmu_add()
|
/linux/arch/xtensa/kernel/ |
H A D | perf_event.c | 177 local64_set(&hwc->period_left, left); in xtensa_perf_event_set_period() 182 local64_set(&hwc->period_left, left); in xtensa_perf_event_set_period() 190 local64_set(&hwc->prev_count, -left); in xtensa_perf_event_set_period()
|
/linux/drivers/perf/ |
H A D | arm_pmu.c | 212 local64_set(&hwc->period_left, left); in armpmu_event_set_period() 219 local64_set(&hwc->period_left, left); in armpmu_event_set_period() 233 local64_set(&hwc->prev_count, (u64)-left); in armpmu_event_set_period() 502 local64_set(&hwc->period_left, hwc->sample_period); in __hw_perf_event_init()
|
H A D | riscv_pmu_legacy.c | 77 local64_set(&hwc->prev_count, initial_val); in pmu_legacy_ctr_start()
|
H A D | qcom_l3_pmu.c | 206 local64_set(&event->hw.prev_count, 0); in qcom_l3_cache__64bit_counter_start() 285 local64_set(&event->hw.prev_count, 0); in qcom_l3_cache__32bit_counter_start()
|
H A D | marvell_pem_pmu.c | 258 local64_set(&hwc->prev_count, in pem_perf_event_start()
|
H A D | arm-ni.c | 390 local64_set(&cd->ccnt->hw.prev_count, S64_MIN); in arm_ni_init_ccnt() 396 local64_set(&cd->evcnt[idx]->hw.prev_count, S32_MIN); in arm_ni_init_evcnt()
|
H A D | arm_smmuv3_pmu.c | 299 local64_set(&hwc->prev_count, new); in smmu_pmu_set_period() 500 local64_set(&hwc->prev_count, 0); in smmu_pmu_event_add()
|
H A D | marvell_cn10k_ddr_pmu.c | 613 local64_set(&hwc->prev_count, 0); in cn10k_ddr_perf_event_start() 733 local64_set(&hwc->prev_count, 0); in cn10k_ddr_perf_event_update_all()
|
/linux/net/netfilter/ |
H A D | nf_conntrack_ecache.c | 169 local64_set(&e->timestamp, ktime_get_real_ns()); in nf_ct_ecache_tstamp_refresh() 318 local64_set(&e->timestamp, ts); in nf_ct_ecache_tstamp_new()
|
/linux/include/linux/ |
H A D | u64_stats_sync.h | 84 local64_set(&p->v, val); in u64_stats_set()
|
/linux/include/net/netfilter/ |
H A D | nf_conntrack_ecache.h | 120 local64_set(&e->timestamp, ktime_get_real_ns()); in nf_conntrack_event_cache()
|
/linux/arch/csky/kernel/ |
H A D | perf_event.c | 887 local64_set(&hwc->period_left, left); in csky_pmu_event_set_period() 894 local64_set(&hwc->period_left, left); in csky_pmu_event_set_period() 906 local64_set(&hwc->prev_count, (u64)(-left)); in csky_pmu_event_set_period() 935 local64_set(&hwc->prev_count, new_raw_count); in csky_perf_event_update()
|
/linux/arch/mips/kernel/ |
H A D | perf_event_mipsxx.c | 420 local64_set(&hwc->period_left, left); in mipspmu_event_set_period() 426 local64_set(&hwc->period_left, left); in mipspmu_event_set_period() 433 local64_set(&hwc->period_left, left); in mipspmu_event_set_period() 436 local64_set(&hwc->prev_count, mipspmu.overflow - left); in mipspmu_event_set_period() 1530 local64_set(&hwc->period_left, hwc->sample_period); in __hw_perf_event_init()
|
/linux/arch/s390/kernel/ |
H A D | perf_pai_ext.c | 357 local64_set(&event->hw.prev_count, new); in paiext_read() 370 local64_set(&event->hw.prev_count, sum); in paiext_start()
|
/linux/arch/x86/events/amd/ |
H A D | uncore.c | 157 local64_set(&hwc->prev_count, new); in amd_uncore_read() 982 local64_set(&hwc->prev_count, 0); in amd_uncore_umc_read() 984 local64_set(&hwc->prev_count, new); in amd_uncore_umc_read()
|
/linux/arch/arm/mm/ |
H A D | cache-l2x0-pmu.c | 147 local64_set(&hw->prev_count, 0); in l2x0_pmu_event_configure()
|
/linux/drivers/perf/hisilicon/ |
H A D | hisi_pcie_pmu.c | 475 local64_set(&hwc->prev_count, HISI_PCIE_INIT_VAL); in hisi_pcie_pmu_set_period() 488 local64_set(&hwc->prev_count, cnt); in hisi_pcie_pmu_set_period()
|
H A D | hisi_uncore_pmu.c | 308 local64_set(&hwc->prev_count, val); in hisi_uncore_pmu_set_event_period()
|