Home
last modified time | relevance | path

Searched refs:prev_count (Results 1 – 25 of 52) sorted by relevance

123

/linux/drivers/perf/
H A Dmarvell_pem_pmu.c238 u64 prev_count, new_count; in pem_perf_event_update() local
241 prev_count = local64_read(&hwc->prev_count); in pem_perf_event_update()
243 } while (local64_xchg(&hwc->prev_count, new_count) != prev_count); in pem_perf_event_update()
245 local64_add((new_count - prev_count), &event->count); in pem_perf_event_update()
258 local64_set(&hwc->prev_count, in pem_perf_event_start()
H A Dmarvell_cn10k_ddr_pmu.c595 u64 prev_count, new_count, mask; in cn10k_ddr_perf_event_update() local
598 prev_count = local64_read(&hwc->prev_count); in cn10k_ddr_perf_event_update()
600 } while (local64_xchg(&hwc->prev_count, new_count) != prev_count); in cn10k_ddr_perf_event_update()
604 local64_add((new_count - prev_count) & mask, &event->count); in cn10k_ddr_perf_event_update()
613 local64_set(&hwc->prev_count, 0); in cn10k_ddr_perf_event_start()
733 local64_set(&hwc->prev_count, 0); in cn10k_ddr_perf_event_update_all()
856 u64 prev_count, new_count; in cn10k_ddr_pmu_overflow_handler() local
863 prev_count in cn10k_ddr_pmu_overflow_handler()
[all...]
H A Darm_dsu_pmu.c334 u64 delta, prev_count, new_count; in dsu_pmu_event_update() local
338 prev_count = local64_read(&hwc->prev_count); in dsu_pmu_event_update()
340 } while (local64_cmpxchg(&hwc->prev_count, prev_count, new_count) != in dsu_pmu_event_update()
341 prev_count); in dsu_pmu_event_update()
342 delta = (new_count - prev_count) & DSU_PMU_COUNTER_MASK(hwc->idx); in dsu_pmu_event_update()
368 local64_set(&event->hw.prev_count, val); in dsu_pmu_set_event_period()
H A Darm_dmc620_pmu.c327 u64 delta, prev_count, new_count; in dmc620_pmu_event_update() local
331 prev_count = local64_read(&hwc->prev_count); in dmc620_pmu_event_update()
333 } while (local64_cmpxchg(&hwc->prev_count, in dmc620_pmu_event_update()
334 prev_count, new_count) != prev_count); in dmc620_pmu_event_update()
335 delta = (new_count - prev_count) & DMC620_CNT_MAX_PERIOD; in dmc620_pmu_event_update()
343 local64_set(&event->hw.prev_count, DMC620_CNT_INIT); in dmc620_pmu_event_set_period()
H A Dqcom_l3_pmu.c205 /* Initialize the hardware counters and reset prev_count*/ in qcom_l3_cache__64bit_counter_start()
206 local64_set(&event->hw.prev_count, 0); in qcom_l3_cache__64bit_counter_start()
247 prev = local64_read(&event->hw.prev_count); in qcom_l3_cache__64bit_counter_update()
253 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in qcom_l3_cache__64bit_counter_update()
284 /* Initialize the hardware counter and reset prev_count*/ in qcom_l3_cache__32bit_counter_start()
285 local64_set(&event->hw.prev_count, 0); in qcom_l3_cache__32bit_counter_start()
323 prev = local64_read(&event->hw.prev_count); in qcom_l3_cache__32bit_counter_update()
325 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in qcom_l3_cache__32bit_counter_update()
H A Dstarfive_starlink_pmu.c167 local64_set(&hwc->prev_count, val); in starlink_pmu_set_event_period()
242 prev_raw_count = local64_read(&hwc->prev_count); in starlink_pmu_update()
250 oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, in starlink_pmu_update()
/linux/drivers/pps/generators/
H A Dpps_gen_tio.c47 u32 prev_count; member
86 tio->prev_count = 0; in pps_tio_disable()
144 if (tio->prev_count && tio->prev_count == event_count) in hrtimer_callback()
146 tio->prev_count = event_count; in hrtimer_callback()
/linux/arch/arm/mm/
H A Dcache-l2x0-pmu.c120 u64 prev_count, new_count, mask; in l2x0_pmu_event_read() local
123 prev_count = local64_read(&hw->prev_count); in l2x0_pmu_event_read()
125 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); in l2x0_pmu_event_read()
128 local64_add((new_count - prev_count) & mask, &event->count); in l2x0_pmu_event_read()
147 local64_set(&hw->prev_count, 0); in l2x0_pmu_event_configure()
/linux/arch/powerpc/perf/
H A D8xx-pmu.c115 local64_set(&event->hw.prev_count, val); in mpc8xx_pmu_add()
128 prev = local64_read(&event->hw.prev_count); in mpc8xx_pmu_read()
149 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in mpc8xx_pmu_read()
H A Dcore-fsl-emb.c175 prev = local64_read(&event->hw.prev_count); in fsl_emb_pmu_read()
178 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in fsl_emb_pmu_read()
314 local64_set(&event->hw.prev_count, val); in fsl_emb_pmu_add()
603 prev = local64_read(&event->hw.prev_count); in record_and_restart()
626 local64_set(&event->hw.prev_count, val); in record_and_restart()
H A Dvpa-pmu.c132 local64_set(&event->hw.prev_count, data); in vpa_pmu_add()
141 prev_data = local64_read(&event->hw.prev_count); in vpa_pmu_read()
H A Dkvm-hv-pmu.c167 prev_val = local64_xchg(&event->hw.prev_count, curr_val); in kvmppc_pmu_event_update()
189 local64_set(&event->hw.prev_count, 0); in kvmppc_pmu_event_init()
/linux/fs/gfs2/
H A Dfile.c781 size_t *prev_count, in should_fault_in_pages() argument
798 if (*prev_count != count) { in should_fault_in_pages()
806 *prev_count = count; in should_fault_in_pages()
816 size_t prev_count = 0, window_size = 0; in gfs2_file_direct_read() local
857 if (should_fault_in_pages(to, iocb, &prev_count, &window_size)) { in gfs2_file_direct_read()
880 size_t prev_count = 0, window_size = 0; in gfs2_file_direct_write() local
926 enough_retries = prev_count == iov_iter_count(from) && in gfs2_file_direct_write()
928 if (should_fault_in_pages(from, iocb, &prev_count, &window_size)) { in gfs2_file_direct_write()
953 size_t prev_count = 0, window_size = 0; in gfs2_file_read_iter() local
996 if (should_fault_in_pages(to, iocb, &prev_count, in gfs2_file_read_iter()
1019 size_t prev_count = 0, window_size = 0; gfs2_file_buffered_write() local
[all...]
/linux/drivers/iommu/intel/
H A Dperfmon.c305 u64 prev_count, new_count, delta; in iommu_pmu_event_update() local
309 prev_count = local64_read(&hwc->prev_count); in iommu_pmu_event_update()
311 if (local64_xchg(&hwc->prev_count, new_count) != prev_count) in iommu_pmu_event_update()
318 delta = (new_count << shift) - (prev_count << shift); in iommu_pmu_event_update()
344 local64_set((&hwc->prev_count), count); in iommu_pmu_start()
/linux/drivers/perf/hisilicon/
H A Dhisi_uncore_pmu.c308 local64_set(&hwc->prev_count, val); in hisi_uncore_pmu_set_event_period()
323 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_event_update()
324 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in hisi_uncore_pmu_event_update()
348 u64 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_start()
H A Dhisi_pcie_pmu.c452 prev_cnt = local64_read(&hwc->prev_count); in hisi_pcie_pmu_event_update()
454 } while (local64_cmpxchg(&hwc->prev_count, prev_cnt, in hisi_pcie_pmu_event_update()
475 local64_set(&hwc->prev_count, HISI_PCIE_INIT_VAL); in hisi_pcie_pmu_set_period()
483 * the hwc->prev_count correctly. Otherwise the final counts user in hisi_pcie_pmu_set_period()
488 local64_set(&hwc->prev_count, cnt); in hisi_pcie_pmu_set_period()
550 prev_cnt = local64_read(&hwc->prev_count); in hisi_pcie_pmu_start()
/linux/arch/xtensa/kernel/
H A Dperf_event.c152 prev_raw_count = local64_read(&hwc->prev_count); in xtensa_perf_event_update()
154 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in xtensa_perf_event_update()
190 local64_set(&hwc->prev_count, -left); in xtensa_perf_event_set_period()
/linux/tools/perf/arch/x86/util/
H A Diostat.c442 struct perf_counts_values *prev_count = in iostat_print_metric() local
445 prev_count_val = prev_count->val; in iostat_print_metric()
446 prev_count->val = count->val; in iostat_print_metric()
/linux/arch/x86/events/amd/
H A Duncore.c143 * we do not have to worry about prev_count changing on us in amd_uncore_read()
146 prev = local64_read(&hwc->prev_count); in amd_uncore_read()
157 local64_set(&hwc->prev_count, new); in amd_uncore_read()
173 wrmsrq(hwc->event_base, (u64)local64_read(&hwc->prev_count)); in amd_uncore_start()
951 wrmsrq(hwc->event_base, (u64)local64_read(&hwc->prev_count)); in amd_uncore_umc_start()
966 prev = local64_read(&hwc->prev_count); in amd_uncore_umc_read()
982 local64_set(&hwc->prev_count, 0); in amd_uncore_umc_read()
984 local64_set(&hwc->prev_count, new); in amd_uncore_umc_read()
/linux/arch/sh/kernel/
H A Dperf_event.c178 prev_raw_count = local64_read(&hwc->prev_count); in sh_perf_event_update()
181 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sh_perf_event_update()
/linux/drivers/misc/
H A Dntsync.c411 __u32 prev_count; in ntsync_sem_release() local
424 prev_count = sem->u.sem.count; in ntsync_sem_release()
434 if (!ret && put_user(prev_count, user_args)) in ntsync_sem_release()
461 __u32 prev_count; in ntsync_mutex_unlock() local
475 prev_count = mutex->u.mutex.count; in ntsync_mutex_unlock()
485 if (!ret && put_user(prev_count, &user_args->count)) in ntsync_mutex_unlock()
/linux/arch/arc/kernel/
H A Dperf_event.c283 u64 prev_raw_count = local64_read(&hwc->prev_count); in arc_perf_event_update()
288 * We aren't afraid of hwc->prev_count changing beneath our feet in arc_perf_event_update()
291 local64_set(&hwc->prev_count, new_raw_count); in arc_perf_event_update()
432 local64_set(&hwc->prev_count, value); in arc_pmu_event_set_period()
550 local64_set(&hwc->prev_count, 0); in arc_pmu_add()
/linux/drivers/gpu/drm/xe/
H A Dxe_pmu.c321 prev = local64_read(&hwc->prev_count); in xe_pmu_event_update()
324 } while (!local64_try_cmpxchg(&hwc->prev_count, &prev, new)); in xe_pmu_event_update()
356 local64_set(&event->hw.prev_count, __xe_pmu_event_read(event)); in xe_pmu_enable()
/linux/drivers/dma/idxd/
H A Dperfmon.c217 prev_raw_count = local64_read(&hwc->prev_count); in perfmon_pmu_event_update()
220 } while (!local64_try_cmpxchg(&hwc->prev_count, in perfmon_pmu_event_update()
330 local64_set(&event->hw.prev_count, cntrdata); in perfmon_pmu_event_start()
/linux/arch/s390/kernel/
H A Dperf_pai_ext.c355 prev = local64_read(&event->hw.prev_count); in paiext_read()
357 local64_set(&event->hw.prev_count, new); in paiext_read()
370 local64_set(&event->hw.prev_count, sum); in paiext_start()

123