/linux/tools/testing/selftests/bpf/progs/ |
H A D | test_rdonly_maps.c | 21 unsigned iters; member 30 unsigned iters = 0, sum = 0; in skip_loop() local 34 iters++; in skip_loop() 39 res.iters = iters; in skip_loop() 49 unsigned iters = 0, sum = 0; in part_loop() local 53 iters++; in part_loop() 58 res.iters = iters; in part_loop() 69 unsigned iters in full_loop() local [all...] |
/linux/kernel/time/ |
H A D | test_udelay.c | 27 static int udelay_test_single(struct seq_file *s, int usecs, uint32_t iters) in udelay_test_single() argument 36 for (i = 0; i < iters; ++i) { in udelay_test_single() 56 do_div(avg, iters); in udelay_test_single() 58 usecs, iters, usecs * 1000, in udelay_test_single() 70 int iters; in udelay_test_show() local 75 iters = udelay_test_iterations; in udelay_test_show() 78 if (usecs > 0 && iters > 0) { in udelay_test_show() 79 return udelay_test_single(s, usecs, iters); in udelay_test_show() 105 int iters; in udelay_test_write() local 114 ret = sscanf(lbuf, "%d %d", &usecs, &iters); in udelay_test_write() [all...] |
/linux/kernel/kcsan/ |
H A D | debugfs.c | 54 * multiple threads, pipe 'microbench=<iters>' from multiple tasks into the 57 static noinline void microbenchmark(unsigned long iters) in microbenchmark() argument 71 pr_info("%s begin | iters: %lu\n", __func__, iters); in microbenchmark() 74 while (iters--) { in microbenchmark() 75 unsigned long addr = iters & ((PAGE_SIZE << 8) - 1); in microbenchmark() 76 int type = !(iters & 0x7f) ? KCSAN_ACCESS_ATOMIC : in microbenchmark() 77 (!(iters & 0xf) ? KCSAN_ACCESS_WRITE : 0); in microbenchmark() 237 unsigned long iters; in debugfs_write() local 239 if (kstrtoul(&arg[strlen("microbench=")], 0, &iters)) in debugfs_write() [all...] |
/linux/arch/powerpc/lib/ |
H A D | qspinlock.c | 442 static __always_inline bool steal_break(u32 val, int iters, bool paravirt, bool sleepy) in steal_break() argument 444 if (iters >= get_steal_spins(paravirt, sleepy)) in steal_break() 448 (iters >= get_remote_steal_spins(paravirt, sleepy))) { in steal_break() 460 int iters = 0; in try_to_steal_lock() local 507 iters++; in try_to_steal_lock() 509 * pv_spin_on_preempted_owner don't increase iters in try_to_steal_lock() 517 iters++; in try_to_steal_lock() 519 } while (!steal_break(val, iters, paravirt, sleepy)); in try_to_steal_lock() 535 int iters = 0; in queued_spin_lock_mcs_queue() local 645 iters in queued_spin_lock_mcs_queue() [all...] |
/linux/tools/testing/selftests/bpf/prog_tests/ |
H A D | rdonly_maps.c | 6 unsigned iters; member 76 if (CHECK(bss.iters != t->exp_iters, "check_iters", in test_rdonly_maps() 77 "prog '%s' iters: %d, expected: %d\n", in test_rdonly_maps() 78 t->prog_name, bss.iters, t->exp_iters)) in test_rdonly_maps()
|
H A D | iters.c | 13 #include "iters.skel.h" 298 RUN_TESTS(iters); in test_iters()
|
/linux/fs/bcachefs/ |
H A D | btree_journal_iter.c | 80 unsigned iters = 0; in bch2_journal_keys_peek_max() local 91 iters++; in bch2_journal_keys_peek_max() 92 if (iters == 10) { in bch2_journal_keys_peek_max() 119 iters++; in bch2_journal_keys_peek_max() 120 if (iters == 10) { in bch2_journal_keys_peek_max() 136 unsigned iters = 0; in bch2_journal_keys_peek_prev_min() local 150 iters++; in bch2_journal_keys_peek_prev_min() 151 if (iters == 10) { in bch2_journal_keys_peek_prev_min() 185 iters++; in bch2_journal_keys_peek_prev_min() 186 if (iters in bch2_journal_keys_peek_prev_min() 560 size_t iters = 0; bch2_btree_and_journal_iter_peek() local [all...] |
/linux/drivers/staging/vc04_services/interface/ |
H A D | TESTING | 47 Functional test - iters:10 58 Ping test - service:echo, iters:1000, version 3
|
/linux/tools/testing/selftests/x86/ |
H A D | ldt_gdt.c | 522 int failures = 0, iters = 5, i; in do_multicpu_tests() local 609 printf("[FAIL]\t%d of %d iterations failed\n", failures, iters); in do_multicpu_tests() 612 printf("[OK]\tAll %d iterations succeeded\n", iters); in do_multicpu_tests()
|
/linux/lib/zstd/decompress/ |
H A D | huf_decompress.c | 754 /* We can safely run iters iterations before running bounds checks */ in HUF_decompress4X1_usingDTable_internal_fast_c_loop() 755 size_t const iters = MIN(oiters, iiters); in HUF_decompress4X1_usingDTable_internal_fast_c_loop() local 756 size_t const symbols = iters * 5; in HUF_decompress4X1_usingDTable_internal_fast_c_loop() 760 * iters iterations, which only happens when op[3] == olimit. in HUF_decompress4X1_usingDTable_internal_fast_c_loop() 1562 * iters loops before running out of input. in HUF_decompress4X2_usingDTable_internal_fast_c_loop() 1564 size_t iters = (size_t)(ip[0] - ilowest) / 7; in HUF_decompress4X2_usingDTable_internal_fast_c_loop() local 1571 iters = MIN(iters, oiters); in HUF_decompress4X2_usingDTable_internal_fast_c_loop() 1575 * op[3] crosses olimit, we know we haven't executed iters in HUF_decompress4X2_usingDTable_internal_fast_c_loop() 1576 * iterations yet. This saves us maintaining an iters counte in HUF_decompress4X2_usingDTable_internal_fast_c_loop() [all...] |
/linux/tools/testing/selftests/rseq/ |
H A D | param_test.c | 1258 const int iters = opt_reps; in test_membarrier_worker_thread() local 1270 for (i = 0; i < iters; ++i) { in test_membarrier_worker_thread()
|
/linux/arch/sparc/lib/ |
H A D | checksum_32.S | 122 subcc %o3, 128, %o3 ! detract from loop iters
|
/linux/drivers/net/ethernet/broadcom/ |
H A D | tg3.c | 3585 const int iters = 10000; in tg3_pause_cpu() local 3587 for (i = 0; i < iters; i++) { in tg3_pause_cpu() 3596 return (i == iters) ? -EBUSY : 0; in tg3_pause_cpu() 3765 const int iters = 5; in tg3_pause_cpu_and_set_pc() local 3770 for (i = 0; i < iters; i++) { in tg3_pause_cpu_and_set_pc() 3779 return (i == iters) ? -EBUSY : 0; in tg3_pause_cpu_and_set_pc() 3826 const int iters = 1000; in tg3_validate_rxcpu_state() local 3833 for (i = 0; i < iters; i++) { in tg3_validate_rxcpu_state() 3840 if (i == iters) { in tg3_validate_rxcpu_state()
|