Lines Matching full:s64

1593 static __always_inline s64
1610 static __always_inline s64
1620 * @i: s64 value to assign
1629 atomic64_set(atomic64_t *v, s64 i) in atomic64_set()
1638 * @i: s64 value to assign
1647 atomic64_set_release(atomic64_t *v, s64 i) in atomic64_set_release()
1656 * @i: s64 value to add
1666 atomic64_add(s64 i, atomic64_t *v) in atomic64_add()
1674 * @i: s64 value to add
1683 static __always_inline s64
1684 atomic64_add_return(s64 i, atomic64_t *v) in atomic64_add_return()
1693 * @i: s64 value to add
1702 static __always_inline s64
1703 atomic64_add_return_acquire(s64 i, atomic64_t *v) in atomic64_add_return_acquire()
1711 * @i: s64 value to add
1720 static __always_inline s64
1721 atomic64_add_return_release(s64 i, atomic64_t *v) in atomic64_add_return_release()
1730 * @i: s64 value to add
1739 static __always_inline s64
1740 atomic64_add_return_relaxed(s64 i, atomic64_t *v) in atomic64_add_return_relaxed()
1748 * @i: s64 value to add
1757 static __always_inline s64
1758 atomic64_fetch_add(s64 i, atomic64_t *v) in atomic64_fetch_add()
1767 * @i: s64 value to add
1776 static __always_inline s64
1777 atomic64_fetch_add_acquire(s64 i, atomic64_t *v) in atomic64_fetch_add_acquire()
1785 * @i: s64 value to add
1794 static __always_inline s64
1795 atomic64_fetch_add_release(s64 i, atomic64_t *v) in atomic64_fetch_add_release()
1804 * @i: s64 value to add
1813 static __always_inline s64
1814 atomic64_fetch_add_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_add_relaxed()
1822 * @i: s64 value to subtract
1832 atomic64_sub(s64 i, atomic64_t *v) in atomic64_sub()
1840 * @i: s64 value to subtract
1849 static __always_inline s64
1850 atomic64_sub_return(s64 i, atomic64_t *v) in atomic64_sub_return()
1859 * @i: s64 value to subtract
1868 static __always_inline s64
1869 atomic64_sub_return_acquire(s64 i, atomic64_t *v) in atomic64_sub_return_acquire()
1877 * @i: s64 value to subtract
1886 static __always_inline s64
1887 atomic64_sub_return_release(s64 i, atomic64_t *v) in atomic64_sub_return_release()
1896 * @i: s64 value to subtract
1905 static __always_inline s64
1906 atomic64_sub_return_relaxed(s64 i, atomic64_t *v) in atomic64_sub_return_relaxed()
1914 * @i: s64 value to subtract
1923 static __always_inline s64
1924 atomic64_fetch_sub(s64 i, atomic64_t *v) in atomic64_fetch_sub()
1933 * @i: s64 value to subtract
1942 static __always_inline s64
1943 atomic64_fetch_sub_acquire(s64 i, atomic64_t *v) in atomic64_fetch_sub_acquire()
1951 * @i: s64 value to subtract
1960 static __always_inline s64
1961 atomic64_fetch_sub_release(s64 i, atomic64_t *v) in atomic64_fetch_sub_release()
1970 * @i: s64 value to subtract
1979 static __always_inline s64
1980 atomic64_fetch_sub_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_sub_relaxed()
2013 static __always_inline s64
2031 static __always_inline s64
2048 static __always_inline s64
2066 static __always_inline s64
2083 static __always_inline s64
2101 static __always_inline s64
2118 static __always_inline s64
2136 static __always_inline s64
2170 static __always_inline s64
2188 static __always_inline s64
2205 static __always_inline s64
2223 static __always_inline s64
2240 static __always_inline s64
2258 static __always_inline s64
2275 static __always_inline s64
2293 static __always_inline s64
2302 * @i: s64 value
2312 atomic64_and(s64 i, atomic64_t *v) in atomic64_and()
2320 * @i: s64 value
2329 static __always_inline s64
2330 atomic64_fetch_and(s64 i, atomic64_t *v) in atomic64_fetch_and()
2339 * @i: s64 value
2348 static __always_inline s64
2349 atomic64_fetch_and_acquire(s64 i, atomic64_t *v) in atomic64_fetch_and_acquire()
2357 * @i: s64 value
2366 static __always_inline s64
2367 atomic64_fetch_and_release(s64 i, atomic64_t *v) in atomic64_fetch_and_release()
2376 * @i: s64 value
2385 static __always_inline s64
2386 atomic64_fetch_and_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_and_relaxed()
2394 * @i: s64 value
2404 atomic64_andnot(s64 i, atomic64_t *v) in atomic64_andnot()
2412 * @i: s64 value
2421 static __always_inline s64
2422 atomic64_fetch_andnot(s64 i, atomic64_t *v) in atomic64_fetch_andnot()
2431 * @i: s64 value
2440 static __always_inline s64
2441 atomic64_fetch_andnot_acquire(s64 i, atomic64_t *v) in atomic64_fetch_andnot_acquire()
2449 * @i: s64 value
2458 static __always_inline s64
2459 atomic64_fetch_andnot_release(s64 i, atomic64_t *v) in atomic64_fetch_andnot_release()
2468 * @i: s64 value
2477 static __always_inline s64
2478 atomic64_fetch_andnot_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_andnot_relaxed()
2486 * @i: s64 value
2496 atomic64_or(s64 i, atomic64_t *v) in atomic64_or()
2504 * @i: s64 value
2513 static __always_inline s64
2514 atomic64_fetch_or(s64 i, atomic64_t *v) in atomic64_fetch_or()
2523 * @i: s64 value
2532 static __always_inline s64
2533 atomic64_fetch_or_acquire(s64 i, atomic64_t *v) in atomic64_fetch_or_acquire()
2541 * @i: s64 value
2550 static __always_inline s64
2551 atomic64_fetch_or_release(s64 i, atomic64_t *v) in atomic64_fetch_or_release()
2560 * @i: s64 value
2569 static __always_inline s64
2570 atomic64_fetch_or_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_or_relaxed()
2578 * @i: s64 value
2588 atomic64_xor(s64 i, atomic64_t *v) in atomic64_xor()
2596 * @i: s64 value
2605 static __always_inline s64
2606 atomic64_fetch_xor(s64 i, atomic64_t *v) in atomic64_fetch_xor()
2615 * @i: s64 value
2624 static __always_inline s64
2625 atomic64_fetch_xor_acquire(s64 i, atomic64_t *v) in atomic64_fetch_xor_acquire()
2633 * @i: s64 value
2642 static __always_inline s64
2643 atomic64_fetch_xor_release(s64 i, atomic64_t *v) in atomic64_fetch_xor_release()
2652 * @i: s64 value
2661 static __always_inline s64
2662 atomic64_fetch_xor_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_xor_relaxed()
2671 * @new: s64 value to assign
2679 static __always_inline s64
2680 atomic64_xchg(atomic64_t *v, s64 new) in atomic64_xchg()
2690 * @new: s64 value to assign
2698 static __always_inline s64
2699 atomic64_xchg_acquire(atomic64_t *v, s64 new) in atomic64_xchg_acquire()
2708 * @new: s64 value to assign
2716 static __always_inline s64
2717 atomic64_xchg_release(atomic64_t *v, s64 new) in atomic64_xchg_release()
2727 * @new: s64 value to assign
2735 static __always_inline s64
2736 atomic64_xchg_relaxed(atomic64_t *v, s64 new) in atomic64_xchg_relaxed()
2745 * @old: s64 value to compare with
2746 * @new: s64 value to assign
2754 static __always_inline s64
2755 atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg()
2765 * @old: s64 value to compare with
2766 * @new: s64 value to assign
2774 static __always_inline s64
2775 atomic64_cmpxchg_acquire(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg_acquire()
2784 * @old: s64 value to compare with
2785 * @new: s64 value to assign
2793 static __always_inline s64
2794 atomic64_cmpxchg_release(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg_release()
2804 * @old: s64 value to compare with
2805 * @new: s64 value to assign
2813 static __always_inline s64
2814 atomic64_cmpxchg_relaxed(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg_relaxed()
2823 * @old: pointer to s64 value to compare with
2824 * @new: s64 value to assign
2834 atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg()
2845 * @old: pointer to s64 value to compare with
2846 * @new: s64 value to assign
2856 atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg_acquire()
2866 * @old: pointer to s64 value to compare with
2867 * @new: s64 value to assign
2877 atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg_release()
2888 * @old: pointer to s64 value to compare with
2889 * @new: s64 value to assign
2899 atomic64_try_cmpxchg_relaxed(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg_relaxed()
2908 * @i: s64 value to add
2918 atomic64_sub_and_test(s64 i, atomic64_t *v) in atomic64_sub_and_test()
2963 * @i: s64 value to add
2973 atomic64_add_negative(s64 i, atomic64_t *v) in atomic64_add_negative()
2982 * @i: s64 value to add
2992 atomic64_add_negative_acquire(s64 i, atomic64_t *v) in atomic64_add_negative_acquire()
3000 * @i: s64 value to add
3010 atomic64_add_negative_release(s64 i, atomic64_t *v) in atomic64_add_negative_release()
3019 * @i: s64 value to add
3029 atomic64_add_negative_relaxed(s64 i, atomic64_t *v) in atomic64_add_negative_relaxed()
3038 * @a: s64 value to add
3039 * @u: s64 value to compare with
3047 static __always_inline s64
3048 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u) in atomic64_fetch_add_unless()
3058 * @a: s64 value to add
3059 * @u: s64 value to compare with
3068 atomic64_add_unless(atomic64_t *v, s64 a, s64 u) in atomic64_add_unless()
3139 static __always_inline s64