Lines Matching refs:s64
2553 static __always_inline s64
2569 static __always_inline s64
2575 s64 ret; in raw_atomic64_read_acquire()
2600 raw_atomic64_set(atomic64_t *v, s64 i) in raw_atomic64_set()
2617 raw_atomic64_set_release(atomic64_t *v, s64 i) in raw_atomic64_set_release()
2643 raw_atomic64_add(s64 i, atomic64_t *v) in raw_atomic64_add()
2659 static __always_inline s64
2660 raw_atomic64_add_return(s64 i, atomic64_t *v) in raw_atomic64_add_return()
2665 s64 ret; in raw_atomic64_add_return()
2686 static __always_inline s64
2687 raw_atomic64_add_return_acquire(s64 i, atomic64_t *v) in raw_atomic64_add_return_acquire()
2692 s64 ret = arch_atomic64_add_return_relaxed(i, v); in raw_atomic64_add_return_acquire()
2713 static __always_inline s64
2714 raw_atomic64_add_return_release(s64 i, atomic64_t *v) in raw_atomic64_add_return_release()
2739 static __always_inline s64
2740 raw_atomic64_add_return_relaxed(s64 i, atomic64_t *v) in raw_atomic64_add_return_relaxed()
2762 static __always_inline s64
2763 raw_atomic64_fetch_add(s64 i, atomic64_t *v) in raw_atomic64_fetch_add()
2768 s64 ret; in raw_atomic64_fetch_add()
2789 static __always_inline s64
2790 raw_atomic64_fetch_add_acquire(s64 i, atomic64_t *v) in raw_atomic64_fetch_add_acquire()
2795 s64 ret = arch_atomic64_fetch_add_relaxed(i, v); in raw_atomic64_fetch_add_acquire()
2816 static __always_inline s64
2817 raw_atomic64_fetch_add_release(s64 i, atomic64_t *v) in raw_atomic64_fetch_add_release()
2842 static __always_inline s64
2843 raw_atomic64_fetch_add_relaxed(s64 i, atomic64_t *v) in raw_atomic64_fetch_add_relaxed()
2866 raw_atomic64_sub(s64 i, atomic64_t *v) in raw_atomic64_sub()
2882 static __always_inline s64
2883 raw_atomic64_sub_return(s64 i, atomic64_t *v) in raw_atomic64_sub_return()
2888 s64 ret; in raw_atomic64_sub_return()
2909 static __always_inline s64
2910 raw_atomic64_sub_return_acquire(s64 i, atomic64_t *v) in raw_atomic64_sub_return_acquire()
2915 s64 ret = arch_atomic64_sub_return_relaxed(i, v); in raw_atomic64_sub_return_acquire()
2936 static __always_inline s64
2937 raw_atomic64_sub_return_release(s64 i, atomic64_t *v) in raw_atomic64_sub_return_release()
2962 static __always_inline s64
2963 raw_atomic64_sub_return_relaxed(s64 i, atomic64_t *v) in raw_atomic64_sub_return_relaxed()
2985 static __always_inline s64
2986 raw_atomic64_fetch_sub(s64 i, atomic64_t *v) in raw_atomic64_fetch_sub()
2991 s64 ret; in raw_atomic64_fetch_sub()
3012 static __always_inline s64
3013 raw_atomic64_fetch_sub_acquire(s64 i, atomic64_t *v) in raw_atomic64_fetch_sub_acquire()
3018 s64 ret = arch_atomic64_fetch_sub_relaxed(i, v); in raw_atomic64_fetch_sub_acquire()
3039 static __always_inline s64
3040 raw_atomic64_fetch_sub_release(s64 i, atomic64_t *v) in raw_atomic64_fetch_sub_release()
3065 static __always_inline s64
3066 raw_atomic64_fetch_sub_relaxed(s64 i, atomic64_t *v) in raw_atomic64_fetch_sub_relaxed()
3107 static __always_inline s64
3113 s64 ret; in raw_atomic64_inc_return()
3133 static __always_inline s64
3139 s64 ret = arch_atomic64_inc_return_relaxed(v); in raw_atomic64_inc_return_acquire()
3159 static __always_inline s64
3184 static __always_inline s64
3206 static __always_inline s64
3212 s64 ret; in raw_atomic64_fetch_inc()
3232 static __always_inline s64
3238 s64 ret = arch_atomic64_fetch_inc_relaxed(v); in raw_atomic64_fetch_inc_acquire()
3258 static __always_inline s64
3283 static __always_inline s64
3325 static __always_inline s64
3331 s64 ret; in raw_atomic64_dec_return()
3351 static __always_inline s64
3357 s64 ret = arch_atomic64_dec_return_relaxed(v); in raw_atomic64_dec_return_acquire()
3377 static __always_inline s64
3402 static __always_inline s64
3424 static __always_inline s64
3430 s64 ret; in raw_atomic64_fetch_dec()
3450 static __always_inline s64
3456 s64 ret = arch_atomic64_fetch_dec_relaxed(v); in raw_atomic64_fetch_dec_acquire()
3476 static __always_inline s64
3501 static __always_inline s64
3525 raw_atomic64_and(s64 i, atomic64_t *v) in raw_atomic64_and()
3541 static __always_inline s64
3542 raw_atomic64_fetch_and(s64 i, atomic64_t *v) in raw_atomic64_fetch_and()
3547 s64 ret; in raw_atomic64_fetch_and()
3568 static __always_inline s64
3569 raw_atomic64_fetch_and_acquire(s64 i, atomic64_t *v) in raw_atomic64_fetch_and_acquire()
3574 s64 ret = arch_atomic64_fetch_and_relaxed(i, v); in raw_atomic64_fetch_and_acquire()
3595 static __always_inline s64
3596 raw_atomic64_fetch_and_release(s64 i, atomic64_t *v) in raw_atomic64_fetch_and_release()
3621 static __always_inline s64
3622 raw_atomic64_fetch_and_relaxed(s64 i, atomic64_t *v) in raw_atomic64_fetch_and_relaxed()
3645 raw_atomic64_andnot(s64 i, atomic64_t *v) in raw_atomic64_andnot()
3665 static __always_inline s64
3666 raw_atomic64_fetch_andnot(s64 i, atomic64_t *v) in raw_atomic64_fetch_andnot()
3671 s64 ret; in raw_atomic64_fetch_andnot()
3692 static __always_inline s64
3693 raw_atomic64_fetch_andnot_acquire(s64 i, atomic64_t *v) in raw_atomic64_fetch_andnot_acquire()
3698 s64 ret = arch_atomic64_fetch_andnot_relaxed(i, v); in raw_atomic64_fetch_andnot_acquire()
3719 static __always_inline s64
3720 raw_atomic64_fetch_andnot_release(s64 i, atomic64_t *v) in raw_atomic64_fetch_andnot_release()
3745 static __always_inline s64
3746 raw_atomic64_fetch_andnot_relaxed(s64 i, atomic64_t *v) in raw_atomic64_fetch_andnot_relaxed()
3769 raw_atomic64_or(s64 i, atomic64_t *v) in raw_atomic64_or()
3785 static __always_inline s64
3786 raw_atomic64_fetch_or(s64 i, atomic64_t *v) in raw_atomic64_fetch_or()
3791 s64 ret; in raw_atomic64_fetch_or()
3812 static __always_inline s64
3813 raw_atomic64_fetch_or_acquire(s64 i, atomic64_t *v) in raw_atomic64_fetch_or_acquire()
3818 s64 ret = arch_atomic64_fetch_or_relaxed(i, v); in raw_atomic64_fetch_or_acquire()
3839 static __always_inline s64
3840 raw_atomic64_fetch_or_release(s64 i, atomic64_t *v) in raw_atomic64_fetch_or_release()
3865 static __always_inline s64
3866 raw_atomic64_fetch_or_relaxed(s64 i, atomic64_t *v) in raw_atomic64_fetch_or_relaxed()
3889 raw_atomic64_xor(s64 i, atomic64_t *v) in raw_atomic64_xor()
3905 static __always_inline s64
3906 raw_atomic64_fetch_xor(s64 i, atomic64_t *v) in raw_atomic64_fetch_xor()
3911 s64 ret; in raw_atomic64_fetch_xor()
3932 static __always_inline s64
3933 raw_atomic64_fetch_xor_acquire(s64 i, atomic64_t *v) in raw_atomic64_fetch_xor_acquire()
3938 s64 ret = arch_atomic64_fetch_xor_relaxed(i, v); in raw_atomic64_fetch_xor_acquire()
3959 static __always_inline s64
3960 raw_atomic64_fetch_xor_release(s64 i, atomic64_t *v) in raw_atomic64_fetch_xor_release()
3985 static __always_inline s64
3986 raw_atomic64_fetch_xor_relaxed(s64 i, atomic64_t *v) in raw_atomic64_fetch_xor_relaxed()
4008 static __always_inline s64
4009 raw_atomic64_xchg(atomic64_t *v, s64 new) in raw_atomic64_xchg()
4014 s64 ret; in raw_atomic64_xchg()
4035 static __always_inline s64
4036 raw_atomic64_xchg_acquire(atomic64_t *v, s64 new) in raw_atomic64_xchg_acquire()
4041 s64 ret = arch_atomic64_xchg_relaxed(v, new); in raw_atomic64_xchg_acquire()
4062 static __always_inline s64
4063 raw_atomic64_xchg_release(atomic64_t *v, s64 new) in raw_atomic64_xchg_release()
4088 static __always_inline s64
4089 raw_atomic64_xchg_relaxed(atomic64_t *v, s64 new) in raw_atomic64_xchg_relaxed()
4112 static __always_inline s64
4113 raw_atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) in raw_atomic64_cmpxchg()
4118 s64 ret; in raw_atomic64_cmpxchg()
4140 static __always_inline s64
4141 raw_atomic64_cmpxchg_acquire(atomic64_t *v, s64 old, s64 new) in raw_atomic64_cmpxchg_acquire()
4146 s64 ret = arch_atomic64_cmpxchg_relaxed(v, old, new); in raw_atomic64_cmpxchg_acquire()
4168 static __always_inline s64
4169 raw_atomic64_cmpxchg_release(atomic64_t *v, s64 old, s64 new) in raw_atomic64_cmpxchg_release()
4195 static __always_inline s64
4196 raw_atomic64_cmpxchg_relaxed(atomic64_t *v, s64 old, s64 new) in raw_atomic64_cmpxchg_relaxed()
4221 raw_atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) in raw_atomic64_try_cmpxchg()
4232 s64 r, o = *old; in raw_atomic64_try_cmpxchg()
4254 raw_atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new) in raw_atomic64_try_cmpxchg_acquire()
4265 s64 r, o = *old; in raw_atomic64_try_cmpxchg_acquire()
4287 raw_atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new) in raw_atomic64_try_cmpxchg_release()
4297 s64 r, o = *old; in raw_atomic64_try_cmpxchg_release()
4319 raw_atomic64_try_cmpxchg_relaxed(atomic64_t *v, s64 *old, s64 new) in raw_atomic64_try_cmpxchg_relaxed()
4326 s64 r, o = *old; in raw_atomic64_try_cmpxchg_relaxed()
4346 raw_atomic64_sub_and_test(s64 i, atomic64_t *v) in raw_atomic64_sub_and_test()
4407 raw_atomic64_add_negative(s64 i, atomic64_t *v) in raw_atomic64_add_negative()
4434 raw_atomic64_add_negative_acquire(s64 i, atomic64_t *v) in raw_atomic64_add_negative_acquire()
4461 raw_atomic64_add_negative_release(s64 i, atomic64_t *v) in raw_atomic64_add_negative_release()
4487 raw_atomic64_add_negative_relaxed(s64 i, atomic64_t *v) in raw_atomic64_add_negative_relaxed()
4510 static __always_inline s64
4511 raw_atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u) in raw_atomic64_fetch_add_unless()
4516 s64 c = raw_atomic64_read(v); in raw_atomic64_fetch_add_unless()
4540 raw_atomic64_add_unless(atomic64_t *v, s64 a, s64 u) in raw_atomic64_add_unless()
4585 s64 c = raw_atomic64_read(v); in raw_atomic64_inc_unless_negative()
4612 s64 c = raw_atomic64_read(v); in raw_atomic64_dec_unless_positive()
4633 static __always_inline s64
4639 s64 dec, c = raw_atomic64_read(v); in raw_atomic64_dec_if_positive()