Lines Matching refs:counter
32 __asm__ __volatile__("lwz %0,0(%1)" : "=r"(t) : "b"(&v->counter)); in arch_atomic_read()
34 __asm__ __volatile__("lwz%U1%X1 %0,%1" : "=r"(t) : "m<>"(v->counter)); in arch_atomic_read()
43 __asm__ __volatile__("stw %1,0(%2)" : "=m"(v->counter) : "r"(i), "b"(&v->counter)); in arch_atomic_set()
45 __asm__ __volatile__("stw%U0%X0 %1,%0" : "=m<>"(v->counter) : "r"(i)); in arch_atomic_set()
58 : "=&r" (t), "+m" (v->counter) \
59 : "r"#sign (a), "r" (&v->counter) \
73 : "=&r" (t), "+m" (v->counter) \
74 : "r"#sign (a), "r" (&v->counter) \
90 : "=&r" (res), "=&r" (t), "+m" (v->counter) \
91 : "r"#sign (a), "r" (&v->counter) \
154 : "r" (&v->counter), "rI" (a), "r" (u) in arch_atomic_fetch_add_unless()
181 : "r" (&v->counter) in arch_atomic_dec_if_positive()
198 __asm__ __volatile__("ld %0,0(%1)" : "=r"(t) : "b"(&v->counter)); in arch_atomic64_read()
200 __asm__ __volatile__("ld%U1%X1 %0,%1" : "=r"(t) : "m<>"(v->counter)); in arch_atomic64_read()
209 __asm__ __volatile__("std %1,0(%2)" : "=m"(v->counter) : "r"(i), "b"(&v->counter)); in arch_atomic64_set()
211 __asm__ __volatile__("std%U0%X0 %1,%0" : "=m<>"(v->counter) : "r"(i)); in arch_atomic64_set()
224 : "=&r" (t), "+m" (v->counter) \
225 : "r" (a), "r" (&v->counter) \
240 : "=&r" (t), "+m" (v->counter) \
241 : "r" (a), "r" (&v->counter) \
258 : "=&r" (res), "=&r" (t), "+m" (v->counter) \
259 : "r" (a), "r" (&v->counter) \
306 : "=&r" (t), "+m" (v->counter) in ATOMIC64_OPS()
307 : "r" (&v->counter) in ATOMIC64_OPS()
321 : "=&r" (t), "+m" (v->counter) in arch_atomic64_inc_return_relaxed()
322 : "r" (&v->counter) in arch_atomic64_inc_return_relaxed()
337 : "=&r" (t), "+m" (v->counter) in arch_atomic64_dec()
338 : "r" (&v->counter) in arch_atomic64_dec()
352 : "=&r" (t), "+m" (v->counter) in arch_atomic64_dec_return_relaxed()
353 : "r" (&v->counter) in arch_atomic64_dec_return_relaxed()
380 : "r" (&v->counter) in arch_atomic64_dec_if_positive()
412 : "r" (&v->counter), "r" (a), "r" (u) in arch_atomic64_fetch_add_unless()
442 : "r" (&v->counter) in arch_atomic64_inc_not_zero()