Lines Matching refs:mb
29 #define ATOMIC_FETCH_OP(name, mb, op, asm_op, cl...) \ in ATOMIC_OP() argument
33 " " #asm_op #mb " %w[i], %w[i], %[v]" \ in ATOMIC_OP()
55 #define ATOMIC_OP_ADD_RETURN(name, mb, cl...) \ argument
61 " ldadd" #mb " %w[i], %w[tmp], %[v]\n" \
86 #define ATOMIC_FETCH_OP_AND(name, mb, cl...) \ argument
91 " ldclr" #mb " %w[i], %w[i], %[v]" \
115 #define ATOMIC_OP_SUB_RETURN(name, mb, cl...) \ argument
122 " ldadd" #mb " %w[i], %w[tmp], %[v]\n" \
138 #define ATOMIC_FETCH_OP_SUB(name, mb, cl...) \ argument
143 " ldadd" #mb " %w[i], %w[i], %[v]" \
174 #define ATOMIC64_FETCH_OP(name, mb, op, asm_op, cl...) \ in ATOMIC64_OP() argument
178 " " #asm_op #mb " %[i], %[i], %[v]" \ in ATOMIC64_OP()
200 #define ATOMIC64_OP_ADD_RETURN(name, mb, cl...) \ argument
206 " ldadd" #mb " %[i], %x[tmp], %[v]\n" \
231 #define ATOMIC64_FETCH_OP_AND(name, mb, cl...) \ argument
236 " ldclr" #mb " %[i], %[i], %[v]" \
260 #define ATOMIC64_OP_SUB_RETURN(name, mb, cl...) \ argument
267 " ldadd" #mb " %[i], %x[tmp], %[v]\n" \
283 #define ATOMIC64_FETCH_OP_SUB(name, mb, cl...) \ argument
288 " ldadd" #mb " %[i], %[i], %[v]" \
323 #define __CMPXCHG_CASE(w, sfx, name, sz, mb, cl...) \ argument
336 " cas" #mb #sfx "\t%" #w "[tmp], %" #w "[new], %[v]\n" \
365 #define __CMPXCHG_DBL(name, mb, cl...) \ argument
382 " casp" #mb "\t%[old1], %[old2], %[new1], %[new2], %[v]\n"\