Lines Matching refs:this_rq
96 extern void calc_global_load_tick(struct rq *this_rq);
97 extern long calc_load_fold_active(struct rq *this_rq, long adjust);
1046 #define this_rq() this_cpu_ptr(&runqueues) macro
1258 rq = this_rq(); in this_rq_lock_irq()
1462 extern int newidle_balance(struct rq *this_rq, struct rq_flags *rf);
1468 static inline int newidle_balance(struct rq *this_rq, struct rq_flags *rf) { return 0; } in newidle_balance() argument
1738 void (*task_woken)(struct rq *this_rq, struct task_struct *task);
1756 void (*switched_from)(struct rq *this_rq, struct task_struct *task);
1757 void (*switched_to) (struct rq *this_rq, struct task_struct *task);
1758 void (*prio_changed) (struct rq *this_rq, struct task_struct *task,
2001 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument
2002 __releases(this_rq->lock) in _double_lock_balance()
2004 __acquires(this_rq->lock) in _double_lock_balance()
2006 raw_spin_unlock(&this_rq->lock); in _double_lock_balance()
2007 double_rq_lock(this_rq, busiest); in _double_lock_balance()
2020 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument
2021 __releases(this_rq->lock) in _double_lock_balance()
2023 __acquires(this_rq->lock) in _double_lock_balance()
2028 if (busiest < this_rq) { in _double_lock_balance()
2029 raw_spin_unlock(&this_rq->lock); in _double_lock_balance()
2031 raw_spin_lock_nested(&this_rq->lock, in _double_lock_balance()
2046 static inline int double_lock_balance(struct rq *this_rq, struct rq *busiest) in double_lock_balance() argument
2050 raw_spin_unlock(&this_rq->lock); in double_lock_balance()
2054 return _double_lock_balance(this_rq, busiest); in double_lock_balance()
2057 static inline void double_unlock_balance(struct rq *this_rq, struct rq *busiest) in double_unlock_balance() argument
2061 lock_set_subclass(&this_rq->lock.dep_map, 0, _RET_IP_); in double_unlock_balance()