/Linux-v4.19/include/linux/ |
D | interval_tree_generic.h | 74 bool leftmost = true; \ 85 leftmost = false; \ 92 leftmost, &ITPREFIX ## _augment); \ 152 ITSTRUCT *node, *leftmost; \ 174 leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \ 175 if (ITSTART(leftmost) > last) \
|
D | rbtree_augmented.h | 47 bool newleft, struct rb_node **leftmost, 165 struct rb_node **leftmost, in __rb_erase_augmented() argument 173 if (leftmost && node == *leftmost) in __rb_erase_augmented() 174 *leftmost = rb_next(node); in __rb_erase_augmented()
|
/Linux-v4.19/kernel/locking/ |
D | rtmutex_common.h | 55 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_top_waiter() local 58 if (leftmost) { in rt_mutex_top_waiter() 59 w = rb_entry(leftmost, struct rt_mutex_waiter, tree_entry); in rt_mutex_top_waiter()
|
D | rtmutex.c | 277 bool leftmost = true; in rt_mutex_enqueue() local 286 leftmost = false; in rt_mutex_enqueue() 291 rb_insert_color_cached(&waiter->tree_entry, &lock->waiters, leftmost); in rt_mutex_enqueue() 310 bool leftmost = true; in rt_mutex_enqueue_pi() local 319 leftmost = false; in rt_mutex_enqueue_pi() 324 rb_insert_color_cached(&waiter->pi_tree_entry, &task->pi_waiters, leftmost); in rt_mutex_enqueue_pi()
|
/Linux-v4.19/lib/ |
D | rbtree.c | 98 bool newleft, struct rb_node **leftmost, in __rb_insert() argument 104 *leftmost = node; in __rb_insert() 467 struct rb_root_cached *root, bool leftmost) in rb_insert_color_cached() argument 469 __rb_insert(node, &root->rb_root, leftmost, in rb_insert_color_cached() 492 bool newleft, struct rb_node **leftmost, in __rb_insert_augmented() argument 495 __rb_insert(node, root, newleft, leftmost, augment_rotate); in __rb_insert_augmented()
|
D | rbtree_test.c | 52 bool leftmost = true; in insert_cached() local 60 leftmost = false; in insert_cached() 65 rb_insert_color_cached(&node->rb, root, leftmost); in insert_cached() 131 bool leftmost = true; in insert_augmented_cached() local 142 leftmost = false; in insert_augmented_cached() 149 leftmost, &augment_callbacks); in insert_augmented_cached()
|
/Linux-v4.19/drivers/gpu/drm/ |
D | drm_mm.c | 183 bool leftmost; in drm_mm_interval_tree_add_node() local 200 leftmost = false; in drm_mm_interval_tree_add_node() 204 leftmost = true; in drm_mm_interval_tree_add_node() 216 leftmost = false; in drm_mm_interval_tree_add_node() 221 rb_insert_augmented_cached(&node->rb, &mm->interval_tree, leftmost, in drm_mm_interval_tree_add_node()
|
/Linux-v4.19/Documentation/media/uapi/v4l/ |
D | pixfmt-intro.rst | 29 leftmost pixel of the topmost row. Following that is the pixel 34 leftmost pixel of the second row from the top, and so on. The last row
|
D | pixfmt-v4l2-mplane.rst | 29 - Distance in bytes between the leftmost pixels in two adjacent
|
D | pixfmt-v4l2.rst | 53 - Distance in bytes between the leftmost pixels in two adjacent
|
D | vidioc-g-fbuf.rst | 163 - Distance in bytes between the leftmost pixels in two adjacent
|
/Linux-v4.19/Documentation/scheduler/ |
D | sched-design-CFS.txt | 69 p->se.vruntime key. CFS picks the "leftmost" task from this tree and sticks to it. 72 to become the "leftmost task" and thus get on the CPU within a deterministic 79 becomes the "leftmost task" of the time-ordered rbtree it maintains (plus a 80 small amount of "granularity" distance relative to the leftmost task so that we 81 do not over-schedule tasks and trash the cache), then the new leftmost task is
|
/Linux-v4.19/kernel/sched/ |
D | deadline.c | 452 bool leftmost = true; in enqueue_pushable_dl_task() local 464 leftmost = false; in enqueue_pushable_dl_task() 468 if (leftmost) in enqueue_pushable_dl_task() 473 &dl_rq->pushable_dl_tasks_root, leftmost); in enqueue_pushable_dl_task() 1326 struct rb_node *leftmost = dl_rq->root.rb_leftmost; in dec_dl_deadline() local 1329 entry = rb_entry(leftmost, struct sched_dl_entity, rb_node); in dec_dl_deadline() 1376 int leftmost = 1; in __enqueue_dl_entity() local 1387 leftmost = 0; in __enqueue_dl_entity() 1392 rb_insert_color_cached(&dl_se->rb_node, &dl_rq->root, leftmost); in __enqueue_dl_entity()
|
D | fair.c | 499 struct rb_node *leftmost = rb_first_cached(&cfs_rq->tasks_timeline); in update_min_vruntime() local 510 if (leftmost) { /* non-empty tree */ in update_min_vruntime() 512 se = rb_entry(leftmost, struct sched_entity, run_node); in update_min_vruntime() 536 bool leftmost = true; in __enqueue_entity() local 552 leftmost = false; in __enqueue_entity() 558 &cfs_rq->tasks_timeline, leftmost); in __enqueue_entity()
|
/Linux-v4.19/Documentation/ |
D | rbtree.txt | 199 Computing the leftmost (smallest) node is quite a common task for binary 212 leftmost node. This allows rb_root_cached to exist wherever rb_root does, 319 * Iterate to find the leftmost such node N. 331 return node; /* node is leftmost match */
|
/Linux-v4.19/fs/ |
D | eventpoll.c | 1265 bool leftmost = true; in ep_rbtree_insert() local 1273 leftmost = false; in ep_rbtree_insert() 1278 rb_insert_color_cached(&epi->rbn, &ep->rbr, leftmost); in ep_rbtree_insert()
|
/Linux-v4.19/block/ |
D | cfq-iosched.c | 1247 bool leftmost = true, rightmost = true; in __cfq_group_service_tree_add() local 1258 leftmost = false; in __cfq_group_service_tree_add() 1266 rb_insert_color_cached(&cfqg->rb_node, &st->rb, leftmost); in __cfq_group_service_tree_add() 2209 bool leftmost = true; in cfq_service_tree_add() local 2264 leftmost = false; in cfq_service_tree_add() 2270 rb_insert_color_cached(&cfqq->rb_node, &st->rb, leftmost); in cfq_service_tree_add()
|
/Linux-v4.19/Documentation/devicetree/ |
D | booting-without-of.txt | 840 is 31. In the case of node names, this is only the leftmost part of
|