Lines Matching full:reorder

233  *   serialization, if present in one of the percpu reorder queues.
236 * the cpu's reorder queue.
242 struct padata_list *reorder; in padata_find_next() local
245 reorder = per_cpu_ptr(pd->reorder_list, cpu); in padata_find_next()
247 spin_lock(&reorder->lock); in padata_find_next()
248 if (list_empty(&reorder->list)) { in padata_find_next()
249 spin_unlock(&reorder->lock); in padata_find_next()
253 padata = list_entry(reorder->list.next, struct padata_priv, list); in padata_find_next()
260 spin_unlock(&reorder->lock); in padata_find_next()
270 spin_unlock(&reorder->lock); in padata_find_next()
280 struct padata_list *reorder; in padata_reorder() local
284 * the reorder queue the time. Calculating in which percpu reorder in padata_reorder()
287 * the objects arrive to the reorder queues. So a cpu could wait to in padata_reorder()
301 * cpu's reorder queue, nothing to do for now. in padata_reorder()
320 * the reorder queues in the meantime. in padata_reorder()
322 * Ensure reorder queue is read after pd->lock is dropped so we see in padata_reorder()
328 reorder = per_cpu_ptr(pd->reorder_list, pd->cpu); in padata_reorder()
329 if (!list_empty(&reorder->list) && padata_find_next(pd, false)) in padata_reorder()
389 struct padata_list *reorder = per_cpu_ptr(pd->reorder_list, hashed_cpu); in padata_do_serial() local
392 spin_lock(&reorder->lock); in padata_do_serial()
394 list_for_each_entry_reverse(cur, &reorder->list, list) in padata_do_serial()
398 spin_unlock(&reorder->lock); in padata_do_serial()
401 * Ensure the addition to the reorder list is ordered correctly in padata_do_serial()
540 /* Initialize per-CPU reorder lists */