Searched refs:workers (Results 1 – 11 of 11) sorted by relevance
/Linux-v5.10/Documentation/core-api/ |
D | workqueue.rst | 34 number of workers as the number of CPUs. The kernel grew a lot of MT 118 number of the currently runnable workers. Generally, work items are 122 workers on the CPU, the worker-pool doesn't start execution of a new 125 are pending work items. This allows using a minimal number of workers 128 Keeping idle workers around doesn't cost other than the memory space 140 Forward progress guarantee relies on that workers can be created when 142 through the use of rescue workers. All work items which might be used 169 worker-pools which host workers which are not bound to any 178 of mostly unused workers across different CPUs as the issuer 200 each other. Each maintains its separate pool of workers and [all …]
|
/Linux-v5.10/net/l2tp/ |
D | Kconfig | 23 with home workers to connect to their offices.
|
/Linux-v5.10/drivers/md/ |
D | raid5.h | 516 struct r5worker *workers; member
|
D | raid5.c | 196 group->workers[0].working = true; in raid5_wakeup_stripe_thread() 198 queue_work_on(sh->cpu, raid5_wq, &group->workers[0].work); in raid5_wakeup_stripe_thread() 203 if (group->workers[i].working == false) { in raid5_wakeup_stripe_thread() 204 group->workers[i].working = true; in raid5_wakeup_stripe_thread() 206 &group->workers[i].work); in raid5_wakeup_stripe_thread() 6924 kfree(old_groups[0].workers); in raid5_store_group_thread_cnt() 6961 struct r5worker *workers; in alloc_thread_groups() local 6970 workers = kcalloc(size, *group_cnt, GFP_NOIO); in alloc_thread_groups() 6973 if (!*worker_groups || !workers) { in alloc_thread_groups() 6974 kfree(workers); in alloc_thread_groups() [all …]
|
/Linux-v5.10/kernel/ |
D | workqueue.c | 170 struct list_head workers; /* A: attached workers */ member 408 list_for_each_entry((worker), &(pool)->workers, node) \ 1862 list_add_tail(&worker->node, &pool->workers); in worker_attach_to_pool() 1886 if (list_empty(&pool->workers)) in worker_detach_from_pool() 3435 INIT_LIST_HEAD(&pool->workers); in init_worker_pool() 3570 if (!list_empty(&pool->workers)) in put_unbound_pool()
|
/Linux-v5.10/drivers/block/mtip32xx/ |
D | mtip32xx.c | 745 int do_irq_enable = 1, i, workers; in mtip_handle_irq() local 766 for (i = 0, workers = 0; i < MTIP_MAX_SLOT_GROUPS; in mtip_handle_irq() 771 workers++; in mtip_handle_irq() 774 atomic_set(&dd->irq_workers_active, workers); in mtip_handle_irq() 775 if (workers) { in mtip_handle_irq()
|
/Linux-v5.10/Documentation/dev-tools/ |
D | kcov.rst | 220 some kernel interface (e.g. vhost workers); as well as from soft
|
/Linux-v5.10/fs/btrfs/ |
D | disk-io.c | 774 btrfs_queue_work(fs_info->workers, &async->work); in btrfs_wq_submit_bio() 1986 btrfs_destroy_workqueue(fs_info->workers); in btrfs_stop_all_workers() 2157 fs_info->workers = in btrfs_init_workqueues() 2209 if (!(fs_info->workers && fs_info->delalloc_workers && in btrfs_init_workqueues()
|
D | super.c | 1781 btrfs_workqueue_set_max(fs_info->workers, new_pool_size); in btrfs_resize_thread_pool()
|
D | ctree.h | 760 struct btrfs_workqueue *workers; member
|
/Linux-v5.10/Documentation/admin-guide/ |
D | kernel-per-CPU-kthreads.rst | 262 d. As of v3.18, Christoph Lameter's on-demand vmstat workers
|