Home
last modified time | relevance | path

Searched refs:workers (Results 1 – 11 of 11) sorted by relevance

/Linux-v4.19/Documentation/core-api/
Dworkqueue.rst34 number of workers as the number of CPUs. The kernel grew a lot of MT
118 number of the currently runnable workers. Generally, work items are
122 workers on the CPU, the worker-pool doesn't start execution of a new
125 are pending work items. This allows using a minimal number of workers
128 Keeping idle workers around doesn't cost other than the memory space
140 Forward progress guarantee relies on that workers can be created when
142 through the use of rescue workers. All work items which might be used
169 worker-pools which host workers which are not bound to any
178 of mostly unused workers across different CPUs as the issuer
200 each other. Each maintains its separate pool of workers and
[all …]
/Linux-v4.19/Documentation/filesystems/pohmelfs/
Dinfo.txt93 -w workers - number of workers per connected client. Default: 1.
98 Number of worker threads specifies how many workers will be created for each client.
/Linux-v4.19/drivers/md/
Draid5.h523 struct r5worker *workers; member
Draid5.c205 group->workers[0].working = true; in raid5_wakeup_stripe_thread()
207 queue_work_on(sh->cpu, raid5_wq, &group->workers[0].work); in raid5_wakeup_stripe_thread()
212 if (group->workers[i].working == false) { in raid5_wakeup_stripe_thread()
213 group->workers[i].working = true; in raid5_wakeup_stripe_thread()
215 &group->workers[i].work); in raid5_wakeup_stripe_thread()
6624 kfree(old_groups[0].workers); in raid5_store_group_thread_cnt()
6661 struct r5worker *workers; in alloc_thread_groups() local
6671 workers = kcalloc(size, *group_cnt, GFP_NOIO); in alloc_thread_groups()
6674 if (!*worker_groups || !workers) { in alloc_thread_groups()
6675 kfree(workers); in alloc_thread_groups()
[all …]
/Linux-v4.19/net/l2tp/
DKconfig22 with home workers to connect to their offices.
/Linux-v4.19/kernel/
Dworkqueue.c169 struct list_head workers; /* A: attached workers */ member
408 list_for_each_entry((worker), &(pool)->workers, node) \
1743 list_add_tail(&worker->node, &pool->workers); in worker_attach_to_pool()
1767 if (list_empty(&pool->workers)) in worker_detach_from_pool()
3304 INIT_LIST_HEAD(&pool->workers); in init_worker_pool()
3387 if (!list_empty(&pool->workers)) in put_unbound_pool()
/Linux-v4.19/drivers/block/mtip32xx/
Dmtip32xx.c788 int do_irq_enable = 1, i, workers; in mtip_handle_irq() local
809 for (i = 0, workers = 0; i < MTIP_MAX_SLOT_GROUPS; in mtip_handle_irq()
814 workers++; in mtip_handle_irq()
817 atomic_set(&dd->irq_workers_active, workers); in mtip_handle_irq()
818 if (workers) { in mtip_handle_irq()
/Linux-v4.19/fs/btrfs/
Ddisk-io.c813 btrfs_queue_work(fs_info->workers, &async->work); in btrfs_wq_submit_bio()
2023 btrfs_destroy_workqueue(fs_info->workers); in btrfs_stop_all_workers()
2184 fs_info->workers = in btrfs_init_workqueues()
2250 if (!(fs_info->workers && fs_info->delalloc_workers && in btrfs_init_workqueues()
Dsuper.c1716 btrfs_workqueue_set_max(fs_info->workers, new_pool_size); in btrfs_resize_thread_pool()
Dctree.h935 struct btrfs_workqueue *workers; member
/Linux-v4.19/Documentation/
Dkernel-per-CPU-kthreads.txt262 d. As of v3.18, Christoph Lameter's on-demand vmstat workers