Searched refs:workers (Results 1 – 11 of 11) sorted by relevance
/Linux-v4.19/Documentation/core-api/ |
D | workqueue.rst | 34 number of workers as the number of CPUs. The kernel grew a lot of MT 118 number of the currently runnable workers. Generally, work items are 122 workers on the CPU, the worker-pool doesn't start execution of a new 125 are pending work items. This allows using a minimal number of workers 128 Keeping idle workers around doesn't cost other than the memory space 140 Forward progress guarantee relies on that workers can be created when 142 through the use of rescue workers. All work items which might be used 169 worker-pools which host workers which are not bound to any 178 of mostly unused workers across different CPUs as the issuer 200 each other. Each maintains its separate pool of workers and [all …]
|
/Linux-v4.19/Documentation/filesystems/pohmelfs/ |
D | info.txt | 93 -w workers - number of workers per connected client. Default: 1. 98 Number of worker threads specifies how many workers will be created for each client.
|
/Linux-v4.19/drivers/md/ |
D | raid5.h | 523 struct r5worker *workers; member
|
D | raid5.c | 205 group->workers[0].working = true; in raid5_wakeup_stripe_thread() 207 queue_work_on(sh->cpu, raid5_wq, &group->workers[0].work); in raid5_wakeup_stripe_thread() 212 if (group->workers[i].working == false) { in raid5_wakeup_stripe_thread() 213 group->workers[i].working = true; in raid5_wakeup_stripe_thread() 215 &group->workers[i].work); in raid5_wakeup_stripe_thread() 6624 kfree(old_groups[0].workers); in raid5_store_group_thread_cnt() 6661 struct r5worker *workers; in alloc_thread_groups() local 6671 workers = kcalloc(size, *group_cnt, GFP_NOIO); in alloc_thread_groups() 6674 if (!*worker_groups || !workers) { in alloc_thread_groups() 6675 kfree(workers); in alloc_thread_groups() [all …]
|
/Linux-v4.19/net/l2tp/ |
D | Kconfig | 22 with home workers to connect to their offices.
|
/Linux-v4.19/kernel/ |
D | workqueue.c | 169 struct list_head workers; /* A: attached workers */ member 408 list_for_each_entry((worker), &(pool)->workers, node) \ 1743 list_add_tail(&worker->node, &pool->workers); in worker_attach_to_pool() 1767 if (list_empty(&pool->workers)) in worker_detach_from_pool() 3304 INIT_LIST_HEAD(&pool->workers); in init_worker_pool() 3387 if (!list_empty(&pool->workers)) in put_unbound_pool()
|
/Linux-v4.19/drivers/block/mtip32xx/ |
D | mtip32xx.c | 788 int do_irq_enable = 1, i, workers; in mtip_handle_irq() local 809 for (i = 0, workers = 0; i < MTIP_MAX_SLOT_GROUPS; in mtip_handle_irq() 814 workers++; in mtip_handle_irq() 817 atomic_set(&dd->irq_workers_active, workers); in mtip_handle_irq() 818 if (workers) { in mtip_handle_irq()
|
/Linux-v4.19/fs/btrfs/ |
D | disk-io.c | 813 btrfs_queue_work(fs_info->workers, &async->work); in btrfs_wq_submit_bio() 2023 btrfs_destroy_workqueue(fs_info->workers); in btrfs_stop_all_workers() 2184 fs_info->workers = in btrfs_init_workqueues() 2250 if (!(fs_info->workers && fs_info->delalloc_workers && in btrfs_init_workqueues()
|
D | super.c | 1716 btrfs_workqueue_set_max(fs_info->workers, new_pool_size); in btrfs_resize_thread_pool()
|
D | ctree.h | 935 struct btrfs_workqueue *workers; member
|
/Linux-v4.19/Documentation/ |
D | kernel-per-CPU-kthreads.txt | 262 d. As of v3.18, Christoph Lameter's on-demand vmstat workers
|