/Linux-v4.19/virt/kvm/ |
D | async_pf.c | 33 struct kvm_async_pf *work) in kvm_async_page_present_sync() argument 36 kvm_arch_async_page_present(vcpu, work); in kvm_async_page_present_sync() 40 struct kvm_async_pf *work) in kvm_async_page_present_async() argument 43 kvm_arch_async_page_present(vcpu, work); in kvm_async_page_present_async() 72 static void async_pf_execute(struct work_struct *work) in async_pf_execute() argument 75 container_of(work, struct kvm_async_pf, work); in async_pf_execute() 122 struct kvm_async_pf *work = in kvm_clear_async_pf_completion_queue() local 124 typeof(*work), queue); in kvm_clear_async_pf_completion_queue() 125 list_del(&work->queue); in kvm_clear_async_pf_completion_queue() 131 if (!work->vcpu) in kvm_clear_async_pf_completion_queue() [all …]
|
/Linux-v4.19/drivers/gpu/drm/ |
D | drm_flip_work.c | 54 void drm_flip_work_queue_task(struct drm_flip_work *work, in drm_flip_work_queue_task() argument 59 spin_lock_irqsave(&work->lock, flags); in drm_flip_work_queue_task() 60 list_add_tail(&task->node, &work->queued); in drm_flip_work_queue_task() 61 spin_unlock_irqrestore(&work->lock, flags); in drm_flip_work_queue_task() 73 void drm_flip_work_queue(struct drm_flip_work *work, void *val) in drm_flip_work_queue() argument 80 drm_flip_work_queue_task(work, task); in drm_flip_work_queue() 82 DRM_ERROR("%s could not allocate task!\n", work->name); in drm_flip_work_queue() 83 work->func(work, val); in drm_flip_work_queue() 98 void drm_flip_work_commit(struct drm_flip_work *work, in drm_flip_work_commit() argument 103 spin_lock_irqsave(&work->lock, flags); in drm_flip_work_commit() [all …]
|
/Linux-v4.19/include/trace/events/ |
D | workqueue.h | 13 TP_PROTO(struct work_struct *work), 15 TP_ARGS(work), 18 __field( void *, work ) 22 __entry->work = work; 25 TP_printk("work struct %p", __entry->work) 43 struct work_struct *work), 45 TP_ARGS(req_cpu, pwq, work), 48 __field( void *, work ) 56 __entry->work = work; 57 __entry->function = work->func; [all …]
|
/Linux-v4.19/kernel/ |
D | task_work.c | 28 task_work_add(struct task_struct *task, struct callback_head *work, bool notify) in task_work_add() argument 36 work->next = head; in task_work_add() 37 } while (cmpxchg(&task->task_works, head, work) != head); in task_work_add() 59 struct callback_head *work; in task_work_cancel() local 71 while ((work = READ_ONCE(*pprev))) { in task_work_cancel() 72 if (work->func != func) in task_work_cancel() 73 pprev = &work->next; in task_work_cancel() 74 else if (cmpxchg(pprev, work, work->next) == work) in task_work_cancel() 79 return work; in task_work_cancel() 93 struct callback_head *work, *head, *next; in task_work_run() local [all …]
|
D | kthread.c | 627 struct kthread_work *work; in kthread_worker_fn() local 650 work = NULL; in kthread_worker_fn() 653 work = list_first_entry(&worker->work_list, in kthread_worker_fn() 655 list_del_init(&work->node); in kthread_worker_fn() 657 worker->current_work = work; in kthread_worker_fn() 660 if (work) { in kthread_worker_fn() 662 work->func(work); in kthread_worker_fn() 768 struct kthread_work *work) in queuing_blocked() argument 772 return !list_empty(&work->node) || work->canceling; in queuing_blocked() 776 struct kthread_work *work) in kthread_insert_work_sanity_check() argument [all …]
|
D | irq_work.c | 29 static bool irq_work_claim(struct irq_work *work) in irq_work_claim() argument 37 flags = work->flags & ~IRQ_WORK_PENDING; in irq_work_claim() 40 oflags = cmpxchg(&work->flags, flags, nflags); in irq_work_claim() 65 bool irq_work_queue_on(struct irq_work *work, int cpu) in irq_work_queue_on() argument 76 if (!irq_work_claim(work)) in irq_work_queue_on() 79 if (llist_add(&work->llnode, &per_cpu(raised_list, cpu))) in irq_work_queue_on() 83 irq_work_queue(work); in irq_work_queue_on() 90 bool irq_work_queue(struct irq_work *work) in irq_work_queue() argument 93 if (!irq_work_claim(work)) in irq_work_queue() 100 if (work->flags & IRQ_WORK_LAZY) { in irq_work_queue() [all …]
|
D | workqueue.c | 440 struct work_struct *work = addr; in work_is_static_object() local 442 return test_bit(WORK_STRUCT_STATIC_BIT, work_data_bits(work)); in work_is_static_object() 451 struct work_struct *work = addr; in work_fixup_init() local 455 cancel_work_sync(work); in work_fixup_init() 456 debug_object_init(work, &work_debug_descr); in work_fixup_init() 469 struct work_struct *work = addr; in work_fixup_free() local 473 cancel_work_sync(work); in work_fixup_free() 474 debug_object_free(work, &work_debug_descr); in work_fixup_free() 489 static inline void debug_work_activate(struct work_struct *work) in debug_work_activate() argument 491 debug_object_activate(work, &work_debug_descr); in debug_work_activate() [all …]
|
/Linux-v4.19/include/linux/ |
D | completion.h | 36 #define COMPLETION_INITIALIZER(work) \ argument 37 { 0, __WAIT_QUEUE_HEAD_INITIALIZER((work).wait) } 39 #define COMPLETION_INITIALIZER_ONSTACK_MAP(work, map) \ argument 40 (*({ init_completion_map(&(work), &(map)); &(work); })) 42 #define COMPLETION_INITIALIZER_ONSTACK(work) \ argument 43 (*({ init_completion(&work); &work; })) 53 #define DECLARE_COMPLETION(work) \ argument 54 struct completion work = COMPLETION_INITIALIZER(work) 69 # define DECLARE_COMPLETION_ONSTACK(work) \ argument 70 struct completion work = COMPLETION_INITIALIZER_ONSTACK(work) [all …]
|
D | workqueue.h | 21 typedef void (*work_func_t)(struct work_struct *work); 28 #define work_data_bits(work) ((unsigned long *)(&(work)->data)) argument 116 struct work_struct work; member 125 struct work_struct work; member 158 static inline struct delayed_work *to_delayed_work(struct work_struct *work) in to_delayed_work() argument 160 return container_of(work, struct delayed_work, work); in to_delayed_work() 163 static inline struct rcu_work *to_rcu_work(struct work_struct *work) in to_rcu_work() argument 165 return container_of(work, struct rcu_work, work); in to_rcu_work() 169 struct work_struct work; member 192 .work = __WORK_INITIALIZER((n).work, (f)), \ [all …]
|
D | kthread.h | 79 typedef void (*kthread_work_func_t)(struct kthread_work *work); 104 struct kthread_work work; member 114 #define KTHREAD_WORK_INIT(work, fn) { \ argument 115 .node = LIST_HEAD_INIT((work).node), \ 120 .work = KTHREAD_WORK_INIT((dwork).work, (fn)), \ 128 #define DEFINE_KTHREAD_WORK(work, fn) \ argument 129 struct kthread_work work = KTHREAD_WORK_INIT(work, fn) 157 #define kthread_init_work(work, fn) \ argument 159 memset((work), 0, sizeof(struct kthread_work)); \ 160 INIT_LIST_HEAD(&(work)->node); \ [all …]
|
/Linux-v4.19/fs/btrfs/ |
D | async-thread.c | 54 static void normal_work_helper(struct btrfs_work *work); 59 struct btrfs_work *work = container_of(arg, struct btrfs_work, \ 61 normal_work_helper(work); \ 71 btrfs_work_owner(const struct btrfs_work *work) in btrfs_work_owner() argument 73 return work->wq->fs_info; in btrfs_work_owner() 258 struct btrfs_work *work; in run_ordered_work() local 268 work = list_entry(list->next, struct btrfs_work, in run_ordered_work() 270 if (!test_bit(WORK_DONE_BIT, &work->flags)) in run_ordered_work() 279 if (test_and_set_bit(WORK_ORDER_DONE_BIT, &work->flags)) in run_ordered_work() 281 trace_btrfs_ordered_sched(work); in run_ordered_work() [all …]
|
/Linux-v4.19/drivers/staging/octeon/ |
D | ethernet-rx.c | 74 static inline int cvm_oct_check_rcv_error(cvmx_wqe_t *work) in cvm_oct_check_rcv_error() argument 79 port = work->word0.pip.cn68xx.pknd; in cvm_oct_check_rcv_error() 81 port = work->word1.cn38xx.ipprt; in cvm_oct_check_rcv_error() 83 if ((work->word2.snoip.err_code == 10) && (work->word1.len <= 64)) { in cvm_oct_check_rcv_error() 90 } else if (work->word2.snoip.err_code == 5 || in cvm_oct_check_rcv_error() 91 work->word2.snoip.err_code == 7) { in cvm_oct_check_rcv_error() 108 cvmx_phys_to_ptr(work->packet_ptr.s.addr); in cvm_oct_check_rcv_error() 111 while (i < work->word1.len - 1) { in cvm_oct_check_rcv_error() 120 work->packet_ptr.s.addr += i + 1; in cvm_oct_check_rcv_error() 121 work->word1.len -= i + 5; in cvm_oct_check_rcv_error() [all …]
|
D | ethernet-tx.c | 528 cvmx_wqe_t *work = cvmx_fpa_alloc(CVMX_FPA_WQE_POOL); in cvm_oct_xmit_pow() local 530 if (unlikely(!work)) { in cvm_oct_xmit_pow() 543 cvmx_fpa_free(work, CVMX_FPA_WQE_POOL, 1); in cvm_oct_xmit_pow() 573 work->word0.pip.cn38xx.hw_chksum = skb->csum; in cvm_oct_xmit_pow() 574 work->word1.len = skb->len; in cvm_oct_xmit_pow() 575 cvmx_wqe_set_port(work, priv->port); in cvm_oct_xmit_pow() 576 cvmx_wqe_set_qos(work, priv->port & 0x7); in cvm_oct_xmit_pow() 577 cvmx_wqe_set_grp(work, pow_send_group); in cvm_oct_xmit_pow() 578 work->word1.tag_type = CVMX_HELPER_INPUT_TAG_TYPE; in cvm_oct_xmit_pow() 579 work->word1.tag = pow_send_group; /* FIXME */ in cvm_oct_xmit_pow() [all …]
|
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_display.c | 44 struct amdgpu_flip_work *work = in amdgpu_display_flip_callback() local 48 schedule_work(&work->flip_work.work); in amdgpu_display_flip_callback() 51 static bool amdgpu_display_flip_handle_fence(struct amdgpu_flip_work *work, in amdgpu_display_flip_handle_fence() argument 61 if (!dma_fence_add_callback(fence, &work->cb, in amdgpu_display_flip_handle_fence() 72 container_of(__work, struct delayed_work, work); in amdgpu_display_flip_work_func() 73 struct amdgpu_flip_work *work = in amdgpu_display_flip_work_func() local 75 struct amdgpu_device *adev = work->adev; in amdgpu_display_flip_work_func() 76 struct amdgpu_crtc *amdgpu_crtc = adev->mode_info.crtcs[work->crtc_id]; in amdgpu_display_flip_work_func() 83 if (amdgpu_display_flip_handle_fence(work, &work->excl)) in amdgpu_display_flip_work_func() 86 for (i = 0; i < work->shared_count; ++i) in amdgpu_display_flip_work_func() [all …]
|
/Linux-v4.19/drivers/net/wireless/st/cw1200/ |
D | sta.h | 61 void cw1200_event_handler(struct work_struct *work); 62 void cw1200_bss_loss_work(struct work_struct *work); 63 void cw1200_bss_params_work(struct work_struct *work); 64 void cw1200_keep_alive_work(struct work_struct *work); 65 void cw1200_tx_failure_work(struct work_struct *work); 81 void cw1200_join_timeout(struct work_struct *work); 82 void cw1200_unjoin_work(struct work_struct *work); 83 void cw1200_join_complete_work(struct work_struct *work); 84 void cw1200_wep_key_work(struct work_struct *work); 87 void cw1200_update_filtering_work(struct work_struct *work); [all …]
|
/Linux-v4.19/fs/ |
D | fs-writeback.c | 177 struct wb_writeback_work *work) in finish_writeback_work() argument 179 struct wb_completion *done = work->done; in finish_writeback_work() 181 if (work->auto_free) in finish_writeback_work() 182 kfree(work); in finish_writeback_work() 188 struct wb_writeback_work *work) in wb_queue_work() argument 190 trace_writeback_queue(wb, work); in wb_queue_work() 192 if (work->done) in wb_queue_work() 193 atomic_inc(&work->done->cnt); in wb_queue_work() 198 list_add_tail(&work->list, &wb->work_list); in wb_queue_work() 201 finish_writeback_work(wb, work); in wb_queue_work() [all …]
|
/Linux-v4.19/drivers/infiniband/core/ |
D | cm.c | 237 struct delayed_work work; member 248 struct cm_work work; /* Must be first. */ member 306 static void cm_work_handler(struct work_struct *work); 743 __be32 remote_id = timewait_info->work.remote_id; in cm_insert_remote_id() 749 if (be32_lt(remote_id, cur_timewait_info->work.remote_id)) in cm_insert_remote_id() 751 else if (be32_gt(remote_id, cur_timewait_info->work.remote_id)) in cm_insert_remote_id() 775 if (be32_lt(remote_id, timewait_info->work.remote_id)) in cm_find_remote_id() 777 else if (be32_gt(remote_id, timewait_info->work.remote_id)) in cm_find_remote_id() 900 struct cm_work *work; in cm_dequeue_work() local 905 work = list_entry(cm_id_priv->work_list.next, struct cm_work, list); in cm_dequeue_work() [all …]
|
/Linux-v4.19/drivers/gpu/drm/i915/ |
D | i915_gem_userptr.c | 42 struct work_struct work; member 61 struct work_struct work; member 65 static void cancel_userptr(struct work_struct *work) in cancel_userptr() argument 67 struct i915_mmu_object *mo = container_of(work, typeof(*mo), work); in cancel_userptr() 73 active = fetch_and_zero(&obj->userptr.work); in cancel_userptr() 151 queue_work(mn->wq, &mo->work); in i915_gem_userptr_mn_invalidate_range_start() 276 INIT_WORK(&mo->work, cancel_userptr); in i915_gem_userptr_init__mmu_notifier() 382 __i915_mm_struct_free__worker(struct work_struct *work) in __i915_mm_struct_free__worker() argument 384 struct i915_mm_struct *mm = container_of(work, typeof(*mm), work); in __i915_mm_struct_free__worker() 399 INIT_WORK(&mm->work, __i915_mm_struct_free__worker); in __i915_mm_struct_free() [all …]
|
/Linux-v4.19/drivers/gpu/drm/i915/selftests/ |
D | igt_wedge_me.h | 17 struct delayed_work work; member 22 static void __igt_wedge_me(struct work_struct *work) in __igt_wedge_me() argument 24 struct igt_wedge_me *w = container_of(work, typeof(*w), work.work); in __igt_wedge_me() 42 INIT_DELAYED_WORK_ONSTACK(&w->work, __igt_wedge_me); in __igt_init_wedge() 43 schedule_delayed_work(&w->work, timeout); in __igt_init_wedge() 48 cancel_delayed_work_sync(&w->work); in __igt_fini_wedge() 49 destroy_delayed_work_on_stack(&w->work); in __igt_fini_wedge()
|
/Linux-v4.19/lib/ |
D | once.c | 8 struct work_struct work; member 14 struct once_work *work; in once_deferred() local 16 work = container_of(w, struct once_work, work); in once_deferred() 17 BUG_ON(!static_key_enabled(work->key)); in once_deferred() 18 static_branch_disable(work->key); in once_deferred() 19 kfree(work); in once_deferred() 30 INIT_WORK(&w->work, once_deferred); in once_disable_jump() 32 schedule_work(&w->work); in once_disable_jump()
|
/Linux-v4.19/kernel/locking/ |
D | test-ww_mutex.c | 33 struct work_struct work; member 44 static void test_mutex_work(struct work_struct *work) in test_mutex_work() argument 46 struct test_mutex *mtx = container_of(work, typeof(*mtx), work); in test_mutex_work() 71 INIT_WORK_ONSTACK(&mtx.work, test_mutex_work); in __test_mutex() 77 schedule_work(&mtx.work); in __test_mutex() 105 flush_work(&mtx.work); in __test_mutex() 106 destroy_work_on_stack(&mtx.work); in __test_mutex() 161 struct work_struct work; member 170 static void test_abba_work(struct work_struct *work) in test_abba_work() argument 172 struct test_abba *abba = container_of(work, typeof(*abba), work); in test_abba_work() [all …]
|
/Linux-v4.19/drivers/misc/mic/scif/ |
D | scif_dma.c | 293 void scif_mmu_notif_handler(struct work_struct *work) in scif_mmu_notif_handler() argument 329 void scif_mmu_notif_handler(struct work_struct *work) in scif_mmu_notif_handler() argument 693 iounmap_remote(void *virt, size_t size, struct scif_copy_work *work) in iounmap_remote() argument 695 scif_iounmap(virt, size, work->remote_dev); in iounmap_remote() 881 scif_rma_list_dma_copy_unaligned(struct scif_copy_work *work, in scif_rma_list_dma_copy_unaligned() argument 885 struct scif_dma_comp_cb *comp_cb = work->comp_cb; in scif_rma_list_dma_copy_unaligned() 888 size_t loop_len, nr_contig_bytes = 0, remaining_len = work->len; in scif_rma_list_dma_copy_unaligned() 899 offset = work->dst_offset; in scif_rma_list_dma_copy_unaligned() 900 window = work->dst_window; in scif_rma_list_dma_copy_unaligned() 902 offset = work->src_offset; in scif_rma_list_dma_copy_unaligned() [all …]
|
/Linux-v4.19/arch/sparc/kernel/ |
D | sun4d_smp.c | 196 struct sun4d_ipi_work *work; in smp4d_ipi_init() local 201 work = &per_cpu(sun4d_ipi_work, cpu); in smp4d_ipi_init() 202 work->single = work->msk = work->resched = 0; in smp4d_ipi_init() 208 struct sun4d_ipi_work *work = this_cpu_ptr(&sun4d_ipi_work); in sun4d_ipi_interrupt() local 210 if (work->single) { in sun4d_ipi_interrupt() 211 work->single = 0; in sun4d_ipi_interrupt() 214 if (work->msk) { in sun4d_ipi_interrupt() 215 work->msk = 0; in sun4d_ipi_interrupt() 218 if (work->resched) { in sun4d_ipi_interrupt() 219 work->resched = 0; in sun4d_ipi_interrupt() [all …]
|
/Linux-v4.19/drivers/md/ |
D | dm-cache-background-tracker.c | 16 struct policy_work work; member 90 cmp = cmp_oblock(w->work.oblock, nw->work.oblock); in __insert_pending() 118 cmp = cmp_oblock(w->work.oblock, oblock); in __find_pending() 178 struct policy_work *work, in btracker_queue() argument 190 memcpy(&w->work, work, sizeof(*work)); in btracker_queue() 202 *pwork = &w->work; in btracker_queue() 206 update_stats(b, &w->work, 1); in btracker_queue() 215 int btracker_issue(struct background_tracker *b, struct policy_work **work) in btracker_issue() argument 224 *work = &w->work; in btracker_issue() 233 struct bt_work *w = container_of(op, struct bt_work, work); in btracker_complete() [all …]
|
/Linux-v4.19/drivers/infiniband/ulp/ipoib/ |
D | ipoib_vlan.c | 204 struct work_struct work; member 218 static void ipoib_vlan_delete_task(struct work_struct *work) in ipoib_vlan_delete_task() argument 221 container_of(work, struct ipoib_vlan_delete_work, work); in ipoib_vlan_delete_task() 262 struct ipoib_vlan_delete_work *work; in ipoib_vlan_delete() local 264 work = kmalloc(sizeof(*work), GFP_KERNEL); in ipoib_vlan_delete() 265 if (!work) { in ipoib_vlan_delete() 273 work->dev = priv->dev; in ipoib_vlan_delete() 274 INIT_WORK(&work->work, ipoib_vlan_delete_task); in ipoib_vlan_delete() 275 queue_work(ipoib_workqueue, &work->work); in ipoib_vlan_delete()
|