Searched refs:ctxs (Results 1 – 12 of 12) sorted by relevance
/Linux-v5.4/drivers/video/fbdev/omap2/omapfb/ |
D | vrfb.c | 67 static struct vrfb_ctx *ctxs; variable 88 omap2_sms_write_rot_control(ctxs[ctx].control, ctx); in restore_hw_context() 89 omap2_sms_write_rot_size(ctxs[ctx].size, ctx); in restore_hw_context() 90 omap2_sms_write_rot_physical_ba(ctxs[ctx].physical_ba, ctx); in restore_hw_context() 211 ctxs[ctx].physical_ba = paddr; in omap_vrfb_setup() 212 ctxs[ctx].size = size; in omap_vrfb_setup() 213 ctxs[ctx].control = control; in omap_vrfb_setup() 308 paddr = ctxs[ctx].base + SMS_ROT_VIRT_BASE(rot); in omap_vrfb_request_ctx() 350 ctxs = devm_kcalloc(&pdev->dev, in vrfb_probe() 354 if (!ctxs) in vrfb_probe() [all …]
|
/Linux-v5.4/arch/x86/mm/ |
D | tlb.c | 68 this_cpu_write(cpu_tlbstate.ctxs[asid].ctx_id, 0); in clear_asid_other() 91 if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) != in choose_new_asid() 96 *need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < in choose_new_asid() 334 VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[prev_asid].ctx_id) != in switch_mm_irqs_off() 362 if (this_cpu_read(cpu_tlbstate.ctxs[prev_asid].tlb_gen) == in switch_mm_irqs_off() 415 this_cpu_write(cpu_tlbstate.ctxs[new_asid].ctx_id, next->context.ctx_id); in switch_mm_irqs_off() 416 this_cpu_write(cpu_tlbstate.ctxs[new_asid].tlb_gen, next_tlb_gen); in switch_mm_irqs_off() 507 this_cpu_write(cpu_tlbstate.ctxs[0].ctx_id, mm->context.ctx_id); in initialize_tlbstate_and_flush() 508 this_cpu_write(cpu_tlbstate.ctxs[0].tlb_gen, tlb_gen); in initialize_tlbstate_and_flush() 511 this_cpu_write(cpu_tlbstate.ctxs[i].ctx_id, 0); in initialize_tlbstate_and_flush() [all …]
|
/Linux-v5.4/block/ |
D | blk-mq-sysfs.c | 20 struct blk_mq_ctxs *ctxs = container_of(kobj, struct blk_mq_ctxs, kobj); in blk_mq_sysfs_release() local 22 free_percpu(ctxs->queue_ctx); in blk_mq_sysfs_release() 23 kfree(ctxs); in blk_mq_sysfs_release() 31 kobject_put(&ctx->ctxs->kobj); in blk_mq_ctx_sysfs_release() 46 kfree(hctx->ctxs); in blk_mq_hw_sysfs_release()
|
D | blk-mq.c | 983 struct blk_mq_ctx *ctx = hctx->ctxs[bitnr]; in flush_busy_ctx() 1018 struct blk_mq_ctx *ctx = hctx->ctxs[bitnr]; in dispatch_rq_from_ctx() 2375 hctx->ctxs = kmalloc_array_node(nr_cpu_ids, sizeof(void *), in blk_mq_alloc_hctx() 2377 if (!hctx->ctxs) in blk_mq_alloc_hctx() 2403 kfree(hctx->ctxs); in blk_mq_alloc_hctx() 2524 hctx->ctxs[hctx->nr_ctx++] = ctx; in blk_mq_map_swqueue() 2644 struct blk_mq_ctxs *ctxs; in blk_mq_alloc_ctxs() local 2647 ctxs = kzalloc(sizeof(*ctxs), GFP_KERNEL); in blk_mq_alloc_ctxs() 2648 if (!ctxs) in blk_mq_alloc_ctxs() 2651 ctxs->queue_ctx = alloc_percpu(struct blk_mq_ctx); in blk_mq_alloc_ctxs() [all …]
|
D | blk-mq.h | 36 struct blk_mq_ctxs *ctxs; member
|
D | blk-mq-sched.c | 126 return hctx->ctxs[idx]; in blk_mq_next_ctx()
|
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | nv04.h | 9 unsigned ctxs:5; member
|
D | dmanv04.c | 101 u32 cm = ((1ULL << c->bits) - 1) << c->ctxs; in nv04_fifo_dma_fini() 104 nvkm_wo32(fctx, c->ctxp + data, cv | (rv << c->ctxs)); in nv04_fifo_dma_fini()
|
/Linux-v5.4/drivers/iommu/ |
D | qcom_iommu.c | 51 struct qcom_iommu_ctx *ctxs[0]; /* indexed by asid-1 */ member 89 return qcom_iommu->ctxs[asid - 1]; in to_ctx() 756 qcom_iommu->ctxs[ctx->asid - 1] = ctx; in qcom_iommu_ctx_probe() 768 qcom_iommu->ctxs[ctx->asid - 1] = NULL; in qcom_iommu_ctx_remove() 813 qcom_iommu = devm_kzalloc(dev, struct_size(qcom_iommu, ctxs, max_asid), in qcom_iommu_device_probe()
|
/Linux-v5.4/include/linux/ |
D | blk-mq.h | 42 struct blk_mq_ctx **ctxs; member 365 ({ ctx = (hctx)->ctxs[(i)]; 1; }); (i)++)
|
/Linux-v5.4/arch/x86/include/asm/ |
D | tlbflush.h | 241 struct tlb_context ctxs[TLB_NR_DYN_ASIDS]; member
|
/Linux-v5.4/kernel/ |
D | workqueue.c | 5218 LIST_HEAD(ctxs); in workqueue_apply_unbound_cpumask() 5238 list_add_tail(&ctx->list, &ctxs); in workqueue_apply_unbound_cpumask() 5241 list_for_each_entry_safe(ctx, n, &ctxs, list) { in workqueue_apply_unbound_cpumask()
|