/Linux-v5.4/drivers/gpu/drm/scheduler/ |
D | sched_entity.c | 52 struct drm_sched_rq **rq_list, in drm_sched_entity_init() argument 58 if (!(entity && rq_list && (num_rq_list == 0 || rq_list[0]))) in drm_sched_entity_init() 66 entity->rq_list = kcalloc(num_rq_list, sizeof(struct drm_sched_rq *), in drm_sched_entity_init() 68 if (!entity->rq_list) in drm_sched_entity_init() 72 entity->rq_list[i] = rq_list[i]; in drm_sched_entity_init() 75 entity->rq = rq_list[0]; in drm_sched_entity_init() 140 struct drm_gpu_scheduler *sched = entity->rq_list[i]->sched; in drm_sched_entity_get_free_sched() 142 if (!entity->rq_list[i]->sched->ready) { in drm_sched_entity_get_free_sched() 150 rq = entity->rq_list[i]; in drm_sched_entity_get_free_sched() 307 kfree(entity->rq_list); in drm_sched_entity_fini() [all …]
|
/Linux-v5.4/block/ |
D | blk-mq-sched.c | 92 LIST_HEAD(rq_list); in blk_mq_do_dispatch_sched() 114 list_add(&rq->queuelist, &rq_list); in blk_mq_do_dispatch_sched() 115 } while (blk_mq_dispatch_rq_list(q, &rq_list, true)); in blk_mq_do_dispatch_sched() 137 LIST_HEAD(rq_list); in blk_mq_do_dispatch_ctx() 160 list_add(&rq->queuelist, &rq_list); in blk_mq_do_dispatch_ctx() 165 } while (blk_mq_dispatch_rq_list(q, &rq_list, true)); in blk_mq_do_dispatch_ctx() 175 LIST_HEAD(rq_list); in blk_mq_sched_dispatch_requests() 190 list_splice_init(&hctx->dispatch, &rq_list); in blk_mq_sched_dispatch_requests() 207 if (!list_empty(&rq_list)) { in blk_mq_sched_dispatch_requests() 209 if (blk_mq_dispatch_rq_list(q, &rq_list, false)) { in blk_mq_sched_dispatch_requests() [all …]
|
D | kyber-iosched.c | 147 struct list_head rq_list[KYBER_NUM_DOMAINS]; member 458 INIT_LIST_HEAD(&kcq->rq_list[i]); in kyber_ctx_queue_init() 572 struct list_head *rq_list = &kcq->rq_list[sched_domain]; in kyber_bio_merge() local 576 merged = blk_mq_bio_list_merge(hctx->queue, rq_list, bio, nr_segs); in kyber_bio_merge() 588 struct list_head *rq_list, bool at_head) in kyber_insert_requests() argument 593 list_for_each_entry_safe(rq, next, rq_list, queuelist) { in kyber_insert_requests() 596 struct list_head *head = &kcq->rq_list[sched_domain]; in kyber_insert_requests() 669 list_splice_tail_init(&kcq->rq_list[flush_data->sched_domain], in flush_busy_kcq()
|
D | blk-mq.c | 745 LIST_HEAD(rq_list); in blk_mq_requeue_work() 749 list_splice_init(&q->requeue_list, &rq_list); in blk_mq_requeue_work() 752 list_for_each_entry_safe(rq, next, &rq_list, queuelist) { in blk_mq_requeue_work() 769 while (!list_empty(&rq_list)) { in blk_mq_requeue_work() 770 rq = list_entry(rq_list.next, struct request, queuelist); in blk_mq_requeue_work() 1729 LIST_HEAD(rq_list); in blk_mq_flush_plug_list() 1752 &rq_list, in blk_mq_flush_plug_list() 1763 list_add_tail(&rq->queuelist, &rq_list); in blk_mq_flush_plug_list() 1772 blk_mq_sched_insert_requests(this_hctx, this_ctx, &rq_list, in blk_mq_flush_plug_list()
|
/Linux-v5.4/include/drm/ |
D | gpu_scheduler.h | 82 struct drm_sched_rq **rq_list; member 312 struct drm_sched_rq **rq_list,
|
/Linux-v5.4/include/linux/mtd/ |
D | blktrans.h | 33 struct list_head rq_list; member
|
/Linux-v5.4/drivers/block/paride/ |
D | pcd.c | 204 struct list_head rq_list; member 322 INIT_LIST_HEAD(&cd->rq_list); in pcd_init_units() 780 if (cd->present && !list_empty(&cd->rq_list)) { in set_next_request() 781 pcd_req = list_first_entry(&cd->rq_list, struct request, in set_next_request() 824 list_add_tail(&bd->rq->queuelist, &cd->rq_list); in pcd_queue_rq()
|
D | pd.c | 240 struct list_head rq_list; member 411 if (list_empty(&disk->rq_list)) in set_next_request() 414 pd_req = list_first_entry(&disk->rq_list, in set_next_request() 765 list_add_tail(&bd->rq->queuelist, &disk->rq_list); in pd_queue_rq() 955 INIT_LIST_HEAD(&disk->rq_list); in pd_detect()
|
D | pf.c | 243 struct list_head rq_list; member 308 INIT_LIST_HEAD(&pf->rq_list); in pf_init_units() 815 if (pf->present && !list_empty(&pf->rq_list)) { in set_next_request() 816 pf_req = list_first_entry(&pf->rq_list, struct request, in set_next_request() 877 list_add_tail(&bd->rq->queuelist, &pf->rq_list); in pf_queue_rq()
|
/Linux-v5.4/drivers/scsi/ |
D | sg.c | 144 struct list_head rq_list; /* head of request list */ member 871 list_for_each_entry(srp, &sfp->rq_list, entry) { in sg_fill_request_table() 998 list_for_each_entry(srp, &sfp->rq_list, entry) { in sg_ioctl() 1012 list_for_each_entry(srp, &sfp->rq_list, entry) { in sg_ioctl() 1184 list_for_each_entry(srp, &sfp->rq_list, entry) { in sg_poll() 2079 list_for_each_entry(resp, &sfp->rq_list, entry) { in sg_get_rq_mark() 2101 if (!list_empty(&sfp->rq_list)) { in sg_add_request() 2115 list_add_tail(&rp->entry, &sfp->rq_list); in sg_add_request() 2130 if (!sfp || !srp || list_empty(&sfp->rq_list)) in sg_remove_request() 2155 INIT_LIST_HEAD(&sfp->rq_list); in sg_add_sfp() [all …]
|
/Linux-v5.4/drivers/block/ |
D | xsysace.c | 210 struct list_head rq_list; member 468 return !list_empty(&ace->rq_list); in ace_has_next_request() 477 rq = list_first_entry_or_null(&ace->rq_list, struct request, queuelist); in ace_get_next_request() 876 list_add_tail(&req->queuelist, &ace->rq_list); in ace_queue_rq() 992 INIT_LIST_HEAD(&ace->rq_list); in ace_setup()
|
/Linux-v5.4/drivers/mtd/ |
D | mtd_blkdevs.c | 130 rq = list_first_entry_or_null(&dev->rq_list, struct request, queuelist); in mtd_next_request() 197 list_add_tail(&bd->rq->queuelist, &dev->rq_list); in mtd_queue_rq() 427 INIT_LIST_HEAD(&new->rq_list); in add_mtd_blktrans_dev()
|
/Linux-v5.4/net/sunrpc/ |
D | xprt.c | 1593 req = list_entry(xprt->free.next, struct rpc_rqst, rq_list); in xprt_alloc_slot() 1594 list_del(&req->rq_list); in xprt_alloc_slot() 1630 list_add(&req->rq_list, &xprt->free); in xprt_free_slot() 1641 req = list_first_entry(&xprt->free, struct rpc_rqst, rq_list); in xprt_free_all_slots() 1642 list_del(&req->rq_list); in xprt_free_all_slots() 1665 list_add(&req->rq_list, &xprt->free); in xprt_alloc()
|
/Linux-v5.4/drivers/block/aoe/ |
D | aoe.h | 173 struct list_head rq_list; member
|
D | aoedev.c | 475 INIT_LIST_HEAD(&d->rq_list); in aoedev_by_aoeaddr()
|
D | aoeblk.c | 278 list_add_tail(&bd->rq->queuelist, &d->rq_list); in aoeblk_queue_rq()
|
D | aoecmd.c | 850 rq = list_first_entry_or_null(&d->rq_list, struct request, in nextbuf()
|
/Linux-v5.4/drivers/ide/ |
D | ide-io.c | 547 list_add(&rq->queuelist, &drive->rq_list); in ide_issue_rq() 903 list_add_tail(&rq->queuelist, &drive->rq_list); in ide_insert_request_head()
|
D | ide-probe.c | 1170 while (!list_empty(&drive->rq_list)) { in drive_rq_insert_work() 1171 rq = list_first_entry(&drive->rq_list, struct request, queuelist); in drive_rq_insert_work() 1219 INIT_LIST_HEAD(&drive->rq_list); in ide_port_init_devices_data()
|
/Linux-v5.4/include/linux/sunrpc/ |
D | xprt.h | 87 struct list_head rq_list; /* Slot allocation list */ member
|
/Linux-v5.4/include/linux/ |
D | ide.h | 624 struct list_head rq_list; member
|