Lines Matching refs:qc

1462 static void ata_qc_complete_internal(struct ata_queued_cmd *qc)  in ata_qc_complete_internal()  argument
1464 struct completion *waiting = qc->private_data; in ata_qc_complete_internal()
1500 struct ata_queued_cmd *qc; in ata_exec_internal_sg() local
1519 qc = __ata_qc_from_tag(ap, ATA_TAG_INTERNAL); in ata_exec_internal_sg()
1521 qc->tag = ATA_TAG_INTERNAL; in ata_exec_internal_sg()
1522 qc->hw_tag = 0; in ata_exec_internal_sg()
1523 qc->scsicmd = NULL; in ata_exec_internal_sg()
1524 qc->ap = ap; in ata_exec_internal_sg()
1525 qc->dev = dev; in ata_exec_internal_sg()
1526 ata_qc_reinit(qc); in ata_exec_internal_sg()
1538 qc->tf = *tf; in ata_exec_internal_sg()
1540 memcpy(qc->cdb, cdb, ATAPI_CDB_LEN); in ata_exec_internal_sg()
1545 qc->tf.feature |= ATAPI_DMADIR; in ata_exec_internal_sg()
1547 qc->flags |= ATA_QCFLAG_RESULT_TF; in ata_exec_internal_sg()
1548 qc->dma_dir = dma_dir; in ata_exec_internal_sg()
1556 ata_sg_init(qc, sgl, n_elem); in ata_exec_internal_sg()
1557 qc->nbytes = buflen; in ata_exec_internal_sg()
1560 qc->private_data = &wait; in ata_exec_internal_sg()
1561 qc->complete_fn = ata_qc_complete_internal; in ata_exec_internal_sg()
1563 ata_qc_issue(qc); in ata_exec_internal_sg()
1594 if (qc->flags & ATA_QCFLAG_ACTIVE) { in ata_exec_internal_sg()
1595 qc->err_mask |= AC_ERR_TIMEOUT; in ata_exec_internal_sg()
1600 ata_qc_complete(qc); in ata_exec_internal_sg()
1612 ap->ops->post_internal_cmd(qc); in ata_exec_internal_sg()
1615 if (qc->flags & ATA_QCFLAG_FAILED) { in ata_exec_internal_sg()
1616 if (qc->result_tf.command & (ATA_ERR | ATA_DF)) in ata_exec_internal_sg()
1617 qc->err_mask |= AC_ERR_DEV; in ata_exec_internal_sg()
1619 if (!qc->err_mask) in ata_exec_internal_sg()
1620 qc->err_mask |= AC_ERR_OTHER; in ata_exec_internal_sg()
1622 if (qc->err_mask & ~AC_ERR_OTHER) in ata_exec_internal_sg()
1623 qc->err_mask &= ~AC_ERR_OTHER; in ata_exec_internal_sg()
1624 } else if (qc->tf.command == ATA_CMD_REQ_SENSE_DATA) { in ata_exec_internal_sg()
1625 qc->result_tf.command |= ATA_SENSE; in ata_exec_internal_sg()
1631 *tf = qc->result_tf; in ata_exec_internal_sg()
1632 err_mask = qc->err_mask; in ata_exec_internal_sg()
1634 ata_qc_free(qc); in ata_exec_internal_sg()
4328 int atapi_check_dma(struct ata_queued_cmd *qc) in atapi_check_dma() argument
4330 struct ata_port *ap = qc->ap; in atapi_check_dma()
4335 if (!(qc->dev->horkage & ATA_HORKAGE_ATAPI_MOD16_DMA) && in atapi_check_dma()
4336 unlikely(qc->nbytes & 15)) in atapi_check_dma()
4340 return ap->ops->check_atapi_dma(qc); in atapi_check_dma()
4360 int ata_std_qc_defer(struct ata_queued_cmd *qc) in ata_std_qc_defer() argument
4362 struct ata_link *link = qc->dev->link; in ata_std_qc_defer()
4364 if (ata_is_ncq(qc->tf.protocol)) { in ata_std_qc_defer()
4376 enum ata_completion_errors ata_noop_qc_prep(struct ata_queued_cmd *qc) in ata_noop_qc_prep() argument
4395 void ata_sg_init(struct ata_queued_cmd *qc, struct scatterlist *sg, in ata_sg_init() argument
4398 qc->sg = sg; in ata_sg_init()
4399 qc->n_elem = n_elem; in ata_sg_init()
4400 qc->cursg = qc->sg; in ata_sg_init()
4414 static void ata_sg_clean(struct ata_queued_cmd *qc) in ata_sg_clean() argument
4416 struct ata_port *ap = qc->ap; in ata_sg_clean()
4417 struct scatterlist *sg = qc->sg; in ata_sg_clean()
4418 int dir = qc->dma_dir; in ata_sg_clean()
4422 VPRINTK("unmapping %u sg elements\n", qc->n_elem); in ata_sg_clean()
4424 if (qc->n_elem) in ata_sg_clean()
4425 dma_unmap_sg(ap->dev, sg, qc->orig_n_elem, dir); in ata_sg_clean()
4427 qc->flags &= ~ATA_QCFLAG_DMAMAP; in ata_sg_clean()
4428 qc->sg = NULL; in ata_sg_clean()
4444 static int ata_sg_setup(struct ata_queued_cmd *qc) in ata_sg_setup() argument
4446 struct ata_port *ap = qc->ap; in ata_sg_setup()
4451 n_elem = dma_map_sg(ap->dev, qc->sg, qc->n_elem, qc->dma_dir); in ata_sg_setup()
4456 qc->orig_n_elem = qc->n_elem; in ata_sg_setup()
4457 qc->n_elem = n_elem; in ata_sg_setup()
4458 qc->flags |= ATA_QCFLAG_DMAMAP; in ata_sg_setup()
4465 static inline void ata_sg_clean(struct ata_queued_cmd *qc) {} in ata_sg_clean() argument
4466 static inline int ata_sg_setup(struct ata_queued_cmd *qc) { return -1; } in ata_sg_setup() argument
4504 struct ata_queued_cmd *qc; in ata_qc_new_init() local
4517 qc = __ata_qc_from_tag(ap, tag); in ata_qc_new_init()
4518 qc->tag = qc->hw_tag = tag; in ata_qc_new_init()
4519 qc->scsicmd = NULL; in ata_qc_new_init()
4520 qc->ap = ap; in ata_qc_new_init()
4521 qc->dev = dev; in ata_qc_new_init()
4523 ata_qc_reinit(qc); in ata_qc_new_init()
4525 return qc; in ata_qc_new_init()
4538 void ata_qc_free(struct ata_queued_cmd *qc) in ata_qc_free() argument
4543 WARN_ON_ONCE(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ in ata_qc_free()
4544 ap = qc->ap; in ata_qc_free()
4546 qc->flags = 0; in ata_qc_free()
4547 tag = qc->tag; in ata_qc_free()
4549 qc->tag = ATA_TAG_POISON; in ata_qc_free()
4555 void __ata_qc_complete(struct ata_queued_cmd *qc) in __ata_qc_complete() argument
4560 WARN_ON_ONCE(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ in __ata_qc_complete()
4561 WARN_ON_ONCE(!(qc->flags & ATA_QCFLAG_ACTIVE)); in __ata_qc_complete()
4562 ap = qc->ap; in __ata_qc_complete()
4563 link = qc->dev->link; in __ata_qc_complete()
4565 if (likely(qc->flags & ATA_QCFLAG_DMAMAP)) in __ata_qc_complete()
4566 ata_sg_clean(qc); in __ata_qc_complete()
4569 if (ata_is_ncq(qc->tf.protocol)) { in __ata_qc_complete()
4570 link->sactive &= ~(1 << qc->hw_tag); in __ata_qc_complete()
4579 if (unlikely(qc->flags & ATA_QCFLAG_CLEAR_EXCL && in __ata_qc_complete()
4587 qc->flags &= ~ATA_QCFLAG_ACTIVE; in __ata_qc_complete()
4588 ap->qc_active &= ~(1ULL << qc->tag); in __ata_qc_complete()
4591 qc->complete_fn(qc); in __ata_qc_complete()
4594 static void fill_result_tf(struct ata_queued_cmd *qc) in fill_result_tf() argument
4596 struct ata_port *ap = qc->ap; in fill_result_tf()
4598 qc->result_tf.flags = qc->tf.flags; in fill_result_tf()
4599 ap->ops->qc_fill_rtf(qc); in fill_result_tf()
4602 static void ata_verify_xfer(struct ata_queued_cmd *qc) in ata_verify_xfer() argument
4604 struct ata_device *dev = qc->dev; in ata_verify_xfer()
4606 if (!ata_is_data(qc->tf.protocol)) in ata_verify_xfer()
4609 if ((dev->mwdma_mask || dev->udma_mask) && ata_is_pio(qc->tf.protocol)) in ata_verify_xfer()
4630 void ata_qc_complete(struct ata_queued_cmd *qc) in ata_qc_complete() argument
4632 struct ata_port *ap = qc->ap; in ata_qc_complete()
4635 ledtrig_disk_activity(!!(qc->tf.flags & ATA_TFLAG_WRITE)); in ata_qc_complete()
4651 struct ata_device *dev = qc->dev; in ata_qc_complete()
4654 if (unlikely(qc->err_mask)) in ata_qc_complete()
4655 qc->flags |= ATA_QCFLAG_FAILED; in ata_qc_complete()
4661 if (unlikely(ata_tag_internal(qc->tag))) { in ata_qc_complete()
4662 fill_result_tf(qc); in ata_qc_complete()
4663 trace_ata_qc_complete_internal(qc); in ata_qc_complete()
4664 __ata_qc_complete(qc); in ata_qc_complete()
4672 if (unlikely(qc->flags & ATA_QCFLAG_FAILED)) { in ata_qc_complete()
4673 fill_result_tf(qc); in ata_qc_complete()
4674 trace_ata_qc_complete_failed(qc); in ata_qc_complete()
4675 ata_qc_schedule_eh(qc); in ata_qc_complete()
4682 if (qc->flags & ATA_QCFLAG_RESULT_TF) in ata_qc_complete()
4683 fill_result_tf(qc); in ata_qc_complete()
4685 trace_ata_qc_complete_done(qc); in ata_qc_complete()
4689 switch (qc->tf.command) { in ata_qc_complete()
4691 if (qc->tf.feature != SETFEATURES_WC_ON && in ata_qc_complete()
4692 qc->tf.feature != SETFEATURES_WC_OFF && in ata_qc_complete()
4693 qc->tf.feature != SETFEATURES_RA_ON && in ata_qc_complete()
4694 qc->tf.feature != SETFEATURES_RA_OFF) in ata_qc_complete()
4710 ata_verify_xfer(qc); in ata_qc_complete()
4712 __ata_qc_complete(qc); in ata_qc_complete()
4714 if (qc->flags & ATA_QCFLAG_EH_SCHEDULED) in ata_qc_complete()
4718 if (qc->err_mask || qc->flags & ATA_QCFLAG_RESULT_TF) in ata_qc_complete()
4719 fill_result_tf(qc); in ata_qc_complete()
4721 __ata_qc_complete(qc); in ata_qc_complete()
4762 void ata_qc_issue(struct ata_queued_cmd *qc) in ata_qc_issue() argument
4764 struct ata_port *ap = qc->ap; in ata_qc_issue()
4765 struct ata_link *link = qc->dev->link; in ata_qc_issue()
4766 u8 prot = qc->tf.protocol; in ata_qc_issue()
4775 WARN_ON_ONCE(link->sactive & (1 << qc->hw_tag)); in ata_qc_issue()
4779 link->sactive |= 1 << qc->hw_tag; in ata_qc_issue()
4784 link->active_tag = qc->tag; in ata_qc_issue()
4787 qc->flags |= ATA_QCFLAG_ACTIVE; in ata_qc_issue()
4788 ap->qc_active |= 1ULL << qc->tag; in ata_qc_issue()
4794 if (ata_is_data(prot) && (!qc->sg || !qc->n_elem || !qc->nbytes)) in ata_qc_issue()
4799 if (ata_sg_setup(qc)) in ata_qc_issue()
4803 if (unlikely(qc->dev->flags & ATA_DFLAG_SLEEPING)) { in ata_qc_issue()
4810 qc->err_mask |= ap->ops->qc_prep(qc); in ata_qc_issue()
4811 if (unlikely(qc->err_mask)) in ata_qc_issue()
4813 trace_ata_qc_issue(qc); in ata_qc_issue()
4814 qc->err_mask |= ap->ops->qc_issue(qc); in ata_qc_issue()
4815 if (unlikely(qc->err_mask)) in ata_qc_issue()
4820 qc->err_mask |= AC_ERR_SYSTEM; in ata_qc_issue()
4822 ata_qc_complete(qc); in ata_qc_issue()
6407 static unsigned int ata_dummy_qc_issue(struct ata_queued_cmd *qc) in ata_dummy_qc_issue() argument