Lines Matching refs:tf
399 void ata_sff_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) in ata_sff_tf_load() argument
402 unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR; in ata_sff_tf_load()
404 if (tf->ctl != ap->last_ctl) { in ata_sff_tf_load()
406 iowrite8(tf->ctl, ioaddr->ctl_addr); in ata_sff_tf_load()
407 ap->last_ctl = tf->ctl; in ata_sff_tf_load()
411 if (is_addr && (tf->flags & ATA_TFLAG_LBA48)) { in ata_sff_tf_load()
413 iowrite8(tf->hob_feature, ioaddr->feature_addr); in ata_sff_tf_load()
414 iowrite8(tf->hob_nsect, ioaddr->nsect_addr); in ata_sff_tf_load()
415 iowrite8(tf->hob_lbal, ioaddr->lbal_addr); in ata_sff_tf_load()
416 iowrite8(tf->hob_lbam, ioaddr->lbam_addr); in ata_sff_tf_load()
417 iowrite8(tf->hob_lbah, ioaddr->lbah_addr); in ata_sff_tf_load()
421 iowrite8(tf->feature, ioaddr->feature_addr); in ata_sff_tf_load()
422 iowrite8(tf->nsect, ioaddr->nsect_addr); in ata_sff_tf_load()
423 iowrite8(tf->lbal, ioaddr->lbal_addr); in ata_sff_tf_load()
424 iowrite8(tf->lbam, ioaddr->lbam_addr); in ata_sff_tf_load()
425 iowrite8(tf->lbah, ioaddr->lbah_addr); in ata_sff_tf_load()
428 if (tf->flags & ATA_TFLAG_DEVICE) in ata_sff_tf_load()
429 iowrite8(tf->device, ioaddr->device_addr); in ata_sff_tf_load()
448 void ata_sff_tf_read(struct ata_port *ap, struct ata_taskfile *tf) in ata_sff_tf_read() argument
452 tf->status = ata_sff_check_status(ap); in ata_sff_tf_read()
453 tf->error = ioread8(ioaddr->error_addr); in ata_sff_tf_read()
454 tf->nsect = ioread8(ioaddr->nsect_addr); in ata_sff_tf_read()
455 tf->lbal = ioread8(ioaddr->lbal_addr); in ata_sff_tf_read()
456 tf->lbam = ioread8(ioaddr->lbam_addr); in ata_sff_tf_read()
457 tf->lbah = ioread8(ioaddr->lbah_addr); in ata_sff_tf_read()
458 tf->device = ioread8(ioaddr->device_addr); in ata_sff_tf_read()
460 if (tf->flags & ATA_TFLAG_LBA48) { in ata_sff_tf_read()
462 iowrite8(tf->ctl | ATA_HOB, ioaddr->ctl_addr); in ata_sff_tf_read()
463 tf->hob_feature = ioread8(ioaddr->error_addr); in ata_sff_tf_read()
464 tf->hob_nsect = ioread8(ioaddr->nsect_addr); in ata_sff_tf_read()
465 tf->hob_lbal = ioread8(ioaddr->lbal_addr); in ata_sff_tf_read()
466 tf->hob_lbam = ioread8(ioaddr->lbam_addr); in ata_sff_tf_read()
467 tf->hob_lbah = ioread8(ioaddr->lbah_addr); in ata_sff_tf_read()
468 iowrite8(tf->ctl, ioaddr->ctl_addr); in ata_sff_tf_read()
469 ap->last_ctl = tf->ctl; in ata_sff_tf_read()
487 void ata_sff_exec_command(struct ata_port *ap, const struct ata_taskfile *tf) in ata_sff_exec_command() argument
489 iowrite8(tf->command, ap->ioaddr.command_addr); in ata_sff_exec_command()
508 const struct ata_taskfile *tf, in ata_tf_to_host() argument
511 trace_ata_tf_load(ap, tf); in ata_tf_to_host()
512 ap->ops->sff_tf_load(ap, tf); in ata_tf_to_host()
513 trace_ata_exec_command(ap, tf, tag); in ata_tf_to_host()
514 ap->ops->sff_exec_command(ap, tf); in ata_tf_to_host()
637 bool do_write = (qc->tf.flags & ATA_TFLAG_WRITE); in ata_pio_xfer()
717 if (is_multi_taskfile(&qc->tf)) { in ata_pio_sectors()
754 switch (qc->tf.protocol) { in atapi_send_cdb()
765 trace_ata_bmdma_start(ap, &qc->tf, qc->tag); in atapi_send_cdb()
787 int rw = (qc->tf.flags & ATA_TFLAG_WRITE) ? WRITE : READ; in __atapi_pio_bytes()
860 int i_write, do_write = (qc->tf.flags & ATA_TFLAG_WRITE) ? 1 : 0; in atapi_pio_bytes()
911 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_hsm_ok_in_wq()
915 if (qc->tf.protocol == ATA_PROT_PIO && in ata_hsm_ok_in_wq()
916 (qc->tf.flags & ATA_TFLAG_WRITE)) in ata_hsm_ok_in_wq()
919 if (ata_is_atapi(qc->tf.protocol) && in ata_hsm_ok_in_wq()
1008 poll_next = (qc->tf.flags & ATA_TFLAG_POLLING); in ata_sff_hsm_move()
1049 if (qc->tf.protocol == ATA_PROT_PIO) { in ata_sff_hsm_move()
1071 if (qc->tf.protocol == ATAPI_PROT_PIO) { in ata_sff_hsm_move()
1148 if (!(qc->tf.flags & ATA_TFLAG_WRITE)) { in ata_sff_hsm_move()
1181 (!(qc->tf.flags & ATA_TFLAG_WRITE))) { in ata_sff_hsm_move()
1356 qc->tf.flags |= ATA_TFLAG_POLLING; in ata_sff_qc_issue()
1362 switch (qc->tf.protocol) { in ata_sff_qc_issue()
1364 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1367 ata_tf_to_host(ap, &qc->tf, qc->tag); in ata_sff_qc_issue()
1370 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1376 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1379 ata_tf_to_host(ap, &qc->tf, qc->tag); in ata_sff_qc_issue()
1381 if (qc->tf.flags & ATA_TFLAG_WRITE) { in ata_sff_qc_issue()
1393 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1406 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1409 ata_tf_to_host(ap, &qc->tf, qc->tag); in ata_sff_qc_issue()
1415 (qc->tf.flags & ATA_TFLAG_POLLING)) in ata_sff_qc_issue()
1550 if (!(qc->tf.flags & ATA_TFLAG_POLLING)) in __ata_sff_interrupt()
1643 if (!qc || qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_lost_interrupt()
1819 struct ata_taskfile tf; in ata_sff_dev_classify() local
1825 memset(&tf, 0, sizeof(tf)); in ata_sff_dev_classify()
1827 ap->ops->sff_tf_read(ap, &tf); in ata_sff_dev_classify()
1828 err = tf.error; in ata_sff_dev_classify()
1844 class = ata_port_classify(ap, &tf); in ata_sff_dev_classify()
2723 if (!ata_is_dma(qc->tf.protocol)) in ata_bmdma_qc_issue()
2730 switch (qc->tf.protocol) { in ata_bmdma_qc_issue()
2732 WARN_ON_ONCE(qc->tf.flags & ATA_TFLAG_POLLING); in ata_bmdma_qc_issue()
2734 trace_ata_tf_load(ap, &qc->tf); in ata_bmdma_qc_issue()
2735 ap->ops->sff_tf_load(ap, &qc->tf); /* load tf registers */ in ata_bmdma_qc_issue()
2736 trace_ata_bmdma_setup(ap, &qc->tf, qc->tag); in ata_bmdma_qc_issue()
2738 trace_ata_bmdma_start(ap, &qc->tf, qc->tag); in ata_bmdma_qc_issue()
2744 WARN_ON_ONCE(qc->tf.flags & ATA_TFLAG_POLLING); in ata_bmdma_qc_issue()
2746 trace_ata_tf_load(ap, &qc->tf); in ata_bmdma_qc_issue()
2747 ap->ops->sff_tf_load(ap, &qc->tf); /* load tf registers */ in ata_bmdma_qc_issue()
2748 trace_ata_bmdma_setup(ap, &qc->tf, qc->tag); in ata_bmdma_qc_issue()
2786 if (ap->hsm_task_state == HSM_ST_LAST && ata_is_dma(qc->tf.protocol)) { in ata_bmdma_port_intr()
2796 trace_ata_bmdma_stop(ap, &qc->tf, qc->tag); in ata_bmdma_port_intr()
2809 if (unlikely(qc->err_mask) && ata_is_dma(qc->tf.protocol)) in ata_bmdma_port_intr()
2861 if (qc && ata_is_dma(qc->tf.protocol)) { in ata_bmdma_error_handler()
2877 trace_ata_bmdma_stop(ap, &qc->tf, qc->tag); in ata_bmdma_error_handler()
2909 if (ata_is_dma(qc->tf.protocol)) { in ata_bmdma_post_internal_cmd()
2911 trace_ata_bmdma_stop(ap, &qc->tf, qc->tag); in ata_bmdma_post_internal_cmd()
2950 unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); in ata_bmdma_setup()
2965 ap->ops->sff_exec_command(ap, &qc->tf); in ata_bmdma_setup()