Lines Matching refs:tf

342 void ata_sff_tf_load(struct ata_port *ap, const struct ata_taskfile *tf)  in ata_sff_tf_load()  argument
345 unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR; in ata_sff_tf_load()
347 if (tf->ctl != ap->last_ctl) { in ata_sff_tf_load()
349 iowrite8(tf->ctl, ioaddr->ctl_addr); in ata_sff_tf_load()
350 ap->last_ctl = tf->ctl; in ata_sff_tf_load()
354 if (is_addr && (tf->flags & ATA_TFLAG_LBA48)) { in ata_sff_tf_load()
356 iowrite8(tf->hob_feature, ioaddr->feature_addr); in ata_sff_tf_load()
357 iowrite8(tf->hob_nsect, ioaddr->nsect_addr); in ata_sff_tf_load()
358 iowrite8(tf->hob_lbal, ioaddr->lbal_addr); in ata_sff_tf_load()
359 iowrite8(tf->hob_lbam, ioaddr->lbam_addr); in ata_sff_tf_load()
360 iowrite8(tf->hob_lbah, ioaddr->lbah_addr); in ata_sff_tf_load()
364 iowrite8(tf->feature, ioaddr->feature_addr); in ata_sff_tf_load()
365 iowrite8(tf->nsect, ioaddr->nsect_addr); in ata_sff_tf_load()
366 iowrite8(tf->lbal, ioaddr->lbal_addr); in ata_sff_tf_load()
367 iowrite8(tf->lbam, ioaddr->lbam_addr); in ata_sff_tf_load()
368 iowrite8(tf->lbah, ioaddr->lbah_addr); in ata_sff_tf_load()
371 if (tf->flags & ATA_TFLAG_DEVICE) in ata_sff_tf_load()
372 iowrite8(tf->device, ioaddr->device_addr); in ata_sff_tf_load()
391 void ata_sff_tf_read(struct ata_port *ap, struct ata_taskfile *tf) in ata_sff_tf_read() argument
395 tf->status = ata_sff_check_status(ap); in ata_sff_tf_read()
396 tf->error = ioread8(ioaddr->error_addr); in ata_sff_tf_read()
397 tf->nsect = ioread8(ioaddr->nsect_addr); in ata_sff_tf_read()
398 tf->lbal = ioread8(ioaddr->lbal_addr); in ata_sff_tf_read()
399 tf->lbam = ioread8(ioaddr->lbam_addr); in ata_sff_tf_read()
400 tf->lbah = ioread8(ioaddr->lbah_addr); in ata_sff_tf_read()
401 tf->device = ioread8(ioaddr->device_addr); in ata_sff_tf_read()
403 if (tf->flags & ATA_TFLAG_LBA48) { in ata_sff_tf_read()
405 iowrite8(tf->ctl | ATA_HOB, ioaddr->ctl_addr); in ata_sff_tf_read()
406 tf->hob_feature = ioread8(ioaddr->error_addr); in ata_sff_tf_read()
407 tf->hob_nsect = ioread8(ioaddr->nsect_addr); in ata_sff_tf_read()
408 tf->hob_lbal = ioread8(ioaddr->lbal_addr); in ata_sff_tf_read()
409 tf->hob_lbam = ioread8(ioaddr->lbam_addr); in ata_sff_tf_read()
410 tf->hob_lbah = ioread8(ioaddr->lbah_addr); in ata_sff_tf_read()
411 iowrite8(tf->ctl, ioaddr->ctl_addr); in ata_sff_tf_read()
412 ap->last_ctl = tf->ctl; in ata_sff_tf_read()
430 void ata_sff_exec_command(struct ata_port *ap, const struct ata_taskfile *tf) in ata_sff_exec_command() argument
432 iowrite8(tf->command, ap->ioaddr.command_addr); in ata_sff_exec_command()
451 const struct ata_taskfile *tf, in ata_tf_to_host() argument
454 trace_ata_tf_load(ap, tf); in ata_tf_to_host()
455 ap->ops->sff_tf_load(ap, tf); in ata_tf_to_host()
456 trace_ata_exec_command(ap, tf, tag); in ata_tf_to_host()
457 ap->ops->sff_exec_command(ap, tf); in ata_tf_to_host()
580 bool do_write = (qc->tf.flags & ATA_TFLAG_WRITE); in ata_pio_xfer()
662 if (is_multi_taskfile(&qc->tf)) { in ata_pio_sectors()
699 switch (qc->tf.protocol) { in atapi_send_cdb()
710 trace_ata_bmdma_start(ap, &qc->tf, qc->tag); in atapi_send_cdb()
732 int rw = (qc->tf.flags & ATA_TFLAG_WRITE) ? WRITE : READ; in __atapi_pio_bytes()
805 int i_write, do_write = (qc->tf.flags & ATA_TFLAG_WRITE) ? 1 : 0; in atapi_pio_bytes()
856 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_hsm_ok_in_wq()
860 if (qc->tf.protocol == ATA_PROT_PIO && in ata_hsm_ok_in_wq()
861 (qc->tf.flags & ATA_TFLAG_WRITE)) in ata_hsm_ok_in_wq()
864 if (ata_is_atapi(qc->tf.protocol) && in ata_hsm_ok_in_wq()
943 poll_next = (qc->tf.flags & ATA_TFLAG_POLLING); in ata_sff_hsm_move()
984 if (qc->tf.protocol == ATA_PROT_PIO) { in ata_sff_hsm_move()
1006 if (qc->tf.protocol == ATAPI_PROT_PIO) { in ata_sff_hsm_move()
1083 if (!(qc->tf.flags & ATA_TFLAG_WRITE)) { in ata_sff_hsm_move()
1116 (!(qc->tf.flags & ATA_TFLAG_WRITE))) { in ata_sff_hsm_move()
1291 qc->tf.flags |= ATA_TFLAG_POLLING; in ata_sff_qc_issue()
1297 switch (qc->tf.protocol) { in ata_sff_qc_issue()
1299 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1302 ata_tf_to_host(ap, &qc->tf, qc->tag); in ata_sff_qc_issue()
1305 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1311 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1314 ata_tf_to_host(ap, &qc->tf, qc->tag); in ata_sff_qc_issue()
1316 if (qc->tf.flags & ATA_TFLAG_WRITE) { in ata_sff_qc_issue()
1328 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1341 if (qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_qc_issue()
1344 ata_tf_to_host(ap, &qc->tf, qc->tag); in ata_sff_qc_issue()
1350 (qc->tf.flags & ATA_TFLAG_POLLING)) in ata_sff_qc_issue()
1481 if (!(qc->tf.flags & ATA_TFLAG_POLLING)) in __ata_sff_interrupt()
1574 if (!qc || qc->tf.flags & ATA_TFLAG_POLLING) in ata_sff_lost_interrupt()
1750 struct ata_taskfile tf; in ata_sff_dev_classify() local
1756 memset(&tf, 0, sizeof(tf)); in ata_sff_dev_classify()
1758 ap->ops->sff_tf_read(ap, &tf); in ata_sff_dev_classify()
1759 err = tf.error; in ata_sff_dev_classify()
1775 class = ata_port_classify(ap, &tf); in ata_sff_dev_classify()
2646 if (!ata_is_dma(qc->tf.protocol)) in ata_bmdma_qc_issue()
2653 switch (qc->tf.protocol) { in ata_bmdma_qc_issue()
2655 WARN_ON_ONCE(qc->tf.flags & ATA_TFLAG_POLLING); in ata_bmdma_qc_issue()
2657 trace_ata_tf_load(ap, &qc->tf); in ata_bmdma_qc_issue()
2658 ap->ops->sff_tf_load(ap, &qc->tf); /* load tf registers */ in ata_bmdma_qc_issue()
2659 trace_ata_bmdma_setup(ap, &qc->tf, qc->tag); in ata_bmdma_qc_issue()
2661 trace_ata_bmdma_start(ap, &qc->tf, qc->tag); in ata_bmdma_qc_issue()
2667 WARN_ON_ONCE(qc->tf.flags & ATA_TFLAG_POLLING); in ata_bmdma_qc_issue()
2669 trace_ata_tf_load(ap, &qc->tf); in ata_bmdma_qc_issue()
2670 ap->ops->sff_tf_load(ap, &qc->tf); /* load tf registers */ in ata_bmdma_qc_issue()
2671 trace_ata_bmdma_setup(ap, &qc->tf, qc->tag); in ata_bmdma_qc_issue()
2709 if (ap->hsm_task_state == HSM_ST_LAST && ata_is_dma(qc->tf.protocol)) { in ata_bmdma_port_intr()
2719 trace_ata_bmdma_stop(ap, &qc->tf, qc->tag); in ata_bmdma_port_intr()
2732 if (unlikely(qc->err_mask) && ata_is_dma(qc->tf.protocol)) in ata_bmdma_port_intr()
2784 if (qc && ata_is_dma(qc->tf.protocol)) { in ata_bmdma_error_handler()
2800 trace_ata_bmdma_stop(ap, &qc->tf, qc->tag); in ata_bmdma_error_handler()
2832 if (ata_is_dma(qc->tf.protocol)) { in ata_bmdma_post_internal_cmd()
2834 trace_ata_bmdma_stop(ap, &qc->tf, qc->tag); in ata_bmdma_post_internal_cmd()
2873 unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); in ata_bmdma_setup()
2888 ap->ops->sff_exec_command(ap, &qc->tf); in ata_bmdma_setup()