Lines Matching refs:pp
607 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_register_mode() local
608 void __iomem *mmio = pp->ctl_block; in nv_adma_register_mode()
612 if (pp->flags & NV_ADMA_PORT_REGISTER_MODE) in nv_adma_register_mode()
640 pp->flags |= NV_ADMA_PORT_REGISTER_MODE; in nv_adma_register_mode()
645 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_mode() local
646 void __iomem *mmio = pp->ctl_block; in nv_adma_mode()
650 if (!(pp->flags & NV_ADMA_PORT_REGISTER_MODE)) in nv_adma_mode()
653 WARN_ON(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE); in nv_adma_mode()
670 pp->flags &= ~NV_ADMA_PORT_REGISTER_MODE; in nv_adma_mode()
676 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_slave_config() local
727 pp->flags &= ~NV_ADMA_ATAPI_SETUP_COMPLETE; in nv_adma_slave_config()
730 pp->flags |= NV_ADMA_ATAPI_SETUP_COMPLETE; in nv_adma_slave_config()
749 rc = dma_set_mask(&pdev->dev, pp->adma_dma_mask); in nv_adma_slave_config()
766 struct nv_adma_port_priv *pp = qc->ap->private_data; in nv_adma_check_atapi_dma() local
767 return !(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE); in nv_adma_check_atapi_dma()
818 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_check_cpb() local
819 u8 flags = pp->cpb[cpb_num].resp_flags; in nv_adma_check_cpb()
895 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_interrupt() local
896 void __iomem *mmio = pp->ctl_block; in nv_adma_interrupt()
904 if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) { in nv_adma_interrupt()
912 if (pp->flags & NV_ADMA_PORT_REGISTER_MODE) { in nv_adma_interrupt()
928 gen_ctl = readl(pp->gen_block + NV_ADMA_GEN_CTL); in nv_adma_interrupt()
1010 struct nv_adma_port_priv *pp = host->ports[0]->private_data; in nv_adma_interrupt() local
1011 writel(notifier_clears[0], pp->notifier_clear_block); in nv_adma_interrupt()
1012 pp = host->ports[1]->private_data; in nv_adma_interrupt()
1013 writel(notifier_clears[1], pp->notifier_clear_block); in nv_adma_interrupt()
1023 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_freeze() local
1024 void __iomem *mmio = pp->ctl_block; in nv_adma_freeze()
1029 if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) in nv_adma_freeze()
1045 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_thaw() local
1046 void __iomem *mmio = pp->ctl_block; in nv_adma_thaw()
1051 if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) in nv_adma_thaw()
1063 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_irq_clear() local
1064 void __iomem *mmio = pp->ctl_block; in nv_adma_irq_clear()
1067 if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) { in nv_adma_irq_clear()
1088 pp = ap->host->ports[0]->private_data; in nv_adma_irq_clear()
1089 writel(notifier_clears[0], pp->notifier_clear_block); in nv_adma_irq_clear()
1090 pp = ap->host->ports[1]->private_data; in nv_adma_irq_clear()
1091 writel(notifier_clears[1], pp->notifier_clear_block); in nv_adma_irq_clear()
1096 struct nv_adma_port_priv *pp = qc->ap->private_data; in nv_adma_post_internal_cmd() local
1098 if (pp->flags & NV_ADMA_PORT_REGISTER_MODE) in nv_adma_post_internal_cmd()
1105 struct nv_adma_port_priv *pp; in nv_adma_port_start() local
1128 pp = devm_kzalloc(dev, sizeof(*pp), GFP_KERNEL); in nv_adma_port_start()
1129 if (!pp) in nv_adma_port_start()
1134 pp->ctl_block = mmio; in nv_adma_port_start()
1135 pp->gen_block = ap->host->iomap[NV_MMIO_BAR] + NV_ADMA_GEN; in nv_adma_port_start()
1136 pp->notifier_clear_block = pp->gen_block + in nv_adma_port_start()
1149 pp->adma_dma_mask = *dev->dma_mask; in nv_adma_port_start()
1162 pp->cpb = mem; in nv_adma_port_start()
1163 pp->cpb_dma = mem_dma; in nv_adma_port_start()
1174 pp->aprd = mem; in nv_adma_port_start()
1175 pp->aprd_dma = mem_dma; in nv_adma_port_start()
1177 ap->private_data = pp; in nv_adma_port_start()
1183 pp->flags = NV_ADMA_PORT_REGISTER_MODE; in nv_adma_port_start()
1205 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_port_stop() local
1206 void __iomem *mmio = pp->ctl_block; in nv_adma_port_stop()
1215 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_port_suspend() local
1216 void __iomem *mmio = pp->ctl_block; in nv_adma_port_suspend()
1232 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_port_resume() local
1233 void __iomem *mmio = pp->ctl_block; in nv_adma_port_resume()
1237 writel(pp->cpb_dma & 0xFFFFFFFF, mmio + NV_ADMA_CPB_BASE_LOW); in nv_adma_port_resume()
1238 writel((pp->cpb_dma >> 16) >> 16, mmio + NV_ADMA_CPB_BASE_HIGH); in nv_adma_port_resume()
1244 pp->flags |= NV_ADMA_PORT_REGISTER_MODE; in nv_adma_port_resume()
1333 struct nv_adma_port_priv *pp = qc->ap->private_data; in nv_adma_fill_sg() local
1342 &pp->aprd[NV_ADMA_SGTBL_LEN * qc->hw_tag + (si-5)]; in nv_adma_fill_sg()
1346 cpb->next_aprd = cpu_to_le64(((u64)(pp->aprd_dma + NV_ADMA_SGTBL_SZ * qc->hw_tag))); in nv_adma_fill_sg()
1353 struct nv_adma_port_priv *pp = qc->ap->private_data; in nv_adma_use_reg_mode() local
1357 if ((pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) || in nv_adma_use_reg_mode()
1370 struct nv_adma_port_priv *pp = qc->ap->private_data; in nv_adma_qc_prep() local
1371 struct nv_adma_cpb *cpb = &pp->cpb[qc->hw_tag]; in nv_adma_qc_prep()
1376 BUG_ON(!(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) && in nv_adma_qc_prep()
1416 struct nv_adma_port_priv *pp = qc->ap->private_data; in nv_adma_qc_issue() local
1417 void __iomem *mmio = pp->ctl_block; in nv_adma_qc_issue()
1434 BUG_ON(!(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) && in nv_adma_qc_issue()
1445 if (curr_ncq != pp->last_issue_ncq) { in nv_adma_qc_issue()
1449 pp->last_issue_ncq = curr_ncq; in nv_adma_qc_issue()
1655 struct nv_adma_port_priv *pp = ap->private_data; in nv_adma_error_handler() local
1656 if (!(pp->flags & NV_ADMA_PORT_REGISTER_MODE)) { in nv_adma_error_handler()
1657 void __iomem *mmio = pp->ctl_block; in nv_adma_error_handler()
1664 u32 gen_ctl = readl(pp->gen_block + NV_ADMA_GEN_CTL); in nv_adma_error_handler()
1677 struct nv_adma_cpb *cpb = &pp->cpb[i]; in nv_adma_error_handler()
1692 pp->cpb[i].ctl_flags &= ~NV_CPB_CTL_CPB_VALID; in nv_adma_error_handler()
1711 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_qc_to_dq() local
1712 struct defer_queue *dq = &pp->defer_queue; in nv_swncq_qc_to_dq()
1722 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_qc_from_dq() local
1723 struct defer_queue *dq = &pp->defer_queue; in nv_swncq_qc_from_dq()
1739 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_fis_reinit() local
1741 pp->dhfis_bits = 0; in nv_swncq_fis_reinit()
1742 pp->dmafis_bits = 0; in nv_swncq_fis_reinit()
1743 pp->sdbfis_bits = 0; in nv_swncq_fis_reinit()
1744 pp->ncq_flags = 0; in nv_swncq_fis_reinit()
1749 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_pp_reinit() local
1750 struct defer_queue *dq = &pp->defer_queue; in nv_swncq_pp_reinit()
1755 pp->qc_active = 0; in nv_swncq_pp_reinit()
1756 pp->last_issue_tag = ATA_TAG_POISON; in nv_swncq_pp_reinit()
1762 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_irq_clear() local
1764 writew(fis, pp->irq_block); in nv_swncq_irq_clear()
1777 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_ncq_stop() local
1787 pp->qc_active, pp->defer_queue.defer_bits, pp->last_issue_tag, in nv_swncq_ncq_stop()
1788 pp->dhfis_bits, pp->dmafis_bits, pp->sdbfis_bits); in nv_swncq_ncq_stop()
1794 sactive = readl(pp->sactive_block); in nv_swncq_ncq_stop()
1795 done_mask = pp->qc_active ^ sactive; in nv_swncq_ncq_stop()
1800 if (pp->qc_active & (1 << i)) in nv_swncq_ncq_stop()
1809 (pp->dhfis_bits >> i) & 0x1, in nv_swncq_ncq_stop()
1810 (pp->dmafis_bits >> i) & 0x1, in nv_swncq_ncq_stop()
1811 (pp->sdbfis_bits >> i) & 0x1, in nv_swncq_ncq_stop()
1949 struct nv_swncq_port_priv *pp; in nv_swncq_port_start() local
1957 pp = devm_kzalloc(dev, sizeof(*pp), GFP_KERNEL); in nv_swncq_port_start()
1958 if (!pp) in nv_swncq_port_start()
1961 pp->prd = dmam_alloc_coherent(dev, ATA_PRD_TBL_SZ * ATA_MAX_QUEUE, in nv_swncq_port_start()
1962 &pp->prd_dma, GFP_KERNEL); in nv_swncq_port_start()
1963 if (!pp->prd) in nv_swncq_port_start()
1965 memset(pp->prd, 0, ATA_PRD_TBL_SZ * ATA_MAX_QUEUE); in nv_swncq_port_start()
1967 ap->private_data = pp; in nv_swncq_port_start()
1968 pp->sactive_block = ap->ioaddr.scr_addr + 4 * SCR_ACTIVE; in nv_swncq_port_start()
1969 pp->irq_block = mmio + NV_INT_STATUS_MCP55 + ap->port_no * 2; in nv_swncq_port_start()
1970 pp->tag_block = mmio + NV_NCQ_REG_MCP55 + ap->port_no * 2; in nv_swncq_port_start()
1992 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_fill_sg() local
1996 prd = pp->prd + ATA_MAX_PRD * qc->hw_tag; in nv_swncq_fill_sg()
2027 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_issue_atacmd() local
2034 writel((1 << qc->hw_tag), pp->sactive_block); in nv_swncq_issue_atacmd()
2035 pp->last_issue_tag = qc->hw_tag; in nv_swncq_issue_atacmd()
2036 pp->dhfis_bits &= ~(1 << qc->hw_tag); in nv_swncq_issue_atacmd()
2037 pp->dmafis_bits &= ~(1 << qc->hw_tag); in nv_swncq_issue_atacmd()
2038 pp->qc_active |= (0x1 << qc->hw_tag); in nv_swncq_issue_atacmd()
2051 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_qc_issue() local
2058 if (!pp->qc_active) in nv_swncq_qc_issue()
2094 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_sdbfis() local
2114 sactive = readl(pp->sactive_block); in nv_swncq_sdbfis()
2115 done_mask = pp->qc_active ^ sactive; in nv_swncq_sdbfis()
2117 pp->qc_active &= ~done_mask; in nv_swncq_sdbfis()
2118 pp->dhfis_bits &= ~done_mask; in nv_swncq_sdbfis()
2119 pp->dmafis_bits &= ~done_mask; in nv_swncq_sdbfis()
2120 pp->sdbfis_bits |= done_mask; in nv_swncq_sdbfis()
2129 if (pp->qc_active & pp->dhfis_bits) in nv_swncq_sdbfis()
2132 if ((pp->ncq_flags & ncq_saw_backout) || in nv_swncq_sdbfis()
2133 (pp->qc_active ^ pp->dhfis_bits)) in nv_swncq_sdbfis()
2142 ap->print_id, ap->qc_active, pp->qc_active, in nv_swncq_sdbfis()
2143 pp->defer_queue.defer_bits, pp->dhfis_bits, in nv_swncq_sdbfis()
2144 pp->dmafis_bits, pp->last_issue_tag); in nv_swncq_sdbfis()
2149 qc = ata_qc_from_tag(ap, pp->last_issue_tag); in nv_swncq_sdbfis()
2154 if (pp->defer_queue.defer_bits) { in nv_swncq_sdbfis()
2166 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_tag() local
2169 tag = readb(pp->tag_block) >> 2; in nv_swncq_tag()
2179 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_dmafis() local
2193 iowrite32(pp->prd_dma + ATA_PRD_TBL_SZ * qc->hw_tag, in nv_swncq_dmafis()
2207 struct nv_swncq_port_priv *pp = ap->private_data; in nv_swncq_host_interrupt() local
2226 if (!pp->qc_active) in nv_swncq_host_interrupt()
2247 pp->ncq_flags |= ncq_saw_backout; in nv_swncq_host_interrupt()
2251 pp->ncq_flags |= ncq_saw_sdb; in nv_swncq_host_interrupt()
2254 ap->print_id, pp->qc_active, pp->dhfis_bits, in nv_swncq_host_interrupt()
2255 pp->dmafis_bits, readl(pp->sactive_block)); in nv_swncq_host_interrupt()
2264 pp->dhfis_bits |= (0x1 << pp->last_issue_tag); in nv_swncq_host_interrupt()
2265 pp->ncq_flags |= ncq_saw_d2h; in nv_swncq_host_interrupt()
2266 if (pp->ncq_flags & (ncq_saw_sdb | ncq_saw_backout)) { in nv_swncq_host_interrupt()
2274 !(pp->ncq_flags & ncq_saw_dmas)) { in nv_swncq_host_interrupt()
2279 if (pp->defer_queue.defer_bits) { in nv_swncq_host_interrupt()
2291 pp->dmafis_bits |= (0x1 << nv_swncq_tag(ap)); in nv_swncq_host_interrupt()
2292 pp->ncq_flags |= ncq_saw_dmas; in nv_swncq_host_interrupt()
2438 struct nv_adma_port_priv *pp; in nv_pci_device_resume() local
2442 pp = host->ports[0]->private_data; in nv_pci_device_resume()
2443 if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) in nv_pci_device_resume()
2449 pp = host->ports[1]->private_data; in nv_pci_device_resume()
2450 if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) in nv_pci_device_resume()