Lines Matching refs:hrrq
710 struct ipr_cmnd *__ipr_get_free_ipr_cmnd(struct ipr_hrr_queue *hrrq) in __ipr_get_free_ipr_cmnd() argument
714 if (likely(!list_empty(&hrrq->hrrq_free_q))) { in __ipr_get_free_ipr_cmnd()
715 ipr_cmd = list_entry(hrrq->hrrq_free_q.next, in __ipr_get_free_ipr_cmnd()
735 __ipr_get_free_ipr_cmnd(&ioa_cfg->hrrq[IPR_INIT_HRRQ]); in ipr_get_free_ipr_cmnd()
759 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_mask_and_clear_interrupts()
760 ioa_cfg->hrrq[i].allow_interrupts = 0; in ipr_mask_and_clear_interrupts()
761 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_mask_and_clear_interrupts()
843 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in __ipr_sata_eh_done()
858 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_sata_eh_done() local
861 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_sata_eh_done()
863 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_sata_eh_done()
886 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in __ipr_scsi_eh_done()
902 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_scsi_eh_done() local
904 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_scsi_eh_done()
906 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_scsi_eh_done()
921 struct ipr_hrr_queue *hrrq; in ipr_fail_all_ops() local
924 for_each_hrrq(hrrq, ioa_cfg) { in ipr_fail_all_ops()
925 spin_lock(&hrrq->_lock); in ipr_fail_all_ops()
927 temp, &hrrq->hrrq_pending_q, queue) { in ipr_fail_all_ops()
945 spin_unlock(&hrrq->_lock); in ipr_fail_all_ops()
996 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_do_req()
1098 unsigned int hrrq; in ipr_get_hrrq_index() local
1101 hrrq = 0; in ipr_get_hrrq_index()
1103 hrrq = atomic_add_return(1, &ioa_cfg->hrrq_index); in ipr_get_hrrq_index()
1104 hrrq = (hrrq % (ioa_cfg->hrrq_num - 1)) + 1; in ipr_get_hrrq_index()
1106 return hrrq; in ipr_get_hrrq_index()
1128 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) { in ipr_send_hcam()
1130 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_send_hcam()
1519 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_process_ccn()
2663 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_process_error()
3354 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) { in ipr_add_remove_thread()
3663 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in ipr_show_adapter_state()
3695 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead && in ipr_store_adapter_state()
3698 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_store_adapter_state()
3699 ioa_cfg->hrrq[i].ioa_is_dead = 0; in ipr_store_adapter_state()
3700 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_store_adapter_state()
3823 irq_poll_disable(&ioa_cfg->hrrq[i].iopoll); in ipr_store_iopoll_weight()
3830 irq_poll_init(&ioa_cfg->hrrq[i].iopoll, in ipr_store_iopoll_weight()
4401 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead && !ioa_cfg->dump_taken) { in ipr_alloc_dump()
5083 list_for_each_entry(loop_cmd, &ipr_cmd->hrrq->hrrq_free_q, queue) { in ipr_cmnd_is_free()
5123 struct ipr_hrr_queue *hrrq; in ipr_wait_for_ops() local
5131 for_each_hrrq(hrrq, ioa_cfg) { in ipr_wait_for_ops()
5132 spin_lock_irqsave(hrrq->lock, flags); in ipr_wait_for_ops()
5133 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in ipr_wait_for_ops()
5142 spin_unlock_irqrestore(hrrq->lock, flags); in ipr_wait_for_ops()
5151 for_each_hrrq(hrrq, ioa_cfg) { in ipr_wait_for_ops()
5152 spin_lock_irqsave(hrrq->lock, flags); in ipr_wait_for_ops()
5153 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in ipr_wait_for_ops()
5162 spin_unlock_irqrestore(hrrq->lock, flags); in ipr_wait_for_ops()
5187 if (!ioa_cfg->in_reset_reload && !ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { in ipr_eh_host_reset()
5202 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { in ipr_eh_host_reset()
5257 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_device_reset()
5337 struct ipr_hrr_queue *hrrq; in __ipr_eh_dev_reset() local
5350 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in __ipr_eh_dev_reset()
5353 for_each_hrrq(hrrq, ioa_cfg) { in __ipr_eh_dev_reset()
5354 spin_lock(&hrrq->_lock); in __ipr_eh_dev_reset()
5355 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in __ipr_eh_dev_reset()
5371 spin_unlock(&hrrq->_lock); in __ipr_eh_dev_reset()
5448 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_bus_reset_done()
5510 struct ipr_hrr_queue *hrrq; in ipr_cancel_op() local
5521 ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in ipr_cancel_op()
5536 for_each_hrrq(hrrq, ioa_cfg) { in ipr_cancel_op()
5537 spin_lock(&hrrq->_lock); in ipr_cancel_op()
5538 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in ipr_cancel_op()
5546 spin_unlock(&hrrq->_lock); in ipr_cancel_op()
5573 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_cancel_op()
5595 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead || ioa_cfg->scan_done) in ipr_scan_finished()
5777 struct ipr_hrr_queue *hrrq; in ipr_iopoll() local
5783 hrrq = container_of(iop, struct ipr_hrr_queue, iopoll); in ipr_iopoll()
5784 ioa_cfg = hrrq->ioa_cfg; in ipr_iopoll()
5786 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_iopoll()
5787 completed_ops = ipr_process_hrrq(hrrq, budget, &doneq); in ipr_iopoll()
5791 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_iopoll()
5812 struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp; in ipr_isr() local
5813 struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg; in ipr_isr()
5822 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_isr()
5824 if (!hrrq->allow_interrupts) { in ipr_isr()
5825 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr()
5830 if (ipr_process_hrrq(hrrq, -1, &doneq)) { in ipr_isr()
5861 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr()
5880 struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp; in ipr_isr_mhrrq() local
5881 struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg; in ipr_isr_mhrrq()
5887 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
5890 if (!hrrq->allow_interrupts) { in ipr_isr_mhrrq()
5891 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
5896 if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) == in ipr_isr_mhrrq()
5897 hrrq->toggle_bit) { in ipr_isr_mhrrq()
5898 irq_poll_sched(&hrrq->iopoll); in ipr_isr_mhrrq()
5899 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
5903 if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) == in ipr_isr_mhrrq()
5904 hrrq->toggle_bit) in ipr_isr_mhrrq()
5906 if (ipr_process_hrrq(hrrq, -1, &doneq)) in ipr_isr_mhrrq()
5910 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
6067 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in __ipr_erp_done()
6082 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_erp_done() local
6085 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_erp_done()
6087 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_erp_done()
6169 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_erp_request_sense() local
6172 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_erp_request_sense()
6174 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_erp_request_sense()
6505 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_erp_start()
6530 spin_lock_irqsave(ipr_cmd->hrrq->lock, lock_flags); in ipr_scsi_done()
6534 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_scsi_done()
6535 spin_unlock_irqrestore(ipr_cmd->hrrq->lock, lock_flags); in ipr_scsi_done()
6538 spin_lock(&ipr_cmd->hrrq->_lock); in ipr_scsi_done()
6540 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_scsi_done()
6566 struct ipr_hrr_queue *hrrq; in ipr_queuecommand() local
6582 hrrq = &ioa_cfg->hrrq[hrrq_id]; in ipr_queuecommand()
6584 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6590 if (unlikely(!hrrq->allow_cmds && !hrrq->ioa_is_dead && !hrrq->removing_ioa)) { in ipr_queuecommand()
6591 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6599 if (unlikely(hrrq->ioa_is_dead || hrrq->removing_ioa || !res)) { in ipr_queuecommand()
6600 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6604 ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq); in ipr_queuecommand()
6606 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6609 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6654 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6655 if (unlikely(rc || (!hrrq->allow_cmds && !hrrq->ioa_is_dead))) { in ipr_queuecommand()
6656 list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q); in ipr_queuecommand()
6657 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6663 if (unlikely(hrrq->ioa_is_dead)) { in ipr_queuecommand()
6664 list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q); in ipr_queuecommand()
6665 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6675 list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_pending_q); in ipr_queuecommand()
6678 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6682 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6686 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6784 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) in ipr_ata_phy_reset()
6815 struct ipr_hrr_queue *hrrq; in ipr_ata_post_internal() local
6825 for_each_hrrq(hrrq, ioa_cfg) { in ipr_ata_post_internal()
6826 spin_lock(&hrrq->_lock); in ipr_ata_post_internal()
6827 list_for_each_entry(ipr_cmd, &hrrq->hrrq_pending_q, queue) { in ipr_ata_post_internal()
6833 spin_unlock(&hrrq->_lock); in ipr_ata_post_internal()
6882 spin_lock(&ipr_cmd->hrrq->_lock); in ipr_sata_done()
6898 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_sata_done()
6899 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_sata_done()
7007 struct ipr_hrr_queue *hrrq; in ipr_qc_defer() local
7011 hrrq = &ioa_cfg->hrrq[hrrq_id]; in ipr_qc_defer()
7014 spin_lock(&hrrq->_lock); in ipr_qc_defer()
7015 if (unlikely(hrrq->ioa_is_dead)) { in ipr_qc_defer()
7016 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7020 if (unlikely(!hrrq->allow_cmds)) { in ipr_qc_defer()
7021 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7025 ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq); in ipr_qc_defer()
7027 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7032 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7061 spin_lock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7062 if (unlikely(!ipr_cmd->hrrq->allow_cmds || in ipr_qc_issue()
7063 ipr_cmd->hrrq->ioa_is_dead)) { in ipr_qc_issue()
7064 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_qc_issue()
7065 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7081 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_qc_issue()
7121 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7126 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7235 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { in ipr_ioa_bringdown_done()
7244 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_ioa_bringdown_done()
7245 ioa_cfg->hrrq[i].ioa_is_dead = 1; in ipr_ioa_bringdown_done()
7246 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_ioa_bringdown_done()
7250 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_ioa_bringdown_done()
7277 spin_lock(&ioa_cfg->hrrq[j]._lock); in ipr_ioa_reset_done()
7278 ioa_cfg->hrrq[j].allow_cmds = 1; in ipr_ioa_reset_done()
7279 spin_unlock(&ioa_cfg->hrrq[j]._lock); in ipr_ioa_reset_done()
7309 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_ioa_reset_done()
7644 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_reset_cmd_failed()
8152 &ioa_cfg->hrrq->hrrq_free_q); in ipr_ioafp_page0_inquiry()
8205 struct ipr_hrr_queue *hrrq; in ipr_ioafp_identify_hrrq() local
8213 hrrq = &ioa_cfg->hrrq[ioa_cfg->identify_hrrq_index]; in ipr_ioafp_identify_hrrq()
8228 ((u64) hrrq->host_rrq_dma >> 24) & 0xff; in ipr_ioafp_identify_hrrq()
8230 ((u64) hrrq->host_rrq_dma >> 16) & 0xff; in ipr_ioafp_identify_hrrq()
8232 ((u64) hrrq->host_rrq_dma >> 8) & 0xff; in ipr_ioafp_identify_hrrq()
8234 ((u64) hrrq->host_rrq_dma) & 0xff; in ipr_ioafp_identify_hrrq()
8236 ((sizeof(u32) * hrrq->size) >> 8) & 0xff; in ipr_ioafp_identify_hrrq()
8238 (sizeof(u32) * hrrq->size) & 0xff; in ipr_ioafp_identify_hrrq()
8246 ((u64) hrrq->host_rrq_dma >> 56) & 0xff; in ipr_ioafp_identify_hrrq()
8248 ((u64) hrrq->host_rrq_dma >> 48) & 0xff; in ipr_ioafp_identify_hrrq()
8250 ((u64) hrrq->host_rrq_dma >> 40) & 0xff; in ipr_ioafp_identify_hrrq()
8252 ((u64) hrrq->host_rrq_dma >> 32) & 0xff; in ipr_ioafp_identify_hrrq()
8321 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_start_timer()
8338 struct ipr_hrr_queue *hrrq; in ipr_init_ioa_mem() local
8340 for_each_hrrq(hrrq, ioa_cfg) { in ipr_init_ioa_mem()
8341 spin_lock(&hrrq->_lock); in ipr_init_ioa_mem()
8342 memset(hrrq->host_rrq, 0, sizeof(u32) * hrrq->size); in ipr_init_ioa_mem()
8345 hrrq->hrrq_start = hrrq->host_rrq; in ipr_init_ioa_mem()
8346 hrrq->hrrq_end = &hrrq->host_rrq[hrrq->size - 1]; in ipr_init_ioa_mem()
8347 hrrq->hrrq_curr = hrrq->hrrq_start; in ipr_init_ioa_mem()
8348 hrrq->toggle_bit = 1; in ipr_init_ioa_mem()
8349 spin_unlock(&hrrq->_lock); in ipr_init_ioa_mem()
8414 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_next_stage()
8441 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_enable_ioa()
8442 ioa_cfg->hrrq[i].allow_interrupts = 1; in ipr_reset_enable_ioa()
8443 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_enable_ioa()
8483 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_enable_ioa()
9005 struct ipr_hrr_queue *hrrq; in ipr_reset_cancel_hcam_done() local
9012 for_each_hrrq(hrrq, ioa_cfg) { in ipr_reset_cancel_hcam_done()
9013 spin_lock(&hrrq->_lock); in ipr_reset_cancel_hcam_done()
9014 list_for_each_entry(loop_cmd, &hrrq->hrrq_pending_q, queue) { in ipr_reset_cancel_hcam_done()
9017 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_reset_cancel_hcam_done()
9021 spin_unlock(&hrrq->_lock); in ipr_reset_cancel_hcam_done()
9046 struct ipr_hrr_queue *hrrq = &ioa_cfg->hrrq[IPR_INIT_HRRQ]; in ipr_reset_cancel_hcam() local
9051 if (!hrrq->ioa_is_dead) { in ipr_reset_cancel_hcam()
9053 list_for_each_entry(hcam_cmd, &hrrq->hrrq_pending_q, queue) { in ipr_reset_cancel_hcam()
9172 !ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { in ipr_reset_shutdown_ioa()
9221 &ipr_cmd->hrrq->hrrq_free_q); in ipr_reset_ioa_job()
9260 spin_lock(&ioa_cfg->hrrq[i]._lock); in _ipr_initiate_ioa_reset()
9261 ioa_cfg->hrrq[i].allow_cmds = 0; in _ipr_initiate_ioa_reset()
9262 spin_unlock(&ioa_cfg->hrrq[i]._lock); in _ipr_initiate_ioa_reset()
9265 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { in _ipr_initiate_ioa_reset()
9296 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in ipr_initiate_ioa_reset()
9312 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_initiate_ioa_reset()
9313 ioa_cfg->hrrq[i].ioa_is_dead = 1; in ipr_initiate_ioa_reset()
9314 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_initiate_ioa_reset()
9324 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { in ipr_initiate_ioa_reset()
9354 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_freeze()
9355 ioa_cfg->hrrq[i].allow_interrupts = 0; in ipr_reset_freeze()
9356 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_freeze()
9359 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_freeze()
9448 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_pci_perm_failure()
9449 ioa_cfg->hrrq[i].allow_cmds = 0; in ipr_pci_perm_failure()
9450 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_pci_perm_failure()
9568 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_free_mem()
9569 ioa_cfg->hrrq[i].host_rrq, in ipr_free_mem()
9570 ioa_cfg->hrrq[i].host_rrq_dma); in ipr_free_mem()
9602 free_irq(pci_irq_vector(pdev, i), &ioa_cfg->hrrq[i]); in ipr_free_irqs()
9664 ioa_cfg->hrrq[i].min_cmd_id = 0; in ipr_alloc_cmd_blks()
9665 ioa_cfg->hrrq[i].max_cmd_id = in ipr_alloc_cmd_blks()
9671 ioa_cfg->hrrq[i].min_cmd_id = in ipr_alloc_cmd_blks()
9674 ioa_cfg->hrrq[i].max_cmd_id = in ipr_alloc_cmd_blks()
9680 ioa_cfg->hrrq[i].min_cmd_id = 0; in ipr_alloc_cmd_blks()
9681 ioa_cfg->hrrq[i].max_cmd_id = (entries_each_hrrq - 1); in ipr_alloc_cmd_blks()
9683 ioa_cfg->hrrq[i].size = entries_each_hrrq; in ipr_alloc_cmd_blks()
9689 ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id - 1; in ipr_alloc_cmd_blks()
9691 ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].size += i; in ipr_alloc_cmd_blks()
9692 ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id += i; in ipr_alloc_cmd_blks()
9734 ipr_cmd->hrrq = &ioa_cfg->hrrq[hrrq_id]; in ipr_alloc_cmd_blks()
9735 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_alloc_cmd_blks()
9736 if (i >= ioa_cfg->hrrq[hrrq_id].max_cmd_id) in ipr_alloc_cmd_blks()
9780 ioa_cfg->hrrq[i].host_rrq = dma_alloc_coherent(&pdev->dev, in ipr_alloc_mem()
9781 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_alloc_mem()
9782 &ioa_cfg->hrrq[i].host_rrq_dma, in ipr_alloc_mem()
9785 if (!ioa_cfg->hrrq[i].host_rrq) { in ipr_alloc_mem()
9788 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_alloc_mem()
9789 ioa_cfg->hrrq[i].host_rrq, in ipr_alloc_mem()
9790 ioa_cfg->hrrq[i].host_rrq_dma); in ipr_alloc_mem()
9793 ioa_cfg->hrrq[i].ioa_cfg = ioa_cfg; in ipr_alloc_mem()
9842 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_alloc_mem()
9843 ioa_cfg->hrrq[i].host_rrq, in ipr_alloc_mem()
9844 ioa_cfg->hrrq[i].host_rrq_dma); in ipr_alloc_mem()
9984 for (i = 0; i < ARRAY_SIZE(ioa_cfg->hrrq); i++) { in ipr_init_ioa_cfg()
9985 INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_free_q); in ipr_init_ioa_cfg()
9986 INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_pending_q); in ipr_init_ioa_cfg()
9987 spin_lock_init(&ioa_cfg->hrrq[i]._lock); in ipr_init_ioa_cfg()
9989 ioa_cfg->hrrq[i].lock = ioa_cfg->host->host_lock; in ipr_init_ioa_cfg()
9991 ioa_cfg->hrrq[i].lock = &ioa_cfg->hrrq[i]._lock; in ipr_init_ioa_cfg()
10056 &ioa_cfg->hrrq[i]); in ipr_request_other_msi_irqs()
10060 &ioa_cfg->hrrq[i]); in ipr_request_other_msi_irqs()
10375 &ioa_cfg->hrrq[0]); in ipr_probe_ioa()
10381 IPR_NAME, &ioa_cfg->hrrq[0]); in ipr_probe_ioa()
10482 spin_lock(&ioa_cfg->hrrq[i]._lock); in __ipr_remove()
10483 ioa_cfg->hrrq[i].removing_ioa = 1; in __ipr_remove()
10484 spin_unlock(&ioa_cfg->hrrq[i]._lock); in __ipr_remove()
10613 irq_poll_init(&ioa_cfg->hrrq[i].iopoll, in ipr_probe()
10644 irq_poll_disable(&ioa_cfg->hrrq[i].iopoll); in ipr_shutdown()
10802 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_halt_done()
10824 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds || in ipr_halt()