Lines Matching refs:txqs

47 		WARN_ON(trans->txqs.bc_table_dword);  in iwl_pcie_gen2_update_byte_tbl()
55 WARN_ON(!trans->txqs.bc_table_dword); in iwl_pcie_gen2_update_byte_tbl()
93 if (num_tbs > trans->txqs.tfd.max_tbs) { in iwl_txq_gen2_tfd_unmap()
165 if (le16_to_cpu(tfd->num_tbs) >= trans->txqs.tfd.max_tbs) { in iwl_txq_gen2_set_tb()
167 trans->txqs.tfd.max_tbs); in iwl_txq_gen2_set_tb()
185 page_ptr = (void *)((u8 *)skb->cb + trans->txqs.page_offs); in get_workaround_page()
283 struct iwl_tso_hdr_page *p = this_cpu_ptr(trans->txqs.tso_hdr_page); in get_page_hdr()
286 page_ptr = (void *)((u8 *)skb->cb + trans->txqs.page_offs); in get_page_hdr()
708 struct iwl_txq *txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_tx()
717 if (WARN_ONCE(!test_bit(txq_id, trans->txqs.queue_used), in iwl_txq_gen2_tx()
736 trans->txqs.dev_cmd_offs); in iwl_txq_gen2_tx()
803 struct iwl_txq *txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_unmap()
810 if (txq_id != trans->txqs.cmd.q_id) { in iwl_txq_gen2_unmap()
841 trans->txqs.tfd.size * txq->n_window, in iwl_txq_gen2_free_memory()
850 dma_pool_free(trans->txqs.bc_pool, in iwl_txq_gen2_free_memory()
872 txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_free()
880 if (txq_id == trans->txqs.cmd.q_id) in iwl_txq_gen2_free()
889 trans->txqs.txq[txq_id] = NULL; in iwl_txq_gen2_free()
891 clear_bit(txq_id, trans->txqs.queue_used); in iwl_txq_gen2_free()
959 page_ptr = (void *)((u8 *)skb->cb + trans->txqs.page_offs); in iwl_txq_free_tso_page()
1023 size_t tfd_sz = trans->txqs.tfd.size * in iwl_txq_alloc()
1032 tfd_sz = trans->txqs.tfd.size * slots_num; in iwl_txq_alloc()
1093 WARN_ON(!trans->txqs.bc_tbl_size); in iwl_txq_dyn_alloc_dma()
1095 bc_tbl_size = trans->txqs.bc_tbl_size; in iwl_txq_dyn_alloc_dma()
1105 txq->bc_tbl.addr = dma_pool_alloc(trans->txqs.bc_pool, GFP_KERNEL, in iwl_txq_dyn_alloc_dma()
1151 if (qid >= ARRAY_SIZE(trans->txqs.txq)) { in iwl_txq_alloc_response()
1157 if (test_and_set_bit(qid, trans->txqs.queue_used)) { in iwl_txq_alloc_response()
1163 if (WARN_ONCE(trans->txqs.txq[qid], in iwl_txq_alloc_response()
1170 trans->txqs.txq[qid] = txq; in iwl_txq_alloc_response()
1236 if (!test_and_clear_bit(queue, trans->txqs.queue_used)) { in iwl_txq_dyn_free()
1251 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_txq_gen2_tx_free()
1254 for (i = 0; i < ARRAY_SIZE(trans->txqs.txq); i++) { in iwl_txq_gen2_tx_free()
1255 if (!trans->txqs.txq[i]) in iwl_txq_gen2_tx_free()
1268 if (!trans->txqs.txq[txq_id]) { in iwl_txq_gen2_init()
1274 trans->txqs.txq[txq_id] = queue; in iwl_txq_gen2_init()
1281 queue = trans->txqs.txq[txq_id]; in iwl_txq_gen2_init()
1285 (txq_id == trans->txqs.cmd.q_id)); in iwl_txq_gen2_init()
1290 trans->txqs.txq[txq_id]->id = txq_id; in iwl_txq_gen2_init()
1291 set_bit(txq_id, trans->txqs.queue_used); in iwl_txq_gen2_init()
1342 if (num_tbs > trans->txqs.tfd.max_tbs) { in iwl_txq_gen1_tfd_unmap()
1400 scd_bc_tbl = trans->txqs.scd_bc_tbls.addr; in iwl_txq_gen1_update_byte_cnt_tbl()
1415 if (trans->txqs.bc_table_dword) in iwl_txq_gen1_update_byte_cnt_tbl()
1433 struct iwlagn_scd_bc_tbl *scd_bc_tbl = trans->txqs.scd_bc_tbls.addr; in iwl_txq_gen1_inval_byte_cnt_tbl()
1443 if (txq_id != trans->txqs.cmd.q_id) in iwl_txq_gen1_inval_byte_cnt_tbl()
1524 struct iwl_txq *txq = trans->txqs.txq[txq_id]; in iwl_txq_reclaim()
1530 if (WARN_ON(txq_id == trans->txqs.cmd.q_id)) in iwl_txq_reclaim()
1535 if (!test_bit(txq_id, trans->txqs.queue_used)) { in iwl_txq_reclaim()
1591 test_bit(txq_id, trans->txqs.queue_stopped)) { in iwl_txq_reclaim()
1620 trans->txqs.dev_cmd_offs); in iwl_txq_reclaim()
1644 struct iwl_txq *txq = trans->txqs.txq[txq_id]; in iwl_txq_set_q_ptrs()
1654 void iwl_trans_txq_freeze_timer(struct iwl_trans *trans, unsigned long txqs, in iwl_trans_txq_freeze_timer() argument
1659 for_each_set_bit(queue, &txqs, BITS_PER_LONG) { in iwl_trans_txq_freeze_timer()
1660 struct iwl_txq *txq = trans->txqs.txq[queue]; in iwl_trans_txq_freeze_timer()
1712 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_trans_txq_send_hcmd_sync()