Lines Matching refs:txqs

74 	memset(trans->txqs.queue_stopped, 0,  in iwl_txq_gen2_tx_stop()
75 sizeof(trans->txqs.queue_stopped)); in iwl_txq_gen2_tx_stop()
76 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_txq_gen2_tx_stop()
79 for (txq_id = 0; txq_id < ARRAY_SIZE(trans->txqs.txq); txq_id++) { in iwl_txq_gen2_tx_stop()
80 if (!trans->txqs.txq[txq_id]) in iwl_txq_gen2_tx_stop()
117 WARN_ON(trans->txqs.bc_table_dword); in iwl_pcie_gen2_update_byte_tbl()
125 WARN_ON(!trans->txqs.bc_table_dword); in iwl_pcie_gen2_update_byte_tbl()
163 if (num_tbs > trans->txqs.tfd.max_tbs) { in iwl_txq_gen2_tfd_unmap()
236 if (le16_to_cpu(tfd->num_tbs) >= trans->txqs.tfd.max_tbs) { in iwl_txq_gen2_set_tb()
238 trans->txqs.tfd.max_tbs); in iwl_txq_gen2_set_tb()
256 page_ptr = (void *)((u8 *)skb->cb + trans->txqs.page_offs); in get_workaround_page()
354 struct iwl_tso_hdr_page *p = this_cpu_ptr(trans->txqs.tso_hdr_page); in get_page_hdr()
357 page_ptr = (void *)((u8 *)skb->cb + trans->txqs.page_offs); in get_page_hdr()
784 struct iwl_txq *txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_tx()
793 if (WARN_ONCE(!test_bit(txq_id, trans->txqs.queue_used), in iwl_txq_gen2_tx()
812 trans->txqs.dev_cmd_offs); in iwl_txq_gen2_tx()
879 struct iwl_txq *txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_unmap()
886 if (txq_id != trans->txqs.cmd.q_id) { in iwl_txq_gen2_unmap()
919 trans->txqs.tfd.size * txq->n_window, in iwl_txq_gen2_free_memory()
928 dma_pool_free(trans->txqs.bc_pool, in iwl_txq_gen2_free_memory()
950 txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_free()
958 if (txq_id == trans->txqs.cmd.q_id) in iwl_txq_gen2_free()
967 trans->txqs.txq[txq_id] = NULL; in iwl_txq_gen2_free()
969 clear_bit(txq_id, trans->txqs.queue_used); in iwl_txq_gen2_free()
1037 page_ptr = (void *)((u8 *)skb->cb + trans->txqs.page_offs); in iwl_txq_free_tso_page()
1101 size_t tfd_sz = trans->txqs.tfd.size * in iwl_txq_alloc()
1110 tfd_sz = trans->txqs.tfd.size * slots_num; in iwl_txq_alloc()
1171 WARN_ON(!trans->txqs.bc_tbl_size); in iwl_txq_dyn_alloc_dma()
1173 bc_tbl_size = trans->txqs.bc_tbl_size; in iwl_txq_dyn_alloc_dma()
1183 txq->bc_tbl.addr = dma_pool_alloc(trans->txqs.bc_pool, GFP_KERNEL, in iwl_txq_dyn_alloc_dma()
1229 if (qid >= ARRAY_SIZE(trans->txqs.txq)) { in iwl_txq_alloc_response()
1235 if (test_and_set_bit(qid, trans->txqs.queue_used)) { in iwl_txq_alloc_response()
1242 trans->txqs.txq[qid] = txq; in iwl_txq_alloc_response()
1308 if (!test_and_clear_bit(queue, trans->txqs.queue_used)) { in iwl_txq_dyn_free()
1316 iwl_txq_gen2_free_memory(trans, trans->txqs.txq[queue]); in iwl_txq_dyn_free()
1318 trans->txqs.txq[queue] = NULL; in iwl_txq_dyn_free()
1327 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_txq_gen2_tx_free()
1330 for (i = 0; i < ARRAY_SIZE(trans->txqs.txq); i++) { in iwl_txq_gen2_tx_free()
1331 if (!trans->txqs.txq[i]) in iwl_txq_gen2_tx_free()
1344 if (!trans->txqs.txq[txq_id]) { in iwl_txq_gen2_init()
1350 trans->txqs.txq[txq_id] = queue; in iwl_txq_gen2_init()
1357 queue = trans->txqs.txq[txq_id]; in iwl_txq_gen2_init()
1361 (txq_id == trans->txqs.cmd.q_id)); in iwl_txq_gen2_init()
1366 trans->txqs.txq[txq_id]->id = txq_id; in iwl_txq_gen2_init()
1367 set_bit(txq_id, trans->txqs.queue_used); in iwl_txq_gen2_init()
1418 if (num_tbs > trans->txqs.tfd.max_tbs) { in iwl_txq_gen1_tfd_unmap()
1476 scd_bc_tbl = trans->txqs.scd_bc_tbls.addr; in iwl_txq_gen1_update_byte_cnt_tbl()
1491 if (trans->txqs.bc_table_dword) in iwl_txq_gen1_update_byte_cnt_tbl()
1509 struct iwlagn_scd_bc_tbl *scd_bc_tbl = trans->txqs.scd_bc_tbls.addr; in iwl_txq_gen1_inval_byte_cnt_tbl()
1519 if (txq_id != trans->txqs.cmd.q_id) in iwl_txq_gen1_inval_byte_cnt_tbl()