Lines Matching refs:rba

503 	struct iwl_rb_allocator *rba = &trans_pcie->rba;  in iwl_pcie_rx_allocator()  local
505 int pending = atomic_xchg(&rba->req_pending, 0); in iwl_pcie_rx_allocator()
510 spin_lock(&rba->lock); in iwl_pcie_rx_allocator()
512 list_replace_init(&rba->rbd_empty, &local_empty); in iwl_pcie_rx_allocator()
513 spin_unlock(&rba->lock); in iwl_pcie_rx_allocator()
562 pending = atomic_xchg(&rba->req_pending, 0); in iwl_pcie_rx_allocator()
568 spin_lock(&rba->lock); in iwl_pcie_rx_allocator()
570 list_splice_tail(&local_allocated, &rba->rbd_allocated); in iwl_pcie_rx_allocator()
572 list_splice_tail_init(&rba->rbd_empty, &local_empty); in iwl_pcie_rx_allocator()
573 spin_unlock(&rba->lock); in iwl_pcie_rx_allocator()
575 atomic_inc(&rba->req_ready); in iwl_pcie_rx_allocator()
578 spin_lock(&rba->lock); in iwl_pcie_rx_allocator()
580 list_splice_tail(&local_empty, &rba->rbd_empty); in iwl_pcie_rx_allocator()
581 spin_unlock(&rba->lock); in iwl_pcie_rx_allocator()
596 struct iwl_rb_allocator *rba = &trans_pcie->rba; in iwl_pcie_rx_allocator_get() local
609 if (atomic_dec_if_positive(&rba->req_ready) < 0) in iwl_pcie_rx_allocator_get()
612 spin_lock(&rba->lock); in iwl_pcie_rx_allocator_get()
616 list_first_entry(&rba->rbd_allocated, in iwl_pcie_rx_allocator_get()
621 spin_unlock(&rba->lock); in iwl_pcie_rx_allocator_get()
632 container_of(rba_p, struct iwl_trans_pcie, rba); in iwl_pcie_rx_allocator_work()
782 struct iwl_rb_allocator *rba = &trans_pcie->rba; in iwl_pcie_rx_alloc() local
793 spin_lock_init(&rba->lock); in iwl_pcie_rx_alloc()
1009 struct iwl_rb_allocator *rba = &trans_pcie->rba; in _iwl_pcie_rx_init() local
1019 cancel_work_sync(&rba->rx_alloc); in _iwl_pcie_rx_init()
1021 spin_lock(&rba->lock); in _iwl_pcie_rx_init()
1022 atomic_set(&rba->req_pending, 0); in _iwl_pcie_rx_init()
1023 atomic_set(&rba->req_ready, 0); in _iwl_pcie_rx_init()
1024 INIT_LIST_HEAD(&rba->rbd_allocated); in _iwl_pcie_rx_init()
1025 INIT_LIST_HEAD(&rba->rbd_empty); in _iwl_pcie_rx_init()
1026 spin_unlock(&rba->lock); in _iwl_pcie_rx_init()
1073 list_add(&rxb->list, &rba->rbd_empty); in _iwl_pcie_rx_init()
1120 struct iwl_rb_allocator *rba = &trans_pcie->rba; in iwl_pcie_rx_free() local
1132 cancel_work_sync(&rba->rx_alloc); in iwl_pcie_rx_free()
1158 struct iwl_rb_allocator *rba = &trans_pcie->rba; in iwl_pcie_rx_reuse_rbd() local
1178 spin_lock(&rba->lock); in iwl_pcie_rx_reuse_rbd()
1179 list_splice_tail_init(&rxq->rx_used, &rba->rbd_empty); in iwl_pcie_rx_reuse_rbd()
1180 spin_unlock(&rba->lock); in iwl_pcie_rx_reuse_rbd()
1182 atomic_inc(&rba->req_pending); in iwl_pcie_rx_reuse_rbd()
1183 queue_work(rba->alloc_wq, &rba->rx_alloc); in iwl_pcie_rx_reuse_rbd()
1424 struct iwl_rb_allocator *rba = &trans_pcie->rba; in iwl_pcie_rx_handle() local
1427 spin_lock(&rba->lock); in iwl_pcie_rx_handle()
1428 list_splice_tail_init(&rxq->rx_used, &rba->rbd_empty); in iwl_pcie_rx_handle()
1429 spin_unlock(&rba->lock); in iwl_pcie_rx_handle()