Lines Matching refs:queue
36 struct data_queue *queue = entry->queue; in rt2x00queue_alloc_rxskb() local
37 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev; in rt2x00queue_alloc_rxskb()
48 frame_size = queue->data_size + queue->desc_size + queue->winfo_size; in rt2x00queue_alloc_rxskb()
106 struct device *dev = entry->queue->rt2x00dev->dev; in rt2x00queue_map_txskb()
122 struct device *dev = entry->queue->rt2x00dev->dev; in rt2x00queue_unmap_skb()
497 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; in rt2x00queue_write_tx_data()
509 entry->queue->qid, DRV_PROJECT); in rt2x00queue_write_tx_data()
538 struct data_queue *queue = entry->queue; in rt2x00queue_write_tx_descriptor() local
540 queue->rt2x00dev->ops->lib->write_tx_desc(entry, txdesc); in rt2x00queue_write_tx_descriptor()
546 rt2x00debug_dump_frame(queue->rt2x00dev, DUMP_FRAME_TX, entry); in rt2x00queue_write_tx_descriptor()
549 static void rt2x00queue_kick_tx_queue(struct data_queue *queue, in rt2x00queue_kick_tx_queue() argument
561 if (rt2x00queue_threshold(queue) || in rt2x00queue_kick_tx_queue()
563 queue->rt2x00dev->ops->lib->kick_queue(queue); in rt2x00queue_kick_tx_queue()
568 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; in rt2x00queue_bar_check()
608 int rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb, in rt2x00queue_write_tx_frame() argument
623 rt2x00queue_create_tx_descriptor(queue->rt2x00dev, skb, &txdesc, sta); in rt2x00queue_write_tx_frame()
648 if (rt2x00_has_cap_flag(queue->rt2x00dev, REQUIRE_COPY_IV)) in rt2x00queue_write_tx_frame()
662 if (rt2x00_has_cap_flag(queue->rt2x00dev, REQUIRE_L2PAD)) in rt2x00queue_write_tx_frame()
664 else if (rt2x00_has_cap_flag(queue->rt2x00dev, REQUIRE_DMA)) in rt2x00queue_write_tx_frame()
670 spin_lock(&queue->tx_lock); in rt2x00queue_write_tx_frame()
672 if (unlikely(rt2x00queue_full(queue))) { in rt2x00queue_write_tx_frame()
673 rt2x00_err(queue->rt2x00dev, "Dropping frame due to full tx queue %d\n", in rt2x00queue_write_tx_frame()
674 queue->qid); in rt2x00queue_write_tx_frame()
679 entry = rt2x00queue_get_entry(queue, Q_INDEX); in rt2x00queue_write_tx_frame()
683 rt2x00_err(queue->rt2x00dev, in rt2x00queue_write_tx_frame()
686 queue->qid, DRV_PROJECT); in rt2x00queue_write_tx_frame()
714 rt2x00queue_kick_tx_queue(queue, &txdesc); in rt2x00queue_write_tx_frame()
722 if (rt2x00queue_threshold(queue)) in rt2x00queue_write_tx_frame()
723 rt2x00queue_pause_queue(queue); in rt2x00queue_write_tx_frame()
725 spin_unlock(&queue->tx_lock); in rt2x00queue_write_tx_frame()
793 bool rt2x00queue_for_each_entry(struct data_queue *queue, in rt2x00queue_for_each_entry() argument
806 rt2x00_err(queue->rt2x00dev, in rt2x00queue_for_each_entry()
818 spin_lock_irqsave(&queue->index_lock, irqflags); in rt2x00queue_for_each_entry()
819 index_start = queue->index[start]; in rt2x00queue_for_each_entry()
820 index_end = queue->index[end]; in rt2x00queue_for_each_entry()
821 spin_unlock_irqrestore(&queue->index_lock, irqflags); in rt2x00queue_for_each_entry()
829 if (fn(&queue->entries[i], data)) in rt2x00queue_for_each_entry()
833 for (i = index_start; i < queue->limit; i++) { in rt2x00queue_for_each_entry()
834 if (fn(&queue->entries[i], data)) in rt2x00queue_for_each_entry()
839 if (fn(&queue->entries[i], data)) in rt2x00queue_for_each_entry()
848 struct queue_entry *rt2x00queue_get_entry(struct data_queue *queue, in rt2x00queue_get_entry() argument
855 rt2x00_err(queue->rt2x00dev, "Entry requested from invalid index type (%d)\n", in rt2x00queue_get_entry()
860 spin_lock_irqsave(&queue->index_lock, irqflags); in rt2x00queue_get_entry()
862 entry = &queue->entries[queue->index[index]]; in rt2x00queue_get_entry()
864 spin_unlock_irqrestore(&queue->index_lock, irqflags); in rt2x00queue_get_entry()
872 struct data_queue *queue = entry->queue; in rt2x00queue_index_inc() local
876 rt2x00_err(queue->rt2x00dev, in rt2x00queue_index_inc()
881 spin_lock_irqsave(&queue->index_lock, irqflags); in rt2x00queue_index_inc()
883 queue->index[index]++; in rt2x00queue_index_inc()
884 if (queue->index[index] >= queue->limit) in rt2x00queue_index_inc()
885 queue->index[index] = 0; in rt2x00queue_index_inc()
890 queue->length++; in rt2x00queue_index_inc()
892 queue->length--; in rt2x00queue_index_inc()
893 queue->count++; in rt2x00queue_index_inc()
896 spin_unlock_irqrestore(&queue->index_lock, irqflags); in rt2x00queue_index_inc()
899 static void rt2x00queue_pause_queue_nocheck(struct data_queue *queue) in rt2x00queue_pause_queue_nocheck() argument
901 switch (queue->qid) { in rt2x00queue_pause_queue_nocheck()
910 ieee80211_stop_queue(queue->rt2x00dev->hw, queue->qid); in rt2x00queue_pause_queue_nocheck()
916 void rt2x00queue_pause_queue(struct data_queue *queue) in rt2x00queue_pause_queue() argument
918 if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) || in rt2x00queue_pause_queue()
919 !test_bit(QUEUE_STARTED, &queue->flags) || in rt2x00queue_pause_queue()
920 test_and_set_bit(QUEUE_PAUSED, &queue->flags)) in rt2x00queue_pause_queue()
923 rt2x00queue_pause_queue_nocheck(queue); in rt2x00queue_pause_queue()
927 void rt2x00queue_unpause_queue(struct data_queue *queue) in rt2x00queue_unpause_queue() argument
929 if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) || in rt2x00queue_unpause_queue()
930 !test_bit(QUEUE_STARTED, &queue->flags) || in rt2x00queue_unpause_queue()
931 !test_and_clear_bit(QUEUE_PAUSED, &queue->flags)) in rt2x00queue_unpause_queue()
934 switch (queue->qid) { in rt2x00queue_unpause_queue()
943 ieee80211_wake_queue(queue->rt2x00dev->hw, queue->qid); in rt2x00queue_unpause_queue()
950 queue->rt2x00dev->ops->lib->kick_queue(queue); in rt2x00queue_unpause_queue()
957 void rt2x00queue_start_queue(struct data_queue *queue) in rt2x00queue_start_queue() argument
959 mutex_lock(&queue->status_lock); in rt2x00queue_start_queue()
961 if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) || in rt2x00queue_start_queue()
962 test_and_set_bit(QUEUE_STARTED, &queue->flags)) { in rt2x00queue_start_queue()
963 mutex_unlock(&queue->status_lock); in rt2x00queue_start_queue()
967 set_bit(QUEUE_PAUSED, &queue->flags); in rt2x00queue_start_queue()
969 queue->rt2x00dev->ops->lib->start_queue(queue); in rt2x00queue_start_queue()
971 rt2x00queue_unpause_queue(queue); in rt2x00queue_start_queue()
973 mutex_unlock(&queue->status_lock); in rt2x00queue_start_queue()
977 void rt2x00queue_stop_queue(struct data_queue *queue) in rt2x00queue_stop_queue() argument
979 mutex_lock(&queue->status_lock); in rt2x00queue_stop_queue()
981 if (!test_and_clear_bit(QUEUE_STARTED, &queue->flags)) { in rt2x00queue_stop_queue()
982 mutex_unlock(&queue->status_lock); in rt2x00queue_stop_queue()
986 rt2x00queue_pause_queue_nocheck(queue); in rt2x00queue_stop_queue()
988 queue->rt2x00dev->ops->lib->stop_queue(queue); in rt2x00queue_stop_queue()
990 mutex_unlock(&queue->status_lock); in rt2x00queue_stop_queue()
994 void rt2x00queue_flush_queue(struct data_queue *queue, bool drop) in rt2x00queue_flush_queue() argument
997 (queue->qid == QID_AC_VO) || in rt2x00queue_flush_queue()
998 (queue->qid == QID_AC_VI) || in rt2x00queue_flush_queue()
999 (queue->qid == QID_AC_BE) || in rt2x00queue_flush_queue()
1000 (queue->qid == QID_AC_BK); in rt2x00queue_flush_queue()
1002 if (rt2x00queue_empty(queue)) in rt2x00queue_flush_queue()
1012 queue->rt2x00dev->ops->lib->kick_queue(queue); in rt2x00queue_flush_queue()
1019 if (likely(queue->rt2x00dev->ops->lib->flush_queue)) in rt2x00queue_flush_queue()
1020 queue->rt2x00dev->ops->lib->flush_queue(queue, drop); in rt2x00queue_flush_queue()
1025 if (unlikely(!rt2x00queue_empty(queue))) in rt2x00queue_flush_queue()
1026 rt2x00_warn(queue->rt2x00dev, "Queue %d failed to flush\n", in rt2x00queue_flush_queue()
1027 queue->qid); in rt2x00queue_flush_queue()
1033 struct data_queue *queue; in rt2x00queue_start_queues() local
1039 tx_queue_for_each(rt2x00dev, queue) in rt2x00queue_start_queues()
1040 rt2x00queue_start_queue(queue); in rt2x00queue_start_queues()
1048 struct data_queue *queue; in rt2x00queue_stop_queues() local
1058 tx_queue_for_each(rt2x00dev, queue) in rt2x00queue_stop_queues()
1059 rt2x00queue_stop_queue(queue); in rt2x00queue_stop_queues()
1067 struct data_queue *queue; in rt2x00queue_flush_queues() local
1069 tx_queue_for_each(rt2x00dev, queue) in rt2x00queue_flush_queues()
1070 rt2x00queue_flush_queue(queue, drop); in rt2x00queue_flush_queues()
1076 static void rt2x00queue_reset(struct data_queue *queue) in rt2x00queue_reset() argument
1081 spin_lock_irqsave(&queue->index_lock, irqflags); in rt2x00queue_reset()
1083 queue->count = 0; in rt2x00queue_reset()
1084 queue->length = 0; in rt2x00queue_reset()
1087 queue->index[i] = 0; in rt2x00queue_reset()
1089 spin_unlock_irqrestore(&queue->index_lock, irqflags); in rt2x00queue_reset()
1094 struct data_queue *queue; in rt2x00queue_init_queues() local
1097 queue_for_each(rt2x00dev, queue) { in rt2x00queue_init_queues()
1098 rt2x00queue_reset(queue); in rt2x00queue_init_queues()
1100 for (i = 0; i < queue->limit; i++) in rt2x00queue_init_queues()
1101 rt2x00dev->ops->lib->clear_entry(&queue->entries[i]); in rt2x00queue_init_queues()
1105 static int rt2x00queue_alloc_entries(struct data_queue *queue) in rt2x00queue_alloc_entries() argument
1111 rt2x00queue_reset(queue); in rt2x00queue_alloc_entries()
1116 entry_size = sizeof(*entries) + queue->priv_size; in rt2x00queue_alloc_entries()
1117 entries = kcalloc(queue->limit, entry_size, GFP_KERNEL); in rt2x00queue_alloc_entries()
1125 for (i = 0; i < queue->limit; i++) { in rt2x00queue_alloc_entries()
1127 entries[i].queue = queue; in rt2x00queue_alloc_entries()
1131 QUEUE_ENTRY_PRIV_OFFSET(entries, i, queue->limit, in rt2x00queue_alloc_entries()
1132 sizeof(*entries), queue->priv_size); in rt2x00queue_alloc_entries()
1137 queue->entries = entries; in rt2x00queue_alloc_entries()
1142 static void rt2x00queue_free_skbs(struct data_queue *queue) in rt2x00queue_free_skbs() argument
1146 if (!queue->entries) in rt2x00queue_free_skbs()
1149 for (i = 0; i < queue->limit; i++) { in rt2x00queue_free_skbs()
1150 rt2x00queue_free_skb(&queue->entries[i]); in rt2x00queue_free_skbs()
1154 static int rt2x00queue_alloc_rxskbs(struct data_queue *queue) in rt2x00queue_alloc_rxskbs() argument
1159 for (i = 0; i < queue->limit; i++) { in rt2x00queue_alloc_rxskbs()
1160 skb = rt2x00queue_alloc_rxskb(&queue->entries[i], GFP_KERNEL); in rt2x00queue_alloc_rxskbs()
1163 queue->entries[i].skb = skb; in rt2x00queue_alloc_rxskbs()
1171 struct data_queue *queue; in rt2x00queue_initialize() local
1178 tx_queue_for_each(rt2x00dev, queue) { in rt2x00queue_initialize()
1179 status = rt2x00queue_alloc_entries(queue); in rt2x00queue_initialize()
1210 struct data_queue *queue; in rt2x00queue_uninitialize() local
1214 queue_for_each(rt2x00dev, queue) { in rt2x00queue_uninitialize()
1215 kfree(queue->entries); in rt2x00queue_uninitialize()
1216 queue->entries = NULL; in rt2x00queue_uninitialize()
1221 struct data_queue *queue, enum data_queue_qid qid) in rt2x00queue_init() argument
1223 mutex_init(&queue->status_lock); in rt2x00queue_init()
1224 spin_lock_init(&queue->tx_lock); in rt2x00queue_init()
1225 spin_lock_init(&queue->index_lock); in rt2x00queue_init()
1227 queue->rt2x00dev = rt2x00dev; in rt2x00queue_init()
1228 queue->qid = qid; in rt2x00queue_init()
1229 queue->txop = 0; in rt2x00queue_init()
1230 queue->aifs = 2; in rt2x00queue_init()
1231 queue->cw_min = 5; in rt2x00queue_init()
1232 queue->cw_max = 10; in rt2x00queue_init()
1234 rt2x00dev->ops->queue_init(queue); in rt2x00queue_init()
1236 queue->threshold = DIV_ROUND_UP(queue->limit, 10); in rt2x00queue_init()
1241 struct data_queue *queue; in rt2x00queue_allocate() local
1255 queue = kcalloc(rt2x00dev->data_queues, sizeof(*queue), GFP_KERNEL); in rt2x00queue_allocate()
1256 if (!queue) in rt2x00queue_allocate()
1262 rt2x00dev->rx = queue; in rt2x00queue_allocate()
1263 rt2x00dev->tx = &queue[1]; in rt2x00queue_allocate()
1264 rt2x00dev->bcn = &queue[1 + rt2x00dev->ops->tx_queues]; in rt2x00queue_allocate()
1265 rt2x00dev->atim = req_atim ? &queue[2 + rt2x00dev->ops->tx_queues] : NULL; in rt2x00queue_allocate()
1279 tx_queue_for_each(rt2x00dev, queue) in rt2x00queue_allocate()
1280 rt2x00queue_init(rt2x00dev, queue, qid++); in rt2x00queue_allocate()