1 // SPDX-License-Identifier: ISC
2 /*
3 * Copyright (c) 2005-2011 Atheros Communications Inc.
4 * Copyright (c) 2011-2017 Qualcomm Atheros, Inc.
5 */
6
7 #include "core.h"
8 #include "hif.h"
9 #include "debug.h"
10
11 /********/
12 /* Send */
13 /********/
14
ath10k_htc_control_tx_complete(struct ath10k * ar,struct sk_buff * skb)15 static void ath10k_htc_control_tx_complete(struct ath10k *ar,
16 struct sk_buff *skb)
17 {
18 kfree_skb(skb);
19 }
20
ath10k_htc_build_tx_ctrl_skb(void * ar)21 static struct sk_buff *ath10k_htc_build_tx_ctrl_skb(void *ar)
22 {
23 struct sk_buff *skb;
24 struct ath10k_skb_cb *skb_cb;
25
26 skb = dev_alloc_skb(ATH10K_HTC_CONTROL_BUFFER_SIZE);
27 if (!skb)
28 return NULL;
29
30 skb_reserve(skb, 20); /* FIXME: why 20 bytes? */
31 WARN_ONCE((unsigned long)skb->data & 3, "unaligned skb");
32
33 skb_cb = ATH10K_SKB_CB(skb);
34 memset(skb_cb, 0, sizeof(*skb_cb));
35
36 ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: skb %pK\n", __func__, skb);
37 return skb;
38 }
39
ath10k_htc_restore_tx_skb(struct ath10k_htc * htc,struct sk_buff * skb)40 static inline void ath10k_htc_restore_tx_skb(struct ath10k_htc *htc,
41 struct sk_buff *skb)
42 {
43 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
44
45 if (htc->ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
46 dma_unmap_single(htc->ar->dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
47 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
48 }
49
ath10k_htc_notify_tx_completion(struct ath10k_htc_ep * ep,struct sk_buff * skb)50 void ath10k_htc_notify_tx_completion(struct ath10k_htc_ep *ep,
51 struct sk_buff *skb)
52 {
53 struct ath10k *ar = ep->htc->ar;
54 struct ath10k_htc_hdr *hdr;
55
56 ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: ep %d skb %pK\n", __func__,
57 ep->eid, skb);
58
59 hdr = (struct ath10k_htc_hdr *)skb->data;
60 ath10k_htc_restore_tx_skb(ep->htc, skb);
61
62 if (!ep->ep_ops.ep_tx_complete) {
63 ath10k_warn(ar, "no tx handler for eid %d\n", ep->eid);
64 dev_kfree_skb_any(skb);
65 return;
66 }
67
68 if (hdr->flags & ATH10K_HTC_FLAG_SEND_BUNDLE) {
69 dev_kfree_skb_any(skb);
70 return;
71 }
72
73 ep->ep_ops.ep_tx_complete(ep->htc->ar, skb);
74 }
75 EXPORT_SYMBOL(ath10k_htc_notify_tx_completion);
76
ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep * ep,struct sk_buff * skb)77 static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
78 struct sk_buff *skb)
79 {
80 struct ath10k_htc_hdr *hdr;
81
82 hdr = (struct ath10k_htc_hdr *)skb->data;
83 memset(hdr, 0, sizeof(struct ath10k_htc_hdr));
84
85 hdr->eid = ep->eid;
86 hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr));
87 hdr->flags = 0;
88 if (ep->tx_credit_flow_enabled && !ep->bundle_tx)
89 hdr->flags |= ATH10K_HTC_FLAG_NEED_CREDIT_UPDATE;
90
91 spin_lock_bh(&ep->htc->tx_lock);
92 hdr->seq_no = ep->seq_no++;
93 spin_unlock_bh(&ep->htc->tx_lock);
94 }
95
ath10k_htc_consume_credit(struct ath10k_htc_ep * ep,unsigned int len,bool consume)96 static int ath10k_htc_consume_credit(struct ath10k_htc_ep *ep,
97 unsigned int len,
98 bool consume)
99 {
100 struct ath10k_htc *htc = ep->htc;
101 struct ath10k *ar = htc->ar;
102 enum ath10k_htc_ep_id eid = ep->eid;
103 int credits, ret = 0;
104
105 if (!ep->tx_credit_flow_enabled)
106 return 0;
107
108 credits = DIV_ROUND_UP(len, ep->tx_credit_size);
109 spin_lock_bh(&htc->tx_lock);
110
111 if (ep->tx_credits < credits) {
112 ath10k_dbg(ar, ATH10K_DBG_HTC,
113 "htc insufficient credits ep %d required %d available %d consume %d\n",
114 eid, credits, ep->tx_credits, consume);
115 ret = -EAGAIN;
116 goto unlock;
117 }
118
119 if (consume) {
120 ep->tx_credits -= credits;
121 ath10k_dbg(ar, ATH10K_DBG_HTC,
122 "htc ep %d consumed %d credits total %d\n",
123 eid, credits, ep->tx_credits);
124 }
125
126 unlock:
127 spin_unlock_bh(&htc->tx_lock);
128 return ret;
129 }
130
ath10k_htc_release_credit(struct ath10k_htc_ep * ep,unsigned int len)131 static void ath10k_htc_release_credit(struct ath10k_htc_ep *ep, unsigned int len)
132 {
133 struct ath10k_htc *htc = ep->htc;
134 struct ath10k *ar = htc->ar;
135 enum ath10k_htc_ep_id eid = ep->eid;
136 int credits;
137
138 if (!ep->tx_credit_flow_enabled)
139 return;
140
141 credits = DIV_ROUND_UP(len, ep->tx_credit_size);
142 spin_lock_bh(&htc->tx_lock);
143 ep->tx_credits += credits;
144 ath10k_dbg(ar, ATH10K_DBG_HTC,
145 "htc ep %d reverted %d credits back total %d\n",
146 eid, credits, ep->tx_credits);
147 spin_unlock_bh(&htc->tx_lock);
148
149 if (ep->ep_ops.ep_tx_credits)
150 ep->ep_ops.ep_tx_credits(htc->ar);
151 }
152
ath10k_htc_send(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,struct sk_buff * skb)153 int ath10k_htc_send(struct ath10k_htc *htc,
154 enum ath10k_htc_ep_id eid,
155 struct sk_buff *skb)
156 {
157 struct ath10k *ar = htc->ar;
158 struct ath10k_htc_ep *ep = &htc->endpoint[eid];
159 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
160 struct ath10k_hif_sg_item sg_item;
161 struct device *dev = htc->ar->dev;
162 int ret;
163 unsigned int skb_len;
164
165 if (htc->ar->state == ATH10K_STATE_WEDGED)
166 return -ECOMM;
167
168 if (eid >= ATH10K_HTC_EP_COUNT) {
169 ath10k_warn(ar, "Invalid endpoint id: %d\n", eid);
170 return -ENOENT;
171 }
172
173 skb_push(skb, sizeof(struct ath10k_htc_hdr));
174
175 skb_len = skb->len;
176 ret = ath10k_htc_consume_credit(ep, skb_len, true);
177 if (ret)
178 goto err_pull;
179
180 ath10k_htc_prepare_tx_skb(ep, skb);
181
182 skb_cb->eid = eid;
183 if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL) {
184 skb_cb->paddr = dma_map_single(dev, skb->data, skb->len,
185 DMA_TO_DEVICE);
186 ret = dma_mapping_error(dev, skb_cb->paddr);
187 if (ret) {
188 ret = -EIO;
189 goto err_credits;
190 }
191 }
192
193 sg_item.transfer_id = ep->eid;
194 sg_item.transfer_context = skb;
195 sg_item.vaddr = skb->data;
196 sg_item.paddr = skb_cb->paddr;
197 sg_item.len = skb->len;
198
199 ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
200 if (ret)
201 goto err_unmap;
202
203 return 0;
204
205 err_unmap:
206 if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
207 dma_unmap_single(dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
208 err_credits:
209 ath10k_htc_release_credit(ep, skb_len);
210 err_pull:
211 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
212 return ret;
213 }
214
ath10k_htc_tx_completion_handler(struct ath10k * ar,struct sk_buff * skb)215 void ath10k_htc_tx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
216 {
217 struct ath10k_htc *htc = &ar->htc;
218 struct ath10k_skb_cb *skb_cb;
219 struct ath10k_htc_ep *ep;
220
221 if (WARN_ON_ONCE(!skb))
222 return;
223
224 skb_cb = ATH10K_SKB_CB(skb);
225 ep = &htc->endpoint[skb_cb->eid];
226
227 ath10k_htc_notify_tx_completion(ep, skb);
228 /* the skb now belongs to the completion handler */
229 }
230 EXPORT_SYMBOL(ath10k_htc_tx_completion_handler);
231
232 /***********/
233 /* Receive */
234 /***********/
235
236 static void
ath10k_htc_process_credit_report(struct ath10k_htc * htc,const struct ath10k_htc_credit_report * report,int len,enum ath10k_htc_ep_id eid)237 ath10k_htc_process_credit_report(struct ath10k_htc *htc,
238 const struct ath10k_htc_credit_report *report,
239 int len,
240 enum ath10k_htc_ep_id eid)
241 {
242 struct ath10k *ar = htc->ar;
243 struct ath10k_htc_ep *ep;
244 int i, n_reports;
245
246 if (len % sizeof(*report))
247 ath10k_warn(ar, "Uneven credit report len %d", len);
248
249 n_reports = len / sizeof(*report);
250
251 spin_lock_bh(&htc->tx_lock);
252 for (i = 0; i < n_reports; i++, report++) {
253 if (report->eid >= ATH10K_HTC_EP_COUNT)
254 break;
255
256 ep = &htc->endpoint[report->eid];
257 ep->tx_credits += report->credits;
258
259 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc ep %d got %d credits (total %d)\n",
260 report->eid, report->credits, ep->tx_credits);
261
262 if (ep->ep_ops.ep_tx_credits) {
263 spin_unlock_bh(&htc->tx_lock);
264 ep->ep_ops.ep_tx_credits(htc->ar);
265 spin_lock_bh(&htc->tx_lock);
266 }
267 }
268 spin_unlock_bh(&htc->tx_lock);
269 }
270
271 static int
ath10k_htc_process_lookahead(struct ath10k_htc * htc,const struct ath10k_htc_lookahead_report * report,int len,enum ath10k_htc_ep_id eid,void * next_lookaheads,int * next_lookaheads_len)272 ath10k_htc_process_lookahead(struct ath10k_htc *htc,
273 const struct ath10k_htc_lookahead_report *report,
274 int len,
275 enum ath10k_htc_ep_id eid,
276 void *next_lookaheads,
277 int *next_lookaheads_len)
278 {
279 struct ath10k *ar = htc->ar;
280
281 /* Invalid lookahead flags are actually transmitted by
282 * the target in the HTC control message.
283 * Since this will happen at every boot we silently ignore
284 * the lookahead in this case
285 */
286 if (report->pre_valid != ((~report->post_valid) & 0xFF))
287 return 0;
288
289 if (next_lookaheads && next_lookaheads_len) {
290 ath10k_dbg(ar, ATH10K_DBG_HTC,
291 "htc rx lookahead found pre_valid 0x%x post_valid 0x%x\n",
292 report->pre_valid, report->post_valid);
293
294 /* look ahead bytes are valid, copy them over */
295 memcpy((u8 *)next_lookaheads, report->lookahead, 4);
296
297 *next_lookaheads_len = 1;
298 }
299
300 return 0;
301 }
302
303 static int
ath10k_htc_process_lookahead_bundle(struct ath10k_htc * htc,const struct ath10k_htc_lookahead_bundle * report,int len,enum ath10k_htc_ep_id eid,void * next_lookaheads,int * next_lookaheads_len)304 ath10k_htc_process_lookahead_bundle(struct ath10k_htc *htc,
305 const struct ath10k_htc_lookahead_bundle *report,
306 int len,
307 enum ath10k_htc_ep_id eid,
308 void *next_lookaheads,
309 int *next_lookaheads_len)
310 {
311 struct ath10k *ar = htc->ar;
312 int bundle_cnt = len / sizeof(*report);
313
314 if (!bundle_cnt || (bundle_cnt > htc->max_msgs_per_htc_bundle)) {
315 ath10k_warn(ar, "Invalid lookahead bundle count: %d\n",
316 bundle_cnt);
317 return -EINVAL;
318 }
319
320 if (next_lookaheads && next_lookaheads_len) {
321 int i;
322
323 for (i = 0; i < bundle_cnt; i++) {
324 memcpy(((u8 *)next_lookaheads) + 4 * i,
325 report->lookahead, 4);
326 report++;
327 }
328
329 *next_lookaheads_len = bundle_cnt;
330 }
331
332 return 0;
333 }
334
ath10k_htc_process_trailer(struct ath10k_htc * htc,u8 * buffer,int length,enum ath10k_htc_ep_id src_eid,void * next_lookaheads,int * next_lookaheads_len)335 int ath10k_htc_process_trailer(struct ath10k_htc *htc,
336 u8 *buffer,
337 int length,
338 enum ath10k_htc_ep_id src_eid,
339 void *next_lookaheads,
340 int *next_lookaheads_len)
341 {
342 struct ath10k_htc_lookahead_bundle *bundle;
343 struct ath10k *ar = htc->ar;
344 int status = 0;
345 struct ath10k_htc_record *record;
346 u8 *orig_buffer;
347 int orig_length;
348 size_t len;
349
350 orig_buffer = buffer;
351 orig_length = length;
352
353 while (length > 0) {
354 record = (struct ath10k_htc_record *)buffer;
355
356 if (length < sizeof(record->hdr)) {
357 status = -EINVAL;
358 break;
359 }
360
361 if (record->hdr.len > length) {
362 /* no room left in buffer for record */
363 ath10k_warn(ar, "Invalid record length: %d\n",
364 record->hdr.len);
365 status = -EINVAL;
366 break;
367 }
368
369 switch (record->hdr.id) {
370 case ATH10K_HTC_RECORD_CREDITS:
371 len = sizeof(struct ath10k_htc_credit_report);
372 if (record->hdr.len < len) {
373 ath10k_warn(ar, "Credit report too long\n");
374 status = -EINVAL;
375 break;
376 }
377 ath10k_htc_process_credit_report(htc,
378 record->credit_report,
379 record->hdr.len,
380 src_eid);
381 break;
382 case ATH10K_HTC_RECORD_LOOKAHEAD:
383 len = sizeof(struct ath10k_htc_lookahead_report);
384 if (record->hdr.len < len) {
385 ath10k_warn(ar, "Lookahead report too long\n");
386 status = -EINVAL;
387 break;
388 }
389 status = ath10k_htc_process_lookahead(htc,
390 record->lookahead_report,
391 record->hdr.len,
392 src_eid,
393 next_lookaheads,
394 next_lookaheads_len);
395 break;
396 case ATH10K_HTC_RECORD_LOOKAHEAD_BUNDLE:
397 bundle = record->lookahead_bundle;
398 status = ath10k_htc_process_lookahead_bundle(htc,
399 bundle,
400 record->hdr.len,
401 src_eid,
402 next_lookaheads,
403 next_lookaheads_len);
404 break;
405 default:
406 ath10k_warn(ar, "Unhandled record: id:%d length:%d\n",
407 record->hdr.id, record->hdr.len);
408 break;
409 }
410
411 if (status)
412 break;
413
414 /* multiple records may be present in a trailer */
415 buffer += sizeof(record->hdr) + record->hdr.len;
416 length -= sizeof(record->hdr) + record->hdr.len;
417 }
418
419 if (status)
420 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc rx bad trailer", "",
421 orig_buffer, orig_length);
422
423 return status;
424 }
425 EXPORT_SYMBOL(ath10k_htc_process_trailer);
426
ath10k_htc_rx_completion_handler(struct ath10k * ar,struct sk_buff * skb)427 void ath10k_htc_rx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
428 {
429 int status = 0;
430 struct ath10k_htc *htc = &ar->htc;
431 struct ath10k_htc_hdr *hdr;
432 struct ath10k_htc_ep *ep;
433 u16 payload_len;
434 u32 trailer_len = 0;
435 size_t min_len;
436 u8 eid;
437 bool trailer_present;
438
439 hdr = (struct ath10k_htc_hdr *)skb->data;
440 skb_pull(skb, sizeof(*hdr));
441
442 eid = hdr->eid;
443
444 if (eid >= ATH10K_HTC_EP_COUNT) {
445 ath10k_warn(ar, "HTC Rx: invalid eid %d\n", eid);
446 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad header", "",
447 hdr, sizeof(*hdr));
448 goto out;
449 }
450
451 ep = &htc->endpoint[eid];
452
453 payload_len = __le16_to_cpu(hdr->len);
454
455 if (payload_len + sizeof(*hdr) > ATH10K_HTC_MAX_LEN) {
456 ath10k_warn(ar, "HTC rx frame too long, len: %zu\n",
457 payload_len + sizeof(*hdr));
458 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len", "",
459 hdr, sizeof(*hdr));
460 goto out;
461 }
462
463 if (skb->len < payload_len) {
464 ath10k_dbg(ar, ATH10K_DBG_HTC,
465 "HTC Rx: insufficient length, got %d, expected %d\n",
466 skb->len, payload_len);
467 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len",
468 "", hdr, sizeof(*hdr));
469 goto out;
470 }
471
472 /* get flags to check for trailer */
473 trailer_present = hdr->flags & ATH10K_HTC_FLAG_TRAILER_PRESENT;
474 if (trailer_present) {
475 u8 *trailer;
476
477 trailer_len = hdr->trailer_len;
478 min_len = sizeof(struct ath10k_ath10k_htc_record_hdr);
479
480 if ((trailer_len < min_len) ||
481 (trailer_len > payload_len)) {
482 ath10k_warn(ar, "Invalid trailer length: %d\n",
483 trailer_len);
484 goto out;
485 }
486
487 trailer = (u8 *)hdr;
488 trailer += sizeof(*hdr);
489 trailer += payload_len;
490 trailer -= trailer_len;
491 status = ath10k_htc_process_trailer(htc, trailer,
492 trailer_len, hdr->eid,
493 NULL, NULL);
494 if (status)
495 goto out;
496
497 skb_trim(skb, skb->len - trailer_len);
498 }
499
500 if (((int)payload_len - (int)trailer_len) <= 0)
501 /* zero length packet with trailer data, just drop these */
502 goto out;
503
504 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc rx completion ep %d skb %pK\n",
505 eid, skb);
506 ep->ep_ops.ep_rx_complete(ar, skb);
507
508 /* skb is now owned by the rx completion handler */
509 skb = NULL;
510 out:
511 kfree_skb(skb);
512 }
513 EXPORT_SYMBOL(ath10k_htc_rx_completion_handler);
514
ath10k_htc_control_rx_complete(struct ath10k * ar,struct sk_buff * skb)515 static void ath10k_htc_control_rx_complete(struct ath10k *ar,
516 struct sk_buff *skb)
517 {
518 struct ath10k_htc *htc = &ar->htc;
519 struct ath10k_htc_msg *msg = (struct ath10k_htc_msg *)skb->data;
520
521 switch (__le16_to_cpu(msg->hdr.message_id)) {
522 case ATH10K_HTC_MSG_READY_ID:
523 case ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID:
524 /* handle HTC control message */
525 if (completion_done(&htc->ctl_resp)) {
526 /* this is a fatal error, target should not be
527 * sending unsolicited messages on the ep 0
528 */
529 ath10k_warn(ar, "HTC rx ctrl still processing\n");
530 complete(&htc->ctl_resp);
531 goto out;
532 }
533
534 htc->control_resp_len =
535 min_t(int, skb->len,
536 ATH10K_HTC_MAX_CTRL_MSG_LEN);
537
538 memcpy(htc->control_resp_buffer, skb->data,
539 htc->control_resp_len);
540
541 complete(&htc->ctl_resp);
542 break;
543 case ATH10K_HTC_MSG_SEND_SUSPEND_COMPLETE:
544 htc->htc_ops.target_send_suspend_complete(ar);
545 break;
546 default:
547 ath10k_warn(ar, "ignoring unsolicited htc ep0 event\n");
548 break;
549 }
550
551 out:
552 kfree_skb(skb);
553 }
554
555 /***************/
556 /* Init/Deinit */
557 /***************/
558
htc_service_name(enum ath10k_htc_svc_id id)559 static const char *htc_service_name(enum ath10k_htc_svc_id id)
560 {
561 switch (id) {
562 case ATH10K_HTC_SVC_ID_RESERVED:
563 return "Reserved";
564 case ATH10K_HTC_SVC_ID_RSVD_CTRL:
565 return "Control";
566 case ATH10K_HTC_SVC_ID_WMI_CONTROL:
567 return "WMI";
568 case ATH10K_HTC_SVC_ID_WMI_DATA_BE:
569 return "DATA BE";
570 case ATH10K_HTC_SVC_ID_WMI_DATA_BK:
571 return "DATA BK";
572 case ATH10K_HTC_SVC_ID_WMI_DATA_VI:
573 return "DATA VI";
574 case ATH10K_HTC_SVC_ID_WMI_DATA_VO:
575 return "DATA VO";
576 case ATH10K_HTC_SVC_ID_NMI_CONTROL:
577 return "NMI Control";
578 case ATH10K_HTC_SVC_ID_NMI_DATA:
579 return "NMI Data";
580 case ATH10K_HTC_SVC_ID_HTT_DATA_MSG:
581 return "HTT Data";
582 case ATH10K_HTC_SVC_ID_HTT_DATA2_MSG:
583 return "HTT Data";
584 case ATH10K_HTC_SVC_ID_HTT_DATA3_MSG:
585 return "HTT Data";
586 case ATH10K_HTC_SVC_ID_TEST_RAW_STREAMS:
587 return "RAW";
588 case ATH10K_HTC_SVC_ID_HTT_LOG_MSG:
589 return "PKTLOG";
590 }
591
592 return "Unknown";
593 }
594
ath10k_htc_reset_endpoint_states(struct ath10k_htc * htc)595 static void ath10k_htc_reset_endpoint_states(struct ath10k_htc *htc)
596 {
597 struct ath10k_htc_ep *ep;
598 int i;
599
600 for (i = ATH10K_HTC_EP_0; i < ATH10K_HTC_EP_COUNT; i++) {
601 ep = &htc->endpoint[i];
602 ep->service_id = ATH10K_HTC_SVC_ID_UNUSED;
603 ep->max_ep_message_len = 0;
604 ep->max_tx_queue_depth = 0;
605 ep->eid = i;
606 ep->htc = htc;
607 ep->tx_credit_flow_enabled = true;
608 }
609 }
610
ath10k_htc_get_credit_allocation(struct ath10k_htc * htc,u16 service_id)611 static u8 ath10k_htc_get_credit_allocation(struct ath10k_htc *htc,
612 u16 service_id)
613 {
614 u8 allocation = 0;
615
616 /* The WMI control service is the only service with flow control.
617 * Let it have all transmit credits.
618 */
619 if (service_id == ATH10K_HTC_SVC_ID_WMI_CONTROL)
620 allocation = htc->total_transmit_credits;
621
622 return allocation;
623 }
624
ath10k_htc_send_bundle(struct ath10k_htc_ep * ep,struct sk_buff * bundle_skb,struct sk_buff_head * tx_save_head)625 static int ath10k_htc_send_bundle(struct ath10k_htc_ep *ep,
626 struct sk_buff *bundle_skb,
627 struct sk_buff_head *tx_save_head)
628 {
629 struct ath10k_hif_sg_item sg_item;
630 struct ath10k_htc *htc = ep->htc;
631 struct ath10k *ar = htc->ar;
632 struct sk_buff *skb;
633 int ret, cn = 0;
634 unsigned int skb_len;
635
636 ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle skb len %d\n", bundle_skb->len);
637 skb_len = bundle_skb->len;
638 ret = ath10k_htc_consume_credit(ep, skb_len, true);
639
640 if (!ret) {
641 sg_item.transfer_id = ep->eid;
642 sg_item.transfer_context = bundle_skb;
643 sg_item.vaddr = bundle_skb->data;
644 sg_item.len = bundle_skb->len;
645
646 ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
647 if (ret)
648 ath10k_htc_release_credit(ep, skb_len);
649 }
650
651 if (ret)
652 dev_kfree_skb_any(bundle_skb);
653
654 for (cn = 0; (skb = skb_dequeue_tail(tx_save_head)); cn++) {
655 if (ret) {
656 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
657 skb_queue_head(&ep->tx_req_head, skb);
658 } else {
659 skb_queue_tail(&ep->tx_complete_head, skb);
660 }
661 }
662
663 if (!ret)
664 queue_work(ar->workqueue_tx_complete, &ar->tx_complete_work);
665
666 ath10k_dbg(ar, ATH10K_DBG_HTC,
667 "bundle tx status %d eid %d req count %d count %d len %d\n",
668 ret, ep->eid, skb_queue_len(&ep->tx_req_head), cn, bundle_skb->len);
669 return ret;
670 }
671
ath10k_htc_send_one_skb(struct ath10k_htc_ep * ep,struct sk_buff * skb)672 static void ath10k_htc_send_one_skb(struct ath10k_htc_ep *ep, struct sk_buff *skb)
673 {
674 struct ath10k_htc *htc = ep->htc;
675 struct ath10k *ar = htc->ar;
676 int ret;
677
678 ret = ath10k_htc_send(htc, ep->eid, skb);
679
680 if (ret)
681 skb_queue_head(&ep->tx_req_head, skb);
682
683 ath10k_dbg(ar, ATH10K_DBG_HTC, "tx one status %d eid %d len %d pending count %d\n",
684 ret, ep->eid, skb->len, skb_queue_len(&ep->tx_req_head));
685 }
686
ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep * ep)687 static int ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep *ep)
688 {
689 struct ath10k_htc *htc = ep->htc;
690 struct sk_buff *bundle_skb, *skb;
691 struct sk_buff_head tx_save_head;
692 struct ath10k_htc_hdr *hdr;
693 u8 *bundle_buf;
694 int ret = 0, credit_pad, credit_remainder, trans_len, bundles_left = 0;
695
696 if (htc->ar->state == ATH10K_STATE_WEDGED)
697 return -ECOMM;
698
699 if (ep->tx_credit_flow_enabled &&
700 ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE)
701 return 0;
702
703 bundles_left = ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
704 bundle_skb = dev_alloc_skb(bundles_left);
705
706 if (!bundle_skb)
707 return -ENOMEM;
708
709 bundle_buf = bundle_skb->data;
710 skb_queue_head_init(&tx_save_head);
711
712 while (true) {
713 skb = skb_dequeue(&ep->tx_req_head);
714 if (!skb)
715 break;
716
717 credit_pad = 0;
718 trans_len = skb->len + sizeof(*hdr);
719 credit_remainder = trans_len % ep->tx_credit_size;
720
721 if (credit_remainder != 0) {
722 credit_pad = ep->tx_credit_size - credit_remainder;
723 trans_len += credit_pad;
724 }
725
726 ret = ath10k_htc_consume_credit(ep,
727 bundle_buf + trans_len - bundle_skb->data,
728 false);
729 if (ret) {
730 skb_queue_head(&ep->tx_req_head, skb);
731 break;
732 }
733
734 if (bundles_left < trans_len) {
735 bundle_skb->len = bundle_buf - bundle_skb->data;
736 ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
737
738 if (ret) {
739 skb_queue_head(&ep->tx_req_head, skb);
740 return ret;
741 }
742
743 if (skb_queue_len(&ep->tx_req_head) == 0) {
744 ath10k_htc_send_one_skb(ep, skb);
745 return ret;
746 }
747
748 if (ep->tx_credit_flow_enabled &&
749 ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE) {
750 skb_queue_head(&ep->tx_req_head, skb);
751 return 0;
752 }
753
754 bundles_left =
755 ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
756 bundle_skb = dev_alloc_skb(bundles_left);
757
758 if (!bundle_skb) {
759 skb_queue_head(&ep->tx_req_head, skb);
760 return -ENOMEM;
761 }
762 bundle_buf = bundle_skb->data;
763 skb_queue_head_init(&tx_save_head);
764 }
765
766 skb_push(skb, sizeof(struct ath10k_htc_hdr));
767 ath10k_htc_prepare_tx_skb(ep, skb);
768
769 memcpy(bundle_buf, skb->data, skb->len);
770 hdr = (struct ath10k_htc_hdr *)bundle_buf;
771 hdr->flags |= ATH10K_HTC_FLAG_SEND_BUNDLE;
772 hdr->pad_len = __cpu_to_le16(credit_pad);
773 bundle_buf += trans_len;
774 bundles_left -= trans_len;
775 skb_queue_tail(&tx_save_head, skb);
776 }
777
778 if (bundle_buf != bundle_skb->data) {
779 bundle_skb->len = bundle_buf - bundle_skb->data;
780 ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
781 } else {
782 dev_kfree_skb_any(bundle_skb);
783 }
784
785 return ret;
786 }
787
ath10k_htc_bundle_tx_work(struct work_struct * work)788 static void ath10k_htc_bundle_tx_work(struct work_struct *work)
789 {
790 struct ath10k *ar = container_of(work, struct ath10k, bundle_tx_work);
791 struct ath10k_htc_ep *ep;
792 struct sk_buff *skb;
793 int i;
794
795 for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
796 ep = &ar->htc.endpoint[i];
797
798 if (!ep->bundle_tx)
799 continue;
800
801 ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx work eid %d count %d\n",
802 ep->eid, skb_queue_len(&ep->tx_req_head));
803
804 if (skb_queue_len(&ep->tx_req_head) >=
805 ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE) {
806 ath10k_htc_send_bundle_skbs(ep);
807 } else {
808 skb = skb_dequeue(&ep->tx_req_head);
809
810 if (!skb)
811 continue;
812 ath10k_htc_send_one_skb(ep, skb);
813 }
814 }
815 }
816
ath10k_htc_tx_complete_work(struct work_struct * work)817 static void ath10k_htc_tx_complete_work(struct work_struct *work)
818 {
819 struct ath10k *ar = container_of(work, struct ath10k, tx_complete_work);
820 struct ath10k_htc_ep *ep;
821 enum ath10k_htc_ep_id eid;
822 struct sk_buff *skb;
823 int i;
824
825 for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
826 ep = &ar->htc.endpoint[i];
827 eid = ep->eid;
828 if (ep->bundle_tx && eid == ar->htt.eid) {
829 ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx complete eid %d pending complete count%d\n",
830 ep->eid, skb_queue_len(&ep->tx_complete_head));
831
832 while (true) {
833 skb = skb_dequeue(&ep->tx_complete_head);
834 if (!skb)
835 break;
836 ath10k_htc_notify_tx_completion(ep, skb);
837 }
838 }
839 }
840 }
841
ath10k_htc_send_hl(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,struct sk_buff * skb)842 int ath10k_htc_send_hl(struct ath10k_htc *htc,
843 enum ath10k_htc_ep_id eid,
844 struct sk_buff *skb)
845 {
846 struct ath10k_htc_ep *ep = &htc->endpoint[eid];
847 struct ath10k *ar = htc->ar;
848
849 if (sizeof(struct ath10k_htc_hdr) + skb->len > ep->tx_credit_size) {
850 ath10k_dbg(ar, ATH10K_DBG_HTC, "tx exceed max len %d\n", skb->len);
851 return -ENOMEM;
852 }
853
854 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc send hl eid %d bundle %d tx count %d len %d\n",
855 eid, ep->bundle_tx, skb_queue_len(&ep->tx_req_head), skb->len);
856
857 if (ep->bundle_tx) {
858 skb_queue_tail(&ep->tx_req_head, skb);
859 queue_work(ar->workqueue, &ar->bundle_tx_work);
860 return 0;
861 } else {
862 return ath10k_htc_send(htc, eid, skb);
863 }
864 }
865
ath10k_htc_setup_tx_req(struct ath10k_htc_ep * ep)866 void ath10k_htc_setup_tx_req(struct ath10k_htc_ep *ep)
867 {
868 if (ep->htc->max_msgs_per_htc_bundle >= ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE &&
869 !ep->bundle_tx) {
870 ep->bundle_tx = true;
871 skb_queue_head_init(&ep->tx_req_head);
872 skb_queue_head_init(&ep->tx_complete_head);
873 }
874 }
875
ath10k_htc_stop_hl(struct ath10k * ar)876 void ath10k_htc_stop_hl(struct ath10k *ar)
877 {
878 struct ath10k_htc_ep *ep;
879 int i;
880
881 cancel_work_sync(&ar->bundle_tx_work);
882 cancel_work_sync(&ar->tx_complete_work);
883
884 for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
885 ep = &ar->htc.endpoint[i];
886
887 if (!ep->bundle_tx)
888 continue;
889
890 ath10k_dbg(ar, ATH10K_DBG_HTC, "stop tx work eid %d count %d\n",
891 ep->eid, skb_queue_len(&ep->tx_req_head));
892
893 skb_queue_purge(&ep->tx_req_head);
894 }
895 }
896
ath10k_htc_wait_target(struct ath10k_htc * htc)897 int ath10k_htc_wait_target(struct ath10k_htc *htc)
898 {
899 struct ath10k *ar = htc->ar;
900 int i, status = 0;
901 unsigned long time_left;
902 struct ath10k_htc_msg *msg;
903 u16 message_id;
904
905 time_left = wait_for_completion_timeout(&htc->ctl_resp,
906 ATH10K_HTC_WAIT_TIMEOUT_HZ);
907 if (!time_left) {
908 /* Workaround: In some cases the PCI HIF doesn't
909 * receive interrupt for the control response message
910 * even if the buffer was completed. It is suspected
911 * iomap writes unmasking PCI CE irqs aren't propagated
912 * properly in KVM PCI-passthrough sometimes.
913 */
914 ath10k_warn(ar, "failed to receive control response completion, polling..\n");
915
916 for (i = 0; i < CE_COUNT; i++)
917 ath10k_hif_send_complete_check(htc->ar, i, 1);
918
919 time_left =
920 wait_for_completion_timeout(&htc->ctl_resp,
921 ATH10K_HTC_WAIT_TIMEOUT_HZ);
922
923 if (!time_left)
924 status = -ETIMEDOUT;
925 }
926
927 if (status < 0) {
928 ath10k_err(ar, "ctl_resp never came in (%d)\n", status);
929 return status;
930 }
931
932 if (htc->control_resp_len < sizeof(msg->hdr) + sizeof(msg->ready)) {
933 ath10k_err(ar, "Invalid HTC ready msg len:%d\n",
934 htc->control_resp_len);
935 return -ECOMM;
936 }
937
938 msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
939 message_id = __le16_to_cpu(msg->hdr.message_id);
940
941 if (message_id != ATH10K_HTC_MSG_READY_ID) {
942 ath10k_err(ar, "Invalid HTC ready msg: 0x%x\n", message_id);
943 return -ECOMM;
944 }
945
946 htc->total_transmit_credits = __le16_to_cpu(msg->ready.credit_count);
947 htc->target_credit_size = __le16_to_cpu(msg->ready.credit_size);
948
949 ath10k_dbg(ar, ATH10K_DBG_HTC,
950 "Target ready! transmit resources: %d size:%d\n",
951 htc->total_transmit_credits,
952 htc->target_credit_size);
953
954 if ((htc->total_transmit_credits == 0) ||
955 (htc->target_credit_size == 0)) {
956 ath10k_err(ar, "Invalid credit size received\n");
957 return -ECOMM;
958 }
959
960 /* The only way to determine if the ready message is an extended
961 * message is from the size.
962 */
963 if (htc->control_resp_len >=
964 sizeof(msg->hdr) + sizeof(msg->ready_ext)) {
965 htc->alt_data_credit_size =
966 __le16_to_cpu(msg->ready_ext.reserved) &
967 ATH10K_HTC_MSG_READY_EXT_ALT_DATA_MASK;
968 htc->max_msgs_per_htc_bundle =
969 min_t(u8, msg->ready_ext.max_msgs_per_htc_bundle,
970 HTC_HOST_MAX_MSG_PER_RX_BUNDLE);
971 ath10k_dbg(ar, ATH10K_DBG_HTC,
972 "Extended ready message RX bundle size %d alt size %d\n",
973 htc->max_msgs_per_htc_bundle,
974 htc->alt_data_credit_size);
975 }
976
977 INIT_WORK(&ar->bundle_tx_work, ath10k_htc_bundle_tx_work);
978 INIT_WORK(&ar->tx_complete_work, ath10k_htc_tx_complete_work);
979
980 return 0;
981 }
982
ath10k_htc_change_tx_credit_flow(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,bool enable)983 void ath10k_htc_change_tx_credit_flow(struct ath10k_htc *htc,
984 enum ath10k_htc_ep_id eid,
985 bool enable)
986 {
987 struct ath10k *ar = htc->ar;
988 struct ath10k_htc_ep *ep = &ar->htc.endpoint[eid];
989
990 ep->tx_credit_flow_enabled = enable;
991 }
992
ath10k_htc_connect_service(struct ath10k_htc * htc,struct ath10k_htc_svc_conn_req * conn_req,struct ath10k_htc_svc_conn_resp * conn_resp)993 int ath10k_htc_connect_service(struct ath10k_htc *htc,
994 struct ath10k_htc_svc_conn_req *conn_req,
995 struct ath10k_htc_svc_conn_resp *conn_resp)
996 {
997 struct ath10k *ar = htc->ar;
998 struct ath10k_htc_msg *msg;
999 struct ath10k_htc_conn_svc *req_msg;
1000 struct ath10k_htc_conn_svc_response resp_msg_dummy;
1001 struct ath10k_htc_conn_svc_response *resp_msg = &resp_msg_dummy;
1002 enum ath10k_htc_ep_id assigned_eid = ATH10K_HTC_EP_COUNT;
1003 struct ath10k_htc_ep *ep;
1004 struct sk_buff *skb;
1005 unsigned int max_msg_size = 0;
1006 int length, status;
1007 unsigned long time_left;
1008 bool disable_credit_flow_ctrl = false;
1009 u16 message_id, service_id, flags = 0;
1010 u8 tx_alloc = 0;
1011
1012 /* special case for HTC pseudo control service */
1013 if (conn_req->service_id == ATH10K_HTC_SVC_ID_RSVD_CTRL) {
1014 disable_credit_flow_ctrl = true;
1015 assigned_eid = ATH10K_HTC_EP_0;
1016 max_msg_size = ATH10K_HTC_MAX_CTRL_MSG_LEN;
1017 memset(&resp_msg_dummy, 0, sizeof(resp_msg_dummy));
1018 goto setup;
1019 }
1020
1021 tx_alloc = ath10k_htc_get_credit_allocation(htc,
1022 conn_req->service_id);
1023 if (!tx_alloc)
1024 ath10k_dbg(ar, ATH10K_DBG_BOOT,
1025 "boot htc service %s does not allocate target credits\n",
1026 htc_service_name(conn_req->service_id));
1027
1028 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1029 if (!skb) {
1030 ath10k_err(ar, "Failed to allocate HTC packet\n");
1031 return -ENOMEM;
1032 }
1033
1034 length = sizeof(msg->hdr) + sizeof(msg->connect_service);
1035 skb_put(skb, length);
1036 memset(skb->data, 0, length);
1037
1038 msg = (struct ath10k_htc_msg *)skb->data;
1039 msg->hdr.message_id =
1040 __cpu_to_le16(ATH10K_HTC_MSG_CONNECT_SERVICE_ID);
1041
1042 flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC);
1043
1044 /* Only enable credit flow control for WMI ctrl service */
1045 if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) {
1046 flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL;
1047 disable_credit_flow_ctrl = true;
1048 }
1049
1050 req_msg = &msg->connect_service;
1051 req_msg->flags = __cpu_to_le16(flags);
1052 req_msg->service_id = __cpu_to_le16(conn_req->service_id);
1053
1054 reinit_completion(&htc->ctl_resp);
1055
1056 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1057 if (status) {
1058 kfree_skb(skb);
1059 return status;
1060 }
1061
1062 /* wait for response */
1063 time_left = wait_for_completion_timeout(&htc->ctl_resp,
1064 ATH10K_HTC_CONN_SVC_TIMEOUT_HZ);
1065 if (!time_left) {
1066 ath10k_err(ar, "Service connect timeout\n");
1067 return -ETIMEDOUT;
1068 }
1069
1070 /* we controlled the buffer creation, it's aligned */
1071 msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
1072 resp_msg = &msg->connect_service_response;
1073 message_id = __le16_to_cpu(msg->hdr.message_id);
1074 service_id = __le16_to_cpu(resp_msg->service_id);
1075
1076 if ((message_id != ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID) ||
1077 (htc->control_resp_len < sizeof(msg->hdr) +
1078 sizeof(msg->connect_service_response))) {
1079 ath10k_err(ar, "Invalid resp message ID 0x%x", message_id);
1080 return -EPROTO;
1081 }
1082
1083 ath10k_dbg(ar, ATH10K_DBG_HTC,
1084 "HTC Service %s connect response: status: 0x%x, assigned ep: 0x%x\n",
1085 htc_service_name(service_id),
1086 resp_msg->status, resp_msg->eid);
1087
1088 conn_resp->connect_resp_code = resp_msg->status;
1089
1090 /* check response status */
1091 if (resp_msg->status != ATH10K_HTC_CONN_SVC_STATUS_SUCCESS) {
1092 ath10k_err(ar, "HTC Service %s connect request failed: 0x%x)\n",
1093 htc_service_name(service_id),
1094 resp_msg->status);
1095 return -EPROTO;
1096 }
1097
1098 assigned_eid = (enum ath10k_htc_ep_id)resp_msg->eid;
1099 max_msg_size = __le16_to_cpu(resp_msg->max_msg_size);
1100
1101 setup:
1102
1103 if (assigned_eid >= ATH10K_HTC_EP_COUNT)
1104 return -EPROTO;
1105
1106 if (max_msg_size == 0)
1107 return -EPROTO;
1108
1109 ep = &htc->endpoint[assigned_eid];
1110 ep->eid = assigned_eid;
1111
1112 if (ep->service_id != ATH10K_HTC_SVC_ID_UNUSED)
1113 return -EPROTO;
1114
1115 /* return assigned endpoint to caller */
1116 conn_resp->eid = assigned_eid;
1117 conn_resp->max_msg_len = __le16_to_cpu(resp_msg->max_msg_size);
1118
1119 /* setup the endpoint */
1120 ep->service_id = conn_req->service_id;
1121 ep->max_tx_queue_depth = conn_req->max_send_queue_depth;
1122 ep->max_ep_message_len = __le16_to_cpu(resp_msg->max_msg_size);
1123 ep->tx_credits = tx_alloc;
1124 ep->tx_credit_size = htc->target_credit_size;
1125
1126 if (conn_req->service_id == ATH10K_HTC_SVC_ID_HTT_DATA_MSG &&
1127 htc->alt_data_credit_size != 0)
1128 ep->tx_credit_size = htc->alt_data_credit_size;
1129
1130 /* copy all the callbacks */
1131 ep->ep_ops = conn_req->ep_ops;
1132
1133 status = ath10k_hif_map_service_to_pipe(htc->ar,
1134 ep->service_id,
1135 &ep->ul_pipe_id,
1136 &ep->dl_pipe_id);
1137 if (status) {
1138 ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC service id: %d\n",
1139 ep->service_id);
1140 return status;
1141 }
1142
1143 ath10k_dbg(ar, ATH10K_DBG_BOOT,
1144 "boot htc service '%s' ul pipe %d dl pipe %d eid %d ready\n",
1145 htc_service_name(ep->service_id), ep->ul_pipe_id,
1146 ep->dl_pipe_id, ep->eid);
1147
1148 if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) {
1149 ep->tx_credit_flow_enabled = false;
1150 ath10k_dbg(ar, ATH10K_DBG_BOOT,
1151 "boot htc service '%s' eid %d TX flow control disabled\n",
1152 htc_service_name(ep->service_id), assigned_eid);
1153 }
1154
1155 return status;
1156 }
1157
ath10k_htc_alloc_skb(struct ath10k * ar,int size)1158 struct sk_buff *ath10k_htc_alloc_skb(struct ath10k *ar, int size)
1159 {
1160 struct sk_buff *skb;
1161
1162 skb = dev_alloc_skb(size + sizeof(struct ath10k_htc_hdr));
1163 if (!skb)
1164 return NULL;
1165
1166 skb_reserve(skb, sizeof(struct ath10k_htc_hdr));
1167
1168 /* FW/HTC requires 4-byte aligned streams */
1169 if (!IS_ALIGNED((unsigned long)skb->data, 4))
1170 ath10k_warn(ar, "Unaligned HTC tx skb\n");
1171
1172 return skb;
1173 }
1174
ath10k_htc_pktlog_process_rx(struct ath10k * ar,struct sk_buff * skb)1175 static void ath10k_htc_pktlog_process_rx(struct ath10k *ar, struct sk_buff *skb)
1176 {
1177 trace_ath10k_htt_pktlog(ar, skb->data, skb->len);
1178 dev_kfree_skb_any(skb);
1179 }
1180
ath10k_htc_pktlog_connect(struct ath10k * ar)1181 static int ath10k_htc_pktlog_connect(struct ath10k *ar)
1182 {
1183 struct ath10k_htc_svc_conn_resp conn_resp;
1184 struct ath10k_htc_svc_conn_req conn_req;
1185 int status;
1186
1187 memset(&conn_req, 0, sizeof(conn_req));
1188 memset(&conn_resp, 0, sizeof(conn_resp));
1189
1190 conn_req.ep_ops.ep_tx_complete = NULL;
1191 conn_req.ep_ops.ep_rx_complete = ath10k_htc_pktlog_process_rx;
1192 conn_req.ep_ops.ep_tx_credits = NULL;
1193
1194 /* connect to control service */
1195 conn_req.service_id = ATH10K_HTC_SVC_ID_HTT_LOG_MSG;
1196 status = ath10k_htc_connect_service(&ar->htc, &conn_req, &conn_resp);
1197 if (status) {
1198 ath10k_warn(ar, "failed to connect to PKTLOG service: %d\n",
1199 status);
1200 return status;
1201 }
1202
1203 return 0;
1204 }
1205
ath10k_htc_pktlog_svc_supported(struct ath10k * ar)1206 static bool ath10k_htc_pktlog_svc_supported(struct ath10k *ar)
1207 {
1208 u8 ul_pipe_id;
1209 u8 dl_pipe_id;
1210 int status;
1211
1212 status = ath10k_hif_map_service_to_pipe(ar, ATH10K_HTC_SVC_ID_HTT_LOG_MSG,
1213 &ul_pipe_id,
1214 &dl_pipe_id);
1215 if (status) {
1216 ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC pktlog service id: %d\n",
1217 ATH10K_HTC_SVC_ID_HTT_LOG_MSG);
1218
1219 return false;
1220 }
1221
1222 return true;
1223 }
1224
ath10k_htc_start(struct ath10k_htc * htc)1225 int ath10k_htc_start(struct ath10k_htc *htc)
1226 {
1227 struct ath10k *ar = htc->ar;
1228 struct sk_buff *skb;
1229 int status = 0;
1230 struct ath10k_htc_msg *msg;
1231
1232 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1233 if (!skb)
1234 return -ENOMEM;
1235
1236 skb_put(skb, sizeof(msg->hdr) + sizeof(msg->setup_complete_ext));
1237 memset(skb->data, 0, skb->len);
1238
1239 msg = (struct ath10k_htc_msg *)skb->data;
1240 msg->hdr.message_id =
1241 __cpu_to_le16(ATH10K_HTC_MSG_SETUP_COMPLETE_EX_ID);
1242
1243 if (ar->hif.bus == ATH10K_BUS_SDIO) {
1244 /* Extra setup params used by SDIO */
1245 msg->setup_complete_ext.flags =
1246 __cpu_to_le32(ATH10K_HTC_SETUP_COMPLETE_FLAGS_RX_BNDL_EN);
1247 msg->setup_complete_ext.max_msgs_per_bundled_recv =
1248 htc->max_msgs_per_htc_bundle;
1249 }
1250 ath10k_dbg(ar, ATH10K_DBG_HTC, "HTC is using TX credit flow control\n");
1251
1252 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1253 if (status) {
1254 kfree_skb(skb);
1255 return status;
1256 }
1257
1258 if (ath10k_htc_pktlog_svc_supported(ar)) {
1259 status = ath10k_htc_pktlog_connect(ar);
1260 if (status) {
1261 ath10k_err(ar, "failed to connect to pktlog: %d\n", status);
1262 return status;
1263 }
1264 }
1265
1266 return 0;
1267 }
1268
1269 /* registered target arrival callback from the HIF layer */
ath10k_htc_init(struct ath10k * ar)1270 int ath10k_htc_init(struct ath10k *ar)
1271 {
1272 int status;
1273 struct ath10k_htc *htc = &ar->htc;
1274 struct ath10k_htc_svc_conn_req conn_req;
1275 struct ath10k_htc_svc_conn_resp conn_resp;
1276
1277 spin_lock_init(&htc->tx_lock);
1278
1279 ath10k_htc_reset_endpoint_states(htc);
1280
1281 htc->ar = ar;
1282
1283 /* setup our pseudo HTC control endpoint connection */
1284 memset(&conn_req, 0, sizeof(conn_req));
1285 memset(&conn_resp, 0, sizeof(conn_resp));
1286 conn_req.ep_ops.ep_tx_complete = ath10k_htc_control_tx_complete;
1287 conn_req.ep_ops.ep_rx_complete = ath10k_htc_control_rx_complete;
1288 conn_req.max_send_queue_depth = ATH10K_NUM_CONTROL_TX_BUFFERS;
1289 conn_req.service_id = ATH10K_HTC_SVC_ID_RSVD_CTRL;
1290
1291 /* connect fake service */
1292 status = ath10k_htc_connect_service(htc, &conn_req, &conn_resp);
1293 if (status) {
1294 ath10k_err(ar, "could not connect to htc service (%d)\n",
1295 status);
1296 return status;
1297 }
1298
1299 init_completion(&htc->ctl_resp);
1300
1301 return 0;
1302 }
1303