xref: /linux/drivers/net/wireless/ath/ath10k/htc.c (revision 06d07429858317ded2db7986113a9e0129cd599b)
1 // SPDX-License-Identifier: ISC
2 /*
3  * Copyright (c) 2005-2011 Atheros Communications Inc.
4  * Copyright (c) 2011-2017 Qualcomm Atheros, Inc.
5  * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
6  */
7 
8 #include "core.h"
9 #include "hif.h"
10 #include "debug.h"
11 
12 /********/
13 /* Send */
14 /********/
15 
ath10k_htc_control_tx_complete(struct ath10k * ar,struct sk_buff * skb)16 static void ath10k_htc_control_tx_complete(struct ath10k *ar,
17 					   struct sk_buff *skb)
18 {
19 	kfree_skb(skb);
20 }
21 
ath10k_htc_build_tx_ctrl_skb(void * ar)22 static struct sk_buff *ath10k_htc_build_tx_ctrl_skb(void *ar)
23 {
24 	struct sk_buff *skb;
25 	struct ath10k_skb_cb *skb_cb;
26 
27 	skb = dev_alloc_skb(ATH10K_HTC_CONTROL_BUFFER_SIZE);
28 	if (!skb)
29 		return NULL;
30 
31 	skb_reserve(skb, 20); /* FIXME: why 20 bytes? */
32 	WARN_ONCE((unsigned long)skb->data & 3, "unaligned skb");
33 
34 	skb_cb = ATH10K_SKB_CB(skb);
35 	memset(skb_cb, 0, sizeof(*skb_cb));
36 
37 	ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: skb %pK\n", __func__, skb);
38 	return skb;
39 }
40 
ath10k_htc_restore_tx_skb(struct ath10k_htc * htc,struct sk_buff * skb)41 static inline void ath10k_htc_restore_tx_skb(struct ath10k_htc *htc,
42 					     struct sk_buff *skb)
43 {
44 	struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
45 
46 	if (htc->ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
47 		dma_unmap_single(htc->ar->dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
48 	skb_pull(skb, sizeof(struct ath10k_htc_hdr));
49 }
50 
ath10k_htc_notify_tx_completion(struct ath10k_htc_ep * ep,struct sk_buff * skb)51 void ath10k_htc_notify_tx_completion(struct ath10k_htc_ep *ep,
52 				     struct sk_buff *skb)
53 {
54 	struct ath10k *ar = ep->htc->ar;
55 	struct ath10k_htc_hdr *hdr;
56 
57 	ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: ep %d skb %pK\n", __func__,
58 		   ep->eid, skb);
59 
60 	/* A corner case where the copy completion is reaching to host but still
61 	 * copy engine is processing it due to which host unmaps corresponding
62 	 * memory and causes SMMU fault, hence as workaround adding delay
63 	 * the unmapping memory to avoid SMMU faults.
64 	 */
65 	if (ar->hw_params.delay_unmap_buffer &&
66 	    ep->ul_pipe_id == 3)
67 		mdelay(2);
68 
69 	hdr = (struct ath10k_htc_hdr *)skb->data;
70 	ath10k_htc_restore_tx_skb(ep->htc, skb);
71 
72 	if (!ep->ep_ops.ep_tx_complete) {
73 		ath10k_warn(ar, "no tx handler for eid %d\n", ep->eid);
74 		dev_kfree_skb_any(skb);
75 		return;
76 	}
77 
78 	if (hdr->flags & ATH10K_HTC_FLAG_SEND_BUNDLE) {
79 		dev_kfree_skb_any(skb);
80 		return;
81 	}
82 
83 	ep->ep_ops.ep_tx_complete(ep->htc->ar, skb);
84 }
85 EXPORT_SYMBOL(ath10k_htc_notify_tx_completion);
86 
ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep * ep,struct sk_buff * skb)87 static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
88 				      struct sk_buff *skb)
89 {
90 	struct ath10k_htc_hdr *hdr;
91 
92 	hdr = (struct ath10k_htc_hdr *)skb->data;
93 	memset(hdr, 0, sizeof(struct ath10k_htc_hdr));
94 
95 	hdr->eid = ep->eid;
96 	hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr));
97 	hdr->flags = 0;
98 	if (ep->tx_credit_flow_enabled && !ep->bundle_tx)
99 		hdr->flags |= ATH10K_HTC_FLAG_NEED_CREDIT_UPDATE;
100 
101 	spin_lock_bh(&ep->htc->tx_lock);
102 	hdr->seq_no = ep->seq_no++;
103 	spin_unlock_bh(&ep->htc->tx_lock);
104 }
105 
ath10k_htc_consume_credit(struct ath10k_htc_ep * ep,unsigned int len,bool consume)106 static int ath10k_htc_consume_credit(struct ath10k_htc_ep *ep,
107 				     unsigned int len,
108 				     bool consume)
109 {
110 	struct ath10k_htc *htc = ep->htc;
111 	struct ath10k *ar = htc->ar;
112 	enum ath10k_htc_ep_id eid = ep->eid;
113 	int credits, ret = 0;
114 
115 	if (!ep->tx_credit_flow_enabled)
116 		return 0;
117 
118 	credits = DIV_ROUND_UP(len, ep->tx_credit_size);
119 	spin_lock_bh(&htc->tx_lock);
120 
121 	if (ep->tx_credits < credits) {
122 		ath10k_dbg(ar, ATH10K_DBG_HTC,
123 			   "htc insufficient credits ep %d required %d available %d consume %d\n",
124 			   eid, credits, ep->tx_credits, consume);
125 		ret = -EAGAIN;
126 		goto unlock;
127 	}
128 
129 	if (consume) {
130 		ep->tx_credits -= credits;
131 		ath10k_dbg(ar, ATH10K_DBG_HTC,
132 			   "htc ep %d consumed %d credits total %d\n",
133 			   eid, credits, ep->tx_credits);
134 	}
135 
136 unlock:
137 	spin_unlock_bh(&htc->tx_lock);
138 	return ret;
139 }
140 
ath10k_htc_release_credit(struct ath10k_htc_ep * ep,unsigned int len)141 static void ath10k_htc_release_credit(struct ath10k_htc_ep *ep, unsigned int len)
142 {
143 	struct ath10k_htc *htc = ep->htc;
144 	struct ath10k *ar = htc->ar;
145 	enum ath10k_htc_ep_id eid = ep->eid;
146 	int credits;
147 
148 	if (!ep->tx_credit_flow_enabled)
149 		return;
150 
151 	credits = DIV_ROUND_UP(len, ep->tx_credit_size);
152 	spin_lock_bh(&htc->tx_lock);
153 	ep->tx_credits += credits;
154 	ath10k_dbg(ar, ATH10K_DBG_HTC,
155 		   "htc ep %d reverted %d credits back total %d\n",
156 		   eid, credits, ep->tx_credits);
157 	spin_unlock_bh(&htc->tx_lock);
158 
159 	if (ep->ep_ops.ep_tx_credits)
160 		ep->ep_ops.ep_tx_credits(htc->ar);
161 }
162 
ath10k_htc_send(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,struct sk_buff * skb)163 int ath10k_htc_send(struct ath10k_htc *htc,
164 		    enum ath10k_htc_ep_id eid,
165 		    struct sk_buff *skb)
166 {
167 	struct ath10k *ar = htc->ar;
168 	struct ath10k_htc_ep *ep = &htc->endpoint[eid];
169 	struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
170 	struct ath10k_hif_sg_item sg_item;
171 	struct device *dev = htc->ar->dev;
172 	int ret;
173 	unsigned int skb_len;
174 
175 	if (htc->ar->state == ATH10K_STATE_WEDGED)
176 		return -ECOMM;
177 
178 	if (eid >= ATH10K_HTC_EP_COUNT) {
179 		ath10k_warn(ar, "Invalid endpoint id: %d\n", eid);
180 		return -ENOENT;
181 	}
182 
183 	skb_push(skb, sizeof(struct ath10k_htc_hdr));
184 
185 	skb_len = skb->len;
186 	ret = ath10k_htc_consume_credit(ep, skb_len, true);
187 	if (ret)
188 		goto err_pull;
189 
190 	ath10k_htc_prepare_tx_skb(ep, skb);
191 
192 	skb_cb->eid = eid;
193 	if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL) {
194 		skb_cb->paddr = dma_map_single(dev, skb->data, skb->len,
195 					       DMA_TO_DEVICE);
196 		ret = dma_mapping_error(dev, skb_cb->paddr);
197 		if (ret) {
198 			ret = -EIO;
199 			goto err_credits;
200 		}
201 	}
202 
203 	sg_item.transfer_id = ep->eid;
204 	sg_item.transfer_context = skb;
205 	sg_item.vaddr = skb->data;
206 	sg_item.paddr = skb_cb->paddr;
207 	sg_item.len = skb->len;
208 
209 	ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
210 	if (ret)
211 		goto err_unmap;
212 
213 	return 0;
214 
215 err_unmap:
216 	if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
217 		dma_unmap_single(dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
218 err_credits:
219 	ath10k_htc_release_credit(ep, skb_len);
220 err_pull:
221 	skb_pull(skb, sizeof(struct ath10k_htc_hdr));
222 	return ret;
223 }
224 
ath10k_htc_tx_completion_handler(struct ath10k * ar,struct sk_buff * skb)225 void ath10k_htc_tx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
226 {
227 	struct ath10k_htc *htc = &ar->htc;
228 	struct ath10k_skb_cb *skb_cb;
229 	struct ath10k_htc_ep *ep;
230 
231 	if (WARN_ON_ONCE(!skb))
232 		return;
233 
234 	skb_cb = ATH10K_SKB_CB(skb);
235 	ep = &htc->endpoint[skb_cb->eid];
236 
237 	ath10k_htc_notify_tx_completion(ep, skb);
238 	/* the skb now belongs to the completion handler */
239 }
240 EXPORT_SYMBOL(ath10k_htc_tx_completion_handler);
241 
242 /***********/
243 /* Receive */
244 /***********/
245 
246 static void
ath10k_htc_process_credit_report(struct ath10k_htc * htc,const struct ath10k_htc_credit_report * report,int len,enum ath10k_htc_ep_id eid)247 ath10k_htc_process_credit_report(struct ath10k_htc *htc,
248 				 const struct ath10k_htc_credit_report *report,
249 				 int len,
250 				 enum ath10k_htc_ep_id eid)
251 {
252 	struct ath10k *ar = htc->ar;
253 	struct ath10k_htc_ep *ep;
254 	int i, n_reports;
255 
256 	if (len % sizeof(*report))
257 		ath10k_warn(ar, "Uneven credit report len %d", len);
258 
259 	n_reports = len / sizeof(*report);
260 
261 	spin_lock_bh(&htc->tx_lock);
262 	for (i = 0; i < n_reports; i++, report++) {
263 		if (report->eid >= ATH10K_HTC_EP_COUNT)
264 			break;
265 
266 		ep = &htc->endpoint[report->eid];
267 		ep->tx_credits += report->credits;
268 
269 		ath10k_dbg(ar, ATH10K_DBG_HTC, "htc ep %d got %d credits (total %d)\n",
270 			   report->eid, report->credits, ep->tx_credits);
271 
272 		if (ep->ep_ops.ep_tx_credits) {
273 			spin_unlock_bh(&htc->tx_lock);
274 			ep->ep_ops.ep_tx_credits(htc->ar);
275 			spin_lock_bh(&htc->tx_lock);
276 		}
277 	}
278 	spin_unlock_bh(&htc->tx_lock);
279 }
280 
281 static int
ath10k_htc_process_lookahead(struct ath10k_htc * htc,const struct ath10k_htc_lookahead_report * report,int len,enum ath10k_htc_ep_id eid,void * next_lookaheads,int * next_lookaheads_len)282 ath10k_htc_process_lookahead(struct ath10k_htc *htc,
283 			     const struct ath10k_htc_lookahead_report *report,
284 			     int len,
285 			     enum ath10k_htc_ep_id eid,
286 			     void *next_lookaheads,
287 			     int *next_lookaheads_len)
288 {
289 	struct ath10k *ar = htc->ar;
290 
291 	/* Invalid lookahead flags are actually transmitted by
292 	 * the target in the HTC control message.
293 	 * Since this will happen at every boot we silently ignore
294 	 * the lookahead in this case
295 	 */
296 	if (report->pre_valid != ((~report->post_valid) & 0xFF))
297 		return 0;
298 
299 	if (next_lookaheads && next_lookaheads_len) {
300 		ath10k_dbg(ar, ATH10K_DBG_HTC,
301 			   "htc rx lookahead found pre_valid 0x%x post_valid 0x%x\n",
302 			   report->pre_valid, report->post_valid);
303 
304 		/* look ahead bytes are valid, copy them over */
305 		memcpy((u8 *)next_lookaheads, report->lookahead, 4);
306 
307 		*next_lookaheads_len = 1;
308 	}
309 
310 	return 0;
311 }
312 
313 static int
ath10k_htc_process_lookahead_bundle(struct ath10k_htc * htc,const struct ath10k_htc_lookahead_bundle * report,int len,enum ath10k_htc_ep_id eid,void * next_lookaheads,int * next_lookaheads_len)314 ath10k_htc_process_lookahead_bundle(struct ath10k_htc *htc,
315 				    const struct ath10k_htc_lookahead_bundle *report,
316 				    int len,
317 				    enum ath10k_htc_ep_id eid,
318 				    void *next_lookaheads,
319 				    int *next_lookaheads_len)
320 {
321 	struct ath10k *ar = htc->ar;
322 	int bundle_cnt = len / sizeof(*report);
323 
324 	if (!bundle_cnt || (bundle_cnt > htc->max_msgs_per_htc_bundle)) {
325 		ath10k_warn(ar, "Invalid lookahead bundle count: %d\n",
326 			    bundle_cnt);
327 		return -EINVAL;
328 	}
329 
330 	if (next_lookaheads && next_lookaheads_len) {
331 		int i;
332 
333 		for (i = 0; i < bundle_cnt; i++) {
334 			memcpy(((u8 *)next_lookaheads) + 4 * i,
335 			       report->lookahead, 4);
336 			report++;
337 		}
338 
339 		*next_lookaheads_len = bundle_cnt;
340 	}
341 
342 	return 0;
343 }
344 
ath10k_htc_process_trailer(struct ath10k_htc * htc,u8 * buffer,int length,enum ath10k_htc_ep_id src_eid,void * next_lookaheads,int * next_lookaheads_len)345 int ath10k_htc_process_trailer(struct ath10k_htc *htc,
346 			       u8 *buffer,
347 			       int length,
348 			       enum ath10k_htc_ep_id src_eid,
349 			       void *next_lookaheads,
350 			       int *next_lookaheads_len)
351 {
352 	struct ath10k_htc_lookahead_bundle *bundle;
353 	struct ath10k *ar = htc->ar;
354 	int status = 0;
355 	struct ath10k_htc_record *record;
356 	u8 *orig_buffer;
357 	int orig_length;
358 	size_t len;
359 
360 	orig_buffer = buffer;
361 	orig_length = length;
362 
363 	while (length > 0) {
364 		record = (struct ath10k_htc_record *)buffer;
365 
366 		if (length < sizeof(record->hdr)) {
367 			status = -EINVAL;
368 			break;
369 		}
370 
371 		if (record->hdr.len > length) {
372 			/* no room left in buffer for record */
373 			ath10k_warn(ar, "Invalid record length: %d\n",
374 				    record->hdr.len);
375 			status = -EINVAL;
376 			break;
377 		}
378 
379 		switch (record->hdr.id) {
380 		case ATH10K_HTC_RECORD_CREDITS:
381 			len = sizeof(struct ath10k_htc_credit_report);
382 			if (record->hdr.len < len) {
383 				ath10k_warn(ar, "Credit report too long\n");
384 				status = -EINVAL;
385 				break;
386 			}
387 			ath10k_htc_process_credit_report(htc,
388 							 record->credit_report,
389 							 record->hdr.len,
390 							 src_eid);
391 			break;
392 		case ATH10K_HTC_RECORD_LOOKAHEAD:
393 			len = sizeof(struct ath10k_htc_lookahead_report);
394 			if (record->hdr.len < len) {
395 				ath10k_warn(ar, "Lookahead report too long\n");
396 				status = -EINVAL;
397 				break;
398 			}
399 			status = ath10k_htc_process_lookahead(htc,
400 							      record->lookahead_report,
401 							      record->hdr.len,
402 							      src_eid,
403 							      next_lookaheads,
404 							      next_lookaheads_len);
405 			break;
406 		case ATH10K_HTC_RECORD_LOOKAHEAD_BUNDLE:
407 			bundle = record->lookahead_bundle;
408 			status = ath10k_htc_process_lookahead_bundle(htc,
409 								     bundle,
410 								     record->hdr.len,
411 								     src_eid,
412 								     next_lookaheads,
413 								     next_lookaheads_len);
414 			break;
415 		default:
416 			ath10k_warn(ar, "Unhandled record: id:%d length:%d\n",
417 				    record->hdr.id, record->hdr.len);
418 			break;
419 		}
420 
421 		if (status)
422 			break;
423 
424 		/* multiple records may be present in a trailer */
425 		buffer += sizeof(record->hdr) + record->hdr.len;
426 		length -= sizeof(record->hdr) + record->hdr.len;
427 	}
428 
429 	if (status)
430 		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc rx bad trailer", "",
431 				orig_buffer, orig_length);
432 
433 	return status;
434 }
435 EXPORT_SYMBOL(ath10k_htc_process_trailer);
436 
ath10k_htc_rx_completion_handler(struct ath10k * ar,struct sk_buff * skb)437 void ath10k_htc_rx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
438 {
439 	int status = 0;
440 	struct ath10k_htc *htc = &ar->htc;
441 	struct ath10k_htc_hdr *hdr;
442 	struct ath10k_htc_ep *ep;
443 	u16 payload_len;
444 	u32 trailer_len = 0;
445 	size_t min_len;
446 	u8 eid;
447 	bool trailer_present;
448 
449 	hdr = (struct ath10k_htc_hdr *)skb->data;
450 	skb_pull(skb, sizeof(*hdr));
451 
452 	eid = hdr->eid;
453 
454 	if (eid >= ATH10K_HTC_EP_COUNT) {
455 		ath10k_warn(ar, "HTC Rx: invalid eid %d\n", eid);
456 		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad header", "",
457 				hdr, sizeof(*hdr));
458 		goto out;
459 	}
460 
461 	ep = &htc->endpoint[eid];
462 	if (ep->service_id == ATH10K_HTC_SVC_ID_UNUSED) {
463 		ath10k_warn(ar, "htc rx endpoint %d is not connected\n", eid);
464 		goto out;
465 	}
466 
467 	payload_len = __le16_to_cpu(hdr->len);
468 
469 	if (payload_len + sizeof(*hdr) > ATH10K_HTC_MAX_LEN) {
470 		ath10k_warn(ar, "HTC rx frame too long, len: %zu\n",
471 			    payload_len + sizeof(*hdr));
472 		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len", "",
473 				hdr, sizeof(*hdr));
474 		goto out;
475 	}
476 
477 	if (skb->len < payload_len) {
478 		ath10k_dbg(ar, ATH10K_DBG_HTC,
479 			   "HTC Rx: insufficient length, got %d, expected %d\n",
480 			   skb->len, payload_len);
481 		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len",
482 				"", hdr, sizeof(*hdr));
483 		goto out;
484 	}
485 
486 	/* get flags to check for trailer */
487 	trailer_present = hdr->flags & ATH10K_HTC_FLAG_TRAILER_PRESENT;
488 	if (trailer_present) {
489 		u8 *trailer;
490 
491 		trailer_len = hdr->trailer_len;
492 		min_len = sizeof(struct ath10k_ath10k_htc_record_hdr);
493 
494 		if ((trailer_len < min_len) ||
495 		    (trailer_len > payload_len)) {
496 			ath10k_warn(ar, "Invalid trailer length: %d\n",
497 				    trailer_len);
498 			goto out;
499 		}
500 
501 		trailer = (u8 *)hdr;
502 		trailer += sizeof(*hdr);
503 		trailer += payload_len;
504 		trailer -= trailer_len;
505 		status = ath10k_htc_process_trailer(htc, trailer,
506 						    trailer_len, hdr->eid,
507 						    NULL, NULL);
508 		if (status)
509 			goto out;
510 
511 		skb_trim(skb, skb->len - trailer_len);
512 	}
513 
514 	if (((int)payload_len - (int)trailer_len) <= 0)
515 		/* zero length packet with trailer data, just drop these */
516 		goto out;
517 
518 	ath10k_dbg(ar, ATH10K_DBG_HTC, "htc rx completion ep %d skb %pK\n",
519 		   eid, skb);
520 	ep->ep_ops.ep_rx_complete(ar, skb);
521 
522 	/* skb is now owned by the rx completion handler */
523 	skb = NULL;
524 out:
525 	kfree_skb(skb);
526 }
527 EXPORT_SYMBOL(ath10k_htc_rx_completion_handler);
528 
ath10k_htc_control_rx_complete(struct ath10k * ar,struct sk_buff * skb)529 static void ath10k_htc_control_rx_complete(struct ath10k *ar,
530 					   struct sk_buff *skb)
531 {
532 	struct ath10k_htc *htc = &ar->htc;
533 	struct ath10k_htc_msg *msg = (struct ath10k_htc_msg *)skb->data;
534 
535 	switch (__le16_to_cpu(msg->hdr.message_id)) {
536 	case ATH10K_HTC_MSG_READY_ID:
537 	case ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID:
538 		/* handle HTC control message */
539 		if (completion_done(&htc->ctl_resp)) {
540 			/* this is a fatal error, target should not be
541 			 * sending unsolicited messages on the ep 0
542 			 */
543 			ath10k_warn(ar, "HTC rx ctrl still processing\n");
544 			complete(&htc->ctl_resp);
545 			goto out;
546 		}
547 
548 		htc->control_resp_len =
549 			min_t(int, skb->len,
550 			      ATH10K_HTC_MAX_CTRL_MSG_LEN);
551 
552 		memcpy(htc->control_resp_buffer, skb->data,
553 		       htc->control_resp_len);
554 
555 		complete(&htc->ctl_resp);
556 		break;
557 	case ATH10K_HTC_MSG_SEND_SUSPEND_COMPLETE:
558 		htc->htc_ops.target_send_suspend_complete(ar);
559 		break;
560 	default:
561 		ath10k_warn(ar, "ignoring unsolicited htc ep0 event\n");
562 		break;
563 	}
564 
565 out:
566 	kfree_skb(skb);
567 }
568 
569 /***************/
570 /* Init/Deinit */
571 /***************/
572 
htc_service_name(enum ath10k_htc_svc_id id)573 static const char *htc_service_name(enum ath10k_htc_svc_id id)
574 {
575 	switch (id) {
576 	case ATH10K_HTC_SVC_ID_RESERVED:
577 		return "Reserved";
578 	case ATH10K_HTC_SVC_ID_RSVD_CTRL:
579 		return "Control";
580 	case ATH10K_HTC_SVC_ID_WMI_CONTROL:
581 		return "WMI";
582 	case ATH10K_HTC_SVC_ID_WMI_DATA_BE:
583 		return "DATA BE";
584 	case ATH10K_HTC_SVC_ID_WMI_DATA_BK:
585 		return "DATA BK";
586 	case ATH10K_HTC_SVC_ID_WMI_DATA_VI:
587 		return "DATA VI";
588 	case ATH10K_HTC_SVC_ID_WMI_DATA_VO:
589 		return "DATA VO";
590 	case ATH10K_HTC_SVC_ID_NMI_CONTROL:
591 		return "NMI Control";
592 	case ATH10K_HTC_SVC_ID_NMI_DATA:
593 		return "NMI Data";
594 	case ATH10K_HTC_SVC_ID_HTT_DATA_MSG:
595 		return "HTT Data";
596 	case ATH10K_HTC_SVC_ID_HTT_DATA2_MSG:
597 		return "HTT Data";
598 	case ATH10K_HTC_SVC_ID_HTT_DATA3_MSG:
599 		return "HTT Data";
600 	case ATH10K_HTC_SVC_ID_TEST_RAW_STREAMS:
601 		return "RAW";
602 	case ATH10K_HTC_SVC_ID_HTT_LOG_MSG:
603 		return "PKTLOG";
604 	}
605 
606 	return "Unknown";
607 }
608 
ath10k_htc_reset_endpoint_states(struct ath10k_htc * htc)609 static void ath10k_htc_reset_endpoint_states(struct ath10k_htc *htc)
610 {
611 	struct ath10k_htc_ep *ep;
612 	int i;
613 
614 	for (i = ATH10K_HTC_EP_0; i < ATH10K_HTC_EP_COUNT; i++) {
615 		ep = &htc->endpoint[i];
616 		ep->service_id = ATH10K_HTC_SVC_ID_UNUSED;
617 		ep->max_ep_message_len = 0;
618 		ep->max_tx_queue_depth = 0;
619 		ep->eid = i;
620 		ep->htc = htc;
621 		ep->tx_credit_flow_enabled = true;
622 	}
623 }
624 
ath10k_htc_get_credit_allocation(struct ath10k_htc * htc,u16 service_id)625 static u8 ath10k_htc_get_credit_allocation(struct ath10k_htc *htc,
626 					   u16 service_id)
627 {
628 	u8 allocation = 0;
629 
630 	/* The WMI control service is the only service with flow control.
631 	 * Let it have all transmit credits.
632 	 */
633 	if (service_id == ATH10K_HTC_SVC_ID_WMI_CONTROL)
634 		allocation = htc->total_transmit_credits;
635 
636 	return allocation;
637 }
638 
ath10k_htc_send_bundle(struct ath10k_htc_ep * ep,struct sk_buff * bundle_skb,struct sk_buff_head * tx_save_head)639 static int ath10k_htc_send_bundle(struct ath10k_htc_ep *ep,
640 				  struct sk_buff *bundle_skb,
641 				  struct sk_buff_head *tx_save_head)
642 {
643 	struct ath10k_hif_sg_item sg_item;
644 	struct ath10k_htc *htc = ep->htc;
645 	struct ath10k *ar = htc->ar;
646 	struct sk_buff *skb;
647 	int ret, cn = 0;
648 	unsigned int skb_len;
649 
650 	ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle skb len %d\n", bundle_skb->len);
651 	skb_len = bundle_skb->len;
652 	ret = ath10k_htc_consume_credit(ep, skb_len, true);
653 
654 	if (!ret) {
655 		sg_item.transfer_id = ep->eid;
656 		sg_item.transfer_context = bundle_skb;
657 		sg_item.vaddr = bundle_skb->data;
658 		sg_item.len = bundle_skb->len;
659 
660 		ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
661 		if (ret)
662 			ath10k_htc_release_credit(ep, skb_len);
663 	}
664 
665 	if (ret)
666 		dev_kfree_skb_any(bundle_skb);
667 
668 	for (cn = 0; (skb = skb_dequeue_tail(tx_save_head)); cn++) {
669 		if (ret) {
670 			skb_pull(skb, sizeof(struct ath10k_htc_hdr));
671 			skb_queue_head(&ep->tx_req_head, skb);
672 		} else {
673 			skb_queue_tail(&ep->tx_complete_head, skb);
674 		}
675 	}
676 
677 	if (!ret)
678 		queue_work(ar->workqueue_tx_complete, &ar->tx_complete_work);
679 
680 	ath10k_dbg(ar, ATH10K_DBG_HTC,
681 		   "bundle tx status %d eid %d req count %d count %d len %d\n",
682 		   ret, ep->eid, skb_queue_len(&ep->tx_req_head), cn, skb_len);
683 	return ret;
684 }
685 
ath10k_htc_send_one_skb(struct ath10k_htc_ep * ep,struct sk_buff * skb)686 static void ath10k_htc_send_one_skb(struct ath10k_htc_ep *ep, struct sk_buff *skb)
687 {
688 	struct ath10k_htc *htc = ep->htc;
689 	struct ath10k *ar = htc->ar;
690 	int ret;
691 
692 	ret = ath10k_htc_send(htc, ep->eid, skb);
693 
694 	if (ret)
695 		skb_queue_head(&ep->tx_req_head, skb);
696 
697 	ath10k_dbg(ar, ATH10K_DBG_HTC, "tx one status %d eid %d len %d pending count %d\n",
698 		   ret, ep->eid, skb->len, skb_queue_len(&ep->tx_req_head));
699 }
700 
ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep * ep)701 static int ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep *ep)
702 {
703 	struct ath10k_htc *htc = ep->htc;
704 	struct sk_buff *bundle_skb, *skb;
705 	struct sk_buff_head tx_save_head;
706 	struct ath10k_htc_hdr *hdr;
707 	u8 *bundle_buf;
708 	int ret = 0, credit_pad, credit_remainder, trans_len, bundles_left = 0;
709 
710 	if (htc->ar->state == ATH10K_STATE_WEDGED)
711 		return -ECOMM;
712 
713 	if (ep->tx_credit_flow_enabled &&
714 	    ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE)
715 		return 0;
716 
717 	bundles_left = ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
718 	bundle_skb = dev_alloc_skb(bundles_left);
719 
720 	if (!bundle_skb)
721 		return -ENOMEM;
722 
723 	bundle_buf = bundle_skb->data;
724 	skb_queue_head_init(&tx_save_head);
725 
726 	while (true) {
727 		skb = skb_dequeue(&ep->tx_req_head);
728 		if (!skb)
729 			break;
730 
731 		credit_pad = 0;
732 		trans_len = skb->len + sizeof(*hdr);
733 		credit_remainder = trans_len % ep->tx_credit_size;
734 
735 		if (credit_remainder != 0) {
736 			credit_pad = ep->tx_credit_size - credit_remainder;
737 			trans_len += credit_pad;
738 		}
739 
740 		ret = ath10k_htc_consume_credit(ep,
741 						bundle_buf + trans_len - bundle_skb->data,
742 						false);
743 		if (ret) {
744 			skb_queue_head(&ep->tx_req_head, skb);
745 			break;
746 		}
747 
748 		if (bundles_left < trans_len) {
749 			bundle_skb->len = bundle_buf - bundle_skb->data;
750 			ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
751 
752 			if (ret) {
753 				skb_queue_head(&ep->tx_req_head, skb);
754 				return ret;
755 			}
756 
757 			if (skb_queue_len(&ep->tx_req_head) == 0) {
758 				ath10k_htc_send_one_skb(ep, skb);
759 				return ret;
760 			}
761 
762 			if (ep->tx_credit_flow_enabled &&
763 			    ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE) {
764 				skb_queue_head(&ep->tx_req_head, skb);
765 				return 0;
766 			}
767 
768 			bundles_left =
769 				ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
770 			bundle_skb = dev_alloc_skb(bundles_left);
771 
772 			if (!bundle_skb) {
773 				skb_queue_head(&ep->tx_req_head, skb);
774 				return -ENOMEM;
775 			}
776 			bundle_buf = bundle_skb->data;
777 			skb_queue_head_init(&tx_save_head);
778 		}
779 
780 		skb_push(skb, sizeof(struct ath10k_htc_hdr));
781 		ath10k_htc_prepare_tx_skb(ep, skb);
782 
783 		memcpy(bundle_buf, skb->data, skb->len);
784 		hdr = (struct ath10k_htc_hdr *)bundle_buf;
785 		hdr->flags |= ATH10K_HTC_FLAG_SEND_BUNDLE;
786 		hdr->pad_len = __cpu_to_le16(credit_pad);
787 		bundle_buf += trans_len;
788 		bundles_left -= trans_len;
789 		skb_queue_tail(&tx_save_head, skb);
790 	}
791 
792 	if (bundle_buf != bundle_skb->data) {
793 		bundle_skb->len = bundle_buf - bundle_skb->data;
794 		ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
795 	} else {
796 		dev_kfree_skb_any(bundle_skb);
797 	}
798 
799 	return ret;
800 }
801 
ath10k_htc_bundle_tx_work(struct work_struct * work)802 static void ath10k_htc_bundle_tx_work(struct work_struct *work)
803 {
804 	struct ath10k *ar = container_of(work, struct ath10k, bundle_tx_work);
805 	struct ath10k_htc_ep *ep;
806 	struct sk_buff *skb;
807 	int i;
808 
809 	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
810 		ep = &ar->htc.endpoint[i];
811 
812 		if (!ep->bundle_tx)
813 			continue;
814 
815 		ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx work eid %d count %d\n",
816 			   ep->eid, skb_queue_len(&ep->tx_req_head));
817 
818 		if (skb_queue_len(&ep->tx_req_head) >=
819 		    ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE) {
820 			ath10k_htc_send_bundle_skbs(ep);
821 		} else {
822 			skb = skb_dequeue(&ep->tx_req_head);
823 
824 			if (!skb)
825 				continue;
826 			ath10k_htc_send_one_skb(ep, skb);
827 		}
828 	}
829 }
830 
ath10k_htc_tx_complete_work(struct work_struct * work)831 static void ath10k_htc_tx_complete_work(struct work_struct *work)
832 {
833 	struct ath10k *ar = container_of(work, struct ath10k, tx_complete_work);
834 	struct ath10k_htc_ep *ep;
835 	enum ath10k_htc_ep_id eid;
836 	struct sk_buff *skb;
837 	int i;
838 
839 	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
840 		ep = &ar->htc.endpoint[i];
841 		eid = ep->eid;
842 		if (ep->bundle_tx && eid == ar->htt.eid) {
843 			ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx complete eid %d pending complete count%d\n",
844 				   ep->eid, skb_queue_len(&ep->tx_complete_head));
845 
846 			while (true) {
847 				skb = skb_dequeue(&ep->tx_complete_head);
848 				if (!skb)
849 					break;
850 				ath10k_htc_notify_tx_completion(ep, skb);
851 			}
852 		}
853 	}
854 }
855 
ath10k_htc_send_hl(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,struct sk_buff * skb)856 int ath10k_htc_send_hl(struct ath10k_htc *htc,
857 		       enum ath10k_htc_ep_id eid,
858 		       struct sk_buff *skb)
859 {
860 	struct ath10k_htc_ep *ep = &htc->endpoint[eid];
861 	struct ath10k *ar = htc->ar;
862 
863 	if (sizeof(struct ath10k_htc_hdr) + skb->len > ep->tx_credit_size) {
864 		ath10k_dbg(ar, ATH10K_DBG_HTC, "tx exceed max len %d\n", skb->len);
865 		return -ENOMEM;
866 	}
867 
868 	ath10k_dbg(ar, ATH10K_DBG_HTC, "htc send hl eid %d bundle %d tx count %d len %d\n",
869 		   eid, ep->bundle_tx, skb_queue_len(&ep->tx_req_head), skb->len);
870 
871 	if (ep->bundle_tx) {
872 		skb_queue_tail(&ep->tx_req_head, skb);
873 		queue_work(ar->workqueue, &ar->bundle_tx_work);
874 		return 0;
875 	} else {
876 		return ath10k_htc_send(htc, eid, skb);
877 	}
878 }
879 
ath10k_htc_setup_tx_req(struct ath10k_htc_ep * ep)880 void ath10k_htc_setup_tx_req(struct ath10k_htc_ep *ep)
881 {
882 	if (ep->htc->max_msgs_per_htc_bundle >= ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE &&
883 	    !ep->bundle_tx) {
884 		ep->bundle_tx = true;
885 		skb_queue_head_init(&ep->tx_req_head);
886 		skb_queue_head_init(&ep->tx_complete_head);
887 	}
888 }
889 
ath10k_htc_stop_hl(struct ath10k * ar)890 void ath10k_htc_stop_hl(struct ath10k *ar)
891 {
892 	struct ath10k_htc_ep *ep;
893 	int i;
894 
895 	cancel_work_sync(&ar->bundle_tx_work);
896 	cancel_work_sync(&ar->tx_complete_work);
897 
898 	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
899 		ep = &ar->htc.endpoint[i];
900 
901 		if (!ep->bundle_tx)
902 			continue;
903 
904 		ath10k_dbg(ar, ATH10K_DBG_HTC, "stop tx work eid %d count %d\n",
905 			   ep->eid, skb_queue_len(&ep->tx_req_head));
906 
907 		skb_queue_purge(&ep->tx_req_head);
908 	}
909 }
910 
ath10k_htc_wait_target(struct ath10k_htc * htc)911 int ath10k_htc_wait_target(struct ath10k_htc *htc)
912 {
913 	struct ath10k *ar = htc->ar;
914 	int i, status = 0;
915 	unsigned long time_left;
916 	struct ath10k_htc_msg *msg;
917 	u16 message_id;
918 
919 	time_left = wait_for_completion_timeout(&htc->ctl_resp,
920 						ATH10K_HTC_WAIT_TIMEOUT_HZ);
921 	if (!time_left) {
922 		/* Workaround: In some cases the PCI HIF doesn't
923 		 * receive interrupt for the control response message
924 		 * even if the buffer was completed. It is suspected
925 		 * iomap writes unmasking PCI CE irqs aren't propagated
926 		 * properly in KVM PCI-passthrough sometimes.
927 		 */
928 		ath10k_warn(ar, "failed to receive control response completion, polling..\n");
929 
930 		for (i = 0; i < CE_COUNT; i++)
931 			ath10k_hif_send_complete_check(htc->ar, i, 1);
932 
933 		time_left =
934 		wait_for_completion_timeout(&htc->ctl_resp,
935 					    ATH10K_HTC_WAIT_TIMEOUT_HZ);
936 
937 		if (!time_left)
938 			status = -ETIMEDOUT;
939 	}
940 
941 	if (status < 0) {
942 		ath10k_err(ar, "ctl_resp never came in (%d)\n", status);
943 		return status;
944 	}
945 
946 	if (htc->control_resp_len < sizeof(msg->hdr) + sizeof(msg->ready)) {
947 		ath10k_err(ar, "Invalid HTC ready msg len:%d\n",
948 			   htc->control_resp_len);
949 		return -ECOMM;
950 	}
951 
952 	msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
953 	message_id   = __le16_to_cpu(msg->hdr.message_id);
954 
955 	if (message_id != ATH10K_HTC_MSG_READY_ID) {
956 		ath10k_err(ar, "Invalid HTC ready msg: 0x%x\n", message_id);
957 		return -ECOMM;
958 	}
959 
960 	if (ar->hw_params.use_fw_tx_credits)
961 		htc->total_transmit_credits = __le16_to_cpu(msg->ready.credit_count);
962 	else
963 		htc->total_transmit_credits = 1;
964 
965 	htc->target_credit_size = __le16_to_cpu(msg->ready.credit_size);
966 
967 	ath10k_dbg(ar, ATH10K_DBG_HTC,
968 		   "Target ready! transmit resources: %d size:%d actual credits:%d\n",
969 		   htc->total_transmit_credits,
970 		   htc->target_credit_size,
971 		   msg->ready.credit_count);
972 
973 	if ((htc->total_transmit_credits == 0) ||
974 	    (htc->target_credit_size == 0)) {
975 		ath10k_err(ar, "Invalid credit size received\n");
976 		return -ECOMM;
977 	}
978 
979 	/* The only way to determine if the ready message is an extended
980 	 * message is from the size.
981 	 */
982 	if (htc->control_resp_len >=
983 	    sizeof(msg->hdr) + sizeof(msg->ready_ext)) {
984 		htc->alt_data_credit_size =
985 			__le16_to_cpu(msg->ready_ext.reserved) &
986 			ATH10K_HTC_MSG_READY_EXT_ALT_DATA_MASK;
987 		htc->max_msgs_per_htc_bundle =
988 			min_t(u8, msg->ready_ext.max_msgs_per_htc_bundle,
989 			      HTC_HOST_MAX_MSG_PER_RX_BUNDLE);
990 		ath10k_dbg(ar, ATH10K_DBG_HTC,
991 			   "Extended ready message RX bundle size %d alt size %d\n",
992 			   htc->max_msgs_per_htc_bundle,
993 			   htc->alt_data_credit_size);
994 	}
995 
996 	INIT_WORK(&ar->bundle_tx_work, ath10k_htc_bundle_tx_work);
997 	INIT_WORK(&ar->tx_complete_work, ath10k_htc_tx_complete_work);
998 
999 	return 0;
1000 }
1001 
ath10k_htc_change_tx_credit_flow(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,bool enable)1002 void ath10k_htc_change_tx_credit_flow(struct ath10k_htc *htc,
1003 				      enum ath10k_htc_ep_id eid,
1004 				      bool enable)
1005 {
1006 	struct ath10k *ar = htc->ar;
1007 	struct ath10k_htc_ep *ep = &ar->htc.endpoint[eid];
1008 
1009 	ep->tx_credit_flow_enabled = enable;
1010 }
1011 
ath10k_htc_connect_service(struct ath10k_htc * htc,struct ath10k_htc_svc_conn_req * conn_req,struct ath10k_htc_svc_conn_resp * conn_resp)1012 int ath10k_htc_connect_service(struct ath10k_htc *htc,
1013 			       struct ath10k_htc_svc_conn_req *conn_req,
1014 			       struct ath10k_htc_svc_conn_resp *conn_resp)
1015 {
1016 	struct ath10k *ar = htc->ar;
1017 	struct ath10k_htc_msg *msg;
1018 	struct ath10k_htc_conn_svc *req_msg;
1019 	struct ath10k_htc_conn_svc_response resp_msg_dummy;
1020 	struct ath10k_htc_conn_svc_response *resp_msg = &resp_msg_dummy;
1021 	enum ath10k_htc_ep_id assigned_eid = ATH10K_HTC_EP_COUNT;
1022 	struct ath10k_htc_ep *ep;
1023 	struct sk_buff *skb;
1024 	unsigned int max_msg_size = 0;
1025 	int length, status;
1026 	unsigned long time_left;
1027 	bool disable_credit_flow_ctrl = false;
1028 	u16 message_id, service_id, flags = 0;
1029 	u8 tx_alloc = 0;
1030 
1031 	/* special case for HTC pseudo control service */
1032 	if (conn_req->service_id == ATH10K_HTC_SVC_ID_RSVD_CTRL) {
1033 		disable_credit_flow_ctrl = true;
1034 		assigned_eid = ATH10K_HTC_EP_0;
1035 		max_msg_size = ATH10K_HTC_MAX_CTRL_MSG_LEN;
1036 		memset(&resp_msg_dummy, 0, sizeof(resp_msg_dummy));
1037 		goto setup;
1038 	}
1039 
1040 	tx_alloc = ath10k_htc_get_credit_allocation(htc,
1041 						    conn_req->service_id);
1042 	if (!tx_alloc)
1043 		ath10k_dbg(ar, ATH10K_DBG_BOOT,
1044 			   "boot htc service %s does not allocate target credits\n",
1045 			   htc_service_name(conn_req->service_id));
1046 
1047 	skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1048 	if (!skb) {
1049 		ath10k_err(ar, "Failed to allocate HTC packet\n");
1050 		return -ENOMEM;
1051 	}
1052 
1053 	length = sizeof(msg->hdr) + sizeof(msg->connect_service);
1054 	skb_put(skb, length);
1055 	memset(skb->data, 0, length);
1056 
1057 	msg = (struct ath10k_htc_msg *)skb->data;
1058 	msg->hdr.message_id =
1059 		__cpu_to_le16(ATH10K_HTC_MSG_CONNECT_SERVICE_ID);
1060 
1061 	flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC);
1062 
1063 	/* Only enable credit flow control for WMI ctrl service */
1064 	if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) {
1065 		flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL;
1066 		disable_credit_flow_ctrl = true;
1067 	}
1068 
1069 	req_msg = &msg->connect_service;
1070 	req_msg->flags = __cpu_to_le16(flags);
1071 	req_msg->service_id = __cpu_to_le16(conn_req->service_id);
1072 
1073 	reinit_completion(&htc->ctl_resp);
1074 
1075 	status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1076 	if (status) {
1077 		kfree_skb(skb);
1078 		return status;
1079 	}
1080 
1081 	/* wait for response */
1082 	time_left = wait_for_completion_timeout(&htc->ctl_resp,
1083 						ATH10K_HTC_CONN_SVC_TIMEOUT_HZ);
1084 	if (!time_left) {
1085 		ath10k_err(ar, "Service connect timeout\n");
1086 		return -ETIMEDOUT;
1087 	}
1088 
1089 	/* we controlled the buffer creation, it's aligned */
1090 	msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
1091 	resp_msg = &msg->connect_service_response;
1092 	message_id = __le16_to_cpu(msg->hdr.message_id);
1093 	service_id = __le16_to_cpu(resp_msg->service_id);
1094 
1095 	if ((message_id != ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID) ||
1096 	    (htc->control_resp_len < sizeof(msg->hdr) +
1097 	     sizeof(msg->connect_service_response))) {
1098 		ath10k_err(ar, "Invalid resp message ID 0x%x", message_id);
1099 		return -EPROTO;
1100 	}
1101 
1102 	ath10k_dbg(ar, ATH10K_DBG_HTC,
1103 		   "HTC Service %s connect response: status: 0x%x, assigned ep: 0x%x\n",
1104 		   htc_service_name(service_id),
1105 		   resp_msg->status, resp_msg->eid);
1106 
1107 	conn_resp->connect_resp_code = resp_msg->status;
1108 
1109 	/* check response status */
1110 	if (resp_msg->status != ATH10K_HTC_CONN_SVC_STATUS_SUCCESS) {
1111 		ath10k_err(ar, "HTC Service %s connect request failed: 0x%x)\n",
1112 			   htc_service_name(service_id),
1113 			   resp_msg->status);
1114 		return -EPROTO;
1115 	}
1116 
1117 	assigned_eid = (enum ath10k_htc_ep_id)resp_msg->eid;
1118 	max_msg_size = __le16_to_cpu(resp_msg->max_msg_size);
1119 
1120 setup:
1121 
1122 	if (assigned_eid >= ATH10K_HTC_EP_COUNT)
1123 		return -EPROTO;
1124 
1125 	if (max_msg_size == 0)
1126 		return -EPROTO;
1127 
1128 	ep = &htc->endpoint[assigned_eid];
1129 	ep->eid = assigned_eid;
1130 
1131 	if (ep->service_id != ATH10K_HTC_SVC_ID_UNUSED)
1132 		return -EPROTO;
1133 
1134 	/* return assigned endpoint to caller */
1135 	conn_resp->eid = assigned_eid;
1136 	conn_resp->max_msg_len = __le16_to_cpu(resp_msg->max_msg_size);
1137 
1138 	/* setup the endpoint */
1139 	ep->service_id = conn_req->service_id;
1140 	ep->max_tx_queue_depth = conn_req->max_send_queue_depth;
1141 	ep->max_ep_message_len = __le16_to_cpu(resp_msg->max_msg_size);
1142 	ep->tx_credits = tx_alloc;
1143 	ep->tx_credit_size = htc->target_credit_size;
1144 
1145 	if (conn_req->service_id == ATH10K_HTC_SVC_ID_HTT_DATA_MSG &&
1146 	    htc->alt_data_credit_size != 0)
1147 		ep->tx_credit_size = htc->alt_data_credit_size;
1148 
1149 	/* copy all the callbacks */
1150 	ep->ep_ops = conn_req->ep_ops;
1151 
1152 	status = ath10k_hif_map_service_to_pipe(htc->ar,
1153 						ep->service_id,
1154 						&ep->ul_pipe_id,
1155 						&ep->dl_pipe_id);
1156 	if (status) {
1157 		ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC service id: %d\n",
1158 			   ep->service_id);
1159 		return status;
1160 	}
1161 
1162 	ath10k_dbg(ar, ATH10K_DBG_BOOT,
1163 		   "boot htc service '%s' ul pipe %d dl pipe %d eid %d ready\n",
1164 		   htc_service_name(ep->service_id), ep->ul_pipe_id,
1165 		   ep->dl_pipe_id, ep->eid);
1166 
1167 	if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) {
1168 		ep->tx_credit_flow_enabled = false;
1169 		ath10k_dbg(ar, ATH10K_DBG_BOOT,
1170 			   "boot htc service '%s' eid %d TX flow control disabled\n",
1171 			   htc_service_name(ep->service_id), assigned_eid);
1172 	}
1173 
1174 	return status;
1175 }
1176 
ath10k_htc_alloc_skb(struct ath10k * ar,int size)1177 struct sk_buff *ath10k_htc_alloc_skb(struct ath10k *ar, int size)
1178 {
1179 	struct sk_buff *skb;
1180 
1181 	skb = dev_alloc_skb(size + sizeof(struct ath10k_htc_hdr));
1182 	if (!skb)
1183 		return NULL;
1184 
1185 	skb_reserve(skb, sizeof(struct ath10k_htc_hdr));
1186 
1187 	/* FW/HTC requires 4-byte aligned streams */
1188 	if (!IS_ALIGNED((unsigned long)skb->data, 4))
1189 		ath10k_warn(ar, "Unaligned HTC tx skb\n");
1190 
1191 	return skb;
1192 }
1193 
ath10k_htc_pktlog_process_rx(struct ath10k * ar,struct sk_buff * skb)1194 static void ath10k_htc_pktlog_process_rx(struct ath10k *ar, struct sk_buff *skb)
1195 {
1196 	trace_ath10k_htt_pktlog(ar, skb->data, skb->len);
1197 	dev_kfree_skb_any(skb);
1198 }
1199 
ath10k_htc_pktlog_connect(struct ath10k * ar)1200 static int ath10k_htc_pktlog_connect(struct ath10k *ar)
1201 {
1202 	struct ath10k_htc_svc_conn_resp conn_resp;
1203 	struct ath10k_htc_svc_conn_req conn_req;
1204 	int status;
1205 
1206 	memset(&conn_req, 0, sizeof(conn_req));
1207 	memset(&conn_resp, 0, sizeof(conn_resp));
1208 
1209 	conn_req.ep_ops.ep_tx_complete = NULL;
1210 	conn_req.ep_ops.ep_rx_complete = ath10k_htc_pktlog_process_rx;
1211 	conn_req.ep_ops.ep_tx_credits = NULL;
1212 
1213 	/* connect to control service */
1214 	conn_req.service_id = ATH10K_HTC_SVC_ID_HTT_LOG_MSG;
1215 	status = ath10k_htc_connect_service(&ar->htc, &conn_req, &conn_resp);
1216 	if (status) {
1217 		ath10k_warn(ar, "failed to connect to PKTLOG service: %d\n",
1218 			    status);
1219 		return status;
1220 	}
1221 
1222 	return 0;
1223 }
1224 
ath10k_htc_pktlog_svc_supported(struct ath10k * ar)1225 static bool ath10k_htc_pktlog_svc_supported(struct ath10k *ar)
1226 {
1227 	u8 ul_pipe_id;
1228 	u8 dl_pipe_id;
1229 	int status;
1230 
1231 	status = ath10k_hif_map_service_to_pipe(ar, ATH10K_HTC_SVC_ID_HTT_LOG_MSG,
1232 						&ul_pipe_id,
1233 						&dl_pipe_id);
1234 	if (status) {
1235 		ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC pktlog service id: %d\n",
1236 			   ATH10K_HTC_SVC_ID_HTT_LOG_MSG);
1237 
1238 		return false;
1239 	}
1240 
1241 	return true;
1242 }
1243 
ath10k_htc_start(struct ath10k_htc * htc)1244 int ath10k_htc_start(struct ath10k_htc *htc)
1245 {
1246 	struct ath10k *ar = htc->ar;
1247 	struct sk_buff *skb;
1248 	int status = 0;
1249 	struct ath10k_htc_msg *msg;
1250 
1251 	skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1252 	if (!skb)
1253 		return -ENOMEM;
1254 
1255 	skb_put(skb, sizeof(msg->hdr) + sizeof(msg->setup_complete_ext));
1256 	memset(skb->data, 0, skb->len);
1257 
1258 	msg = (struct ath10k_htc_msg *)skb->data;
1259 	msg->hdr.message_id =
1260 		__cpu_to_le16(ATH10K_HTC_MSG_SETUP_COMPLETE_EX_ID);
1261 
1262 	if (ar->hif.bus == ATH10K_BUS_SDIO) {
1263 		/* Extra setup params used by SDIO */
1264 		msg->setup_complete_ext.flags =
1265 			__cpu_to_le32(ATH10K_HTC_SETUP_COMPLETE_FLAGS_RX_BNDL_EN);
1266 		msg->setup_complete_ext.max_msgs_per_bundled_recv =
1267 			htc->max_msgs_per_htc_bundle;
1268 	}
1269 	ath10k_dbg(ar, ATH10K_DBG_HTC, "HTC is using TX credit flow control\n");
1270 
1271 	status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1272 	if (status) {
1273 		kfree_skb(skb);
1274 		return status;
1275 	}
1276 
1277 	if (ath10k_htc_pktlog_svc_supported(ar)) {
1278 		status = ath10k_htc_pktlog_connect(ar);
1279 		if (status) {
1280 			ath10k_err(ar, "failed to connect to pktlog: %d\n", status);
1281 			return status;
1282 		}
1283 	}
1284 
1285 	return 0;
1286 }
1287 
1288 /* registered target arrival callback from the HIF layer */
ath10k_htc_init(struct ath10k * ar)1289 int ath10k_htc_init(struct ath10k *ar)
1290 {
1291 	int status;
1292 	struct ath10k_htc *htc = &ar->htc;
1293 	struct ath10k_htc_svc_conn_req conn_req;
1294 	struct ath10k_htc_svc_conn_resp conn_resp;
1295 
1296 	spin_lock_init(&htc->tx_lock);
1297 
1298 	ath10k_htc_reset_endpoint_states(htc);
1299 
1300 	htc->ar = ar;
1301 
1302 	/* setup our pseudo HTC control endpoint connection */
1303 	memset(&conn_req, 0, sizeof(conn_req));
1304 	memset(&conn_resp, 0, sizeof(conn_resp));
1305 	conn_req.ep_ops.ep_tx_complete = ath10k_htc_control_tx_complete;
1306 	conn_req.ep_ops.ep_rx_complete = ath10k_htc_control_rx_complete;
1307 	conn_req.max_send_queue_depth = ATH10K_NUM_CONTROL_TX_BUFFERS;
1308 	conn_req.service_id = ATH10K_HTC_SVC_ID_RSVD_CTRL;
1309 
1310 	/* connect fake service */
1311 	status = ath10k_htc_connect_service(htc, &conn_req, &conn_resp);
1312 	if (status) {
1313 		ath10k_err(ar, "could not connect to htc service (%d)\n",
1314 			   status);
1315 		return status;
1316 	}
1317 
1318 	init_completion(&htc->ctl_resp);
1319 
1320 	return 0;
1321 }
1322