xref: /linux/net/nfc/nci/core.c (revision c5aec4c76af1a2d89ee2f2d4d5463b2ad2d85de5)
1 /*
2  *  The NFC Controller Interface is the communication protocol between an
3  *  NFC Controller (NFCC) and a Device Host (DH).
4  *
5  *  Copyright (C) 2011 Texas Instruments, Inc.
6  *
7  *  Written by Ilan Elias <ilane@ti.com>
8  *
9  *  Acknowledgements:
10  *  This file is based on hci_core.c, which was written
11  *  by Maxim Krasnyansky.
12  *
13  *  This program is free software; you can redistribute it and/or modify
14  *  it under the terms of the GNU General Public License version 2
15  *  as published by the Free Software Foundation
16  *
17  *  This program is distributed in the hope that it will be useful,
18  *  but WITHOUT ANY WARRANTY; without even the implied warranty of
19  *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
20  *  GNU General Public License for more details.
21  *
22  *  You should have received a copy of the GNU General Public License
23  *  along with this program; if not, see <http://www.gnu.org/licenses/>.
24  *
25  */
26 
27 #define pr_fmt(fmt) KBUILD_MODNAME ": %s: " fmt, __func__
28 
29 #include <linux/module.h>
30 #include <linux/types.h>
31 #include <linux/workqueue.h>
32 #include <linux/completion.h>
33 #include <linux/export.h>
34 #include <linux/sched.h>
35 #include <linux/bitops.h>
36 #include <linux/skbuff.h>
37 
38 #include "../nfc.h"
39 #include <net/nfc/nci.h>
40 #include <net/nfc/nci_core.h>
41 #include <linux/nfc.h>
42 
43 static void nci_cmd_work(struct work_struct *work);
44 static void nci_rx_work(struct work_struct *work);
45 static void nci_tx_work(struct work_struct *work);
46 
47 /* ---- NCI requests ---- */
48 
49 void nci_req_complete(struct nci_dev *ndev, int result)
50 {
51 	if (ndev->req_status == NCI_REQ_PEND) {
52 		ndev->req_result = result;
53 		ndev->req_status = NCI_REQ_DONE;
54 		complete(&ndev->req_completion);
55 	}
56 }
57 
58 static void nci_req_cancel(struct nci_dev *ndev, int err)
59 {
60 	if (ndev->req_status == NCI_REQ_PEND) {
61 		ndev->req_result = err;
62 		ndev->req_status = NCI_REQ_CANCELED;
63 		complete(&ndev->req_completion);
64 	}
65 }
66 
67 /* Execute request and wait for completion. */
68 static int __nci_request(struct nci_dev *ndev,
69 			 void (*req)(struct nci_dev *ndev, unsigned long opt),
70 			 unsigned long opt, __u32 timeout)
71 {
72 	int rc = 0;
73 	long completion_rc;
74 
75 	ndev->req_status = NCI_REQ_PEND;
76 
77 	reinit_completion(&ndev->req_completion);
78 	req(ndev, opt);
79 	completion_rc =
80 		wait_for_completion_interruptible_timeout(&ndev->req_completion,
81 							  timeout);
82 
83 	pr_debug("wait_for_completion return %ld\n", completion_rc);
84 
85 	if (completion_rc > 0) {
86 		switch (ndev->req_status) {
87 		case NCI_REQ_DONE:
88 			rc = nci_to_errno(ndev->req_result);
89 			break;
90 
91 		case NCI_REQ_CANCELED:
92 			rc = -ndev->req_result;
93 			break;
94 
95 		default:
96 			rc = -ETIMEDOUT;
97 			break;
98 		}
99 	} else {
100 		pr_err("wait_for_completion_interruptible_timeout failed %ld\n",
101 		       completion_rc);
102 
103 		rc = ((completion_rc == 0) ? (-ETIMEDOUT) : (completion_rc));
104 	}
105 
106 	ndev->req_status = ndev->req_result = 0;
107 
108 	return rc;
109 }
110 
111 static inline int nci_request(struct nci_dev *ndev,
112 			      void (*req)(struct nci_dev *ndev,
113 					  unsigned long opt),
114 			      unsigned long opt, __u32 timeout)
115 {
116 	int rc;
117 
118 	if (!test_bit(NCI_UP, &ndev->flags))
119 		return -ENETDOWN;
120 
121 	/* Serialize all requests */
122 	mutex_lock(&ndev->req_lock);
123 	rc = __nci_request(ndev, req, opt, timeout);
124 	mutex_unlock(&ndev->req_lock);
125 
126 	return rc;
127 }
128 
129 static void nci_reset_req(struct nci_dev *ndev, unsigned long opt)
130 {
131 	struct nci_core_reset_cmd cmd;
132 
133 	cmd.reset_type = NCI_RESET_TYPE_RESET_CONFIG;
134 	nci_send_cmd(ndev, NCI_OP_CORE_RESET_CMD, 1, &cmd);
135 }
136 
137 static void nci_init_req(struct nci_dev *ndev, unsigned long opt)
138 {
139 	nci_send_cmd(ndev, NCI_OP_CORE_INIT_CMD, 0, NULL);
140 }
141 
142 static void nci_init_complete_req(struct nci_dev *ndev, unsigned long opt)
143 {
144 	struct nci_rf_disc_map_cmd cmd;
145 	struct disc_map_config *cfg = cmd.mapping_configs;
146 	__u8 *num = &cmd.num_mapping_configs;
147 	int i;
148 
149 	/* set rf mapping configurations */
150 	*num = 0;
151 
152 	/* by default mapping is set to NCI_RF_INTERFACE_FRAME */
153 	for (i = 0; i < ndev->num_supported_rf_interfaces; i++) {
154 		if (ndev->supported_rf_interfaces[i] ==
155 		    NCI_RF_INTERFACE_ISO_DEP) {
156 			cfg[*num].rf_protocol = NCI_RF_PROTOCOL_ISO_DEP;
157 			cfg[*num].mode = NCI_DISC_MAP_MODE_POLL |
158 				NCI_DISC_MAP_MODE_LISTEN;
159 			cfg[*num].rf_interface = NCI_RF_INTERFACE_ISO_DEP;
160 			(*num)++;
161 		} else if (ndev->supported_rf_interfaces[i] ==
162 			   NCI_RF_INTERFACE_NFC_DEP) {
163 			cfg[*num].rf_protocol = NCI_RF_PROTOCOL_NFC_DEP;
164 			cfg[*num].mode = NCI_DISC_MAP_MODE_POLL |
165 				NCI_DISC_MAP_MODE_LISTEN;
166 			cfg[*num].rf_interface = NCI_RF_INTERFACE_NFC_DEP;
167 			(*num)++;
168 		}
169 
170 		if (*num == NCI_MAX_NUM_MAPPING_CONFIGS)
171 			break;
172 	}
173 
174 	nci_send_cmd(ndev, NCI_OP_RF_DISCOVER_MAP_CMD,
175 		     (1 + ((*num) * sizeof(struct disc_map_config))), &cmd);
176 }
177 
178 struct nci_set_config_param {
179 	__u8	id;
180 	size_t	len;
181 	__u8	*val;
182 };
183 
184 static void nci_set_config_req(struct nci_dev *ndev, unsigned long opt)
185 {
186 	struct nci_set_config_param *param = (struct nci_set_config_param *)opt;
187 	struct nci_core_set_config_cmd cmd;
188 
189 	BUG_ON(param->len > NCI_MAX_PARAM_LEN);
190 
191 	cmd.num_params = 1;
192 	cmd.param.id = param->id;
193 	cmd.param.len = param->len;
194 	memcpy(cmd.param.val, param->val, param->len);
195 
196 	nci_send_cmd(ndev, NCI_OP_CORE_SET_CONFIG_CMD, (3 + param->len), &cmd);
197 }
198 
199 static void nci_rf_discover_req(struct nci_dev *ndev, unsigned long opt)
200 {
201 	struct nci_rf_disc_cmd cmd;
202 	__u32 protocols = opt;
203 
204 	cmd.num_disc_configs = 0;
205 
206 	if ((cmd.num_disc_configs < NCI_MAX_NUM_RF_CONFIGS) &&
207 	    (protocols & NFC_PROTO_JEWEL_MASK ||
208 	     protocols & NFC_PROTO_MIFARE_MASK ||
209 	     protocols & NFC_PROTO_ISO14443_MASK ||
210 	     protocols & NFC_PROTO_NFC_DEP_MASK)) {
211 		cmd.disc_configs[cmd.num_disc_configs].rf_tech_and_mode =
212 			NCI_NFC_A_PASSIVE_POLL_MODE;
213 		cmd.disc_configs[cmd.num_disc_configs].frequency = 1;
214 		cmd.num_disc_configs++;
215 	}
216 
217 	if ((cmd.num_disc_configs < NCI_MAX_NUM_RF_CONFIGS) &&
218 	    (protocols & NFC_PROTO_ISO14443_B_MASK)) {
219 		cmd.disc_configs[cmd.num_disc_configs].rf_tech_and_mode =
220 			NCI_NFC_B_PASSIVE_POLL_MODE;
221 		cmd.disc_configs[cmd.num_disc_configs].frequency = 1;
222 		cmd.num_disc_configs++;
223 	}
224 
225 	if ((cmd.num_disc_configs < NCI_MAX_NUM_RF_CONFIGS) &&
226 	    (protocols & NFC_PROTO_FELICA_MASK ||
227 	     protocols & NFC_PROTO_NFC_DEP_MASK)) {
228 		cmd.disc_configs[cmd.num_disc_configs].rf_tech_and_mode =
229 			NCI_NFC_F_PASSIVE_POLL_MODE;
230 		cmd.disc_configs[cmd.num_disc_configs].frequency = 1;
231 		cmd.num_disc_configs++;
232 	}
233 
234 	nci_send_cmd(ndev, NCI_OP_RF_DISCOVER_CMD,
235 		     (1 + (cmd.num_disc_configs * sizeof(struct disc_config))),
236 		     &cmd);
237 }
238 
239 struct nci_rf_discover_select_param {
240 	__u8	rf_discovery_id;
241 	__u8	rf_protocol;
242 };
243 
244 static void nci_rf_discover_select_req(struct nci_dev *ndev, unsigned long opt)
245 {
246 	struct nci_rf_discover_select_param *param =
247 		(struct nci_rf_discover_select_param *)opt;
248 	struct nci_rf_discover_select_cmd cmd;
249 
250 	cmd.rf_discovery_id = param->rf_discovery_id;
251 	cmd.rf_protocol = param->rf_protocol;
252 
253 	switch (cmd.rf_protocol) {
254 	case NCI_RF_PROTOCOL_ISO_DEP:
255 		cmd.rf_interface = NCI_RF_INTERFACE_ISO_DEP;
256 		break;
257 
258 	case NCI_RF_PROTOCOL_NFC_DEP:
259 		cmd.rf_interface = NCI_RF_INTERFACE_NFC_DEP;
260 		break;
261 
262 	default:
263 		cmd.rf_interface = NCI_RF_INTERFACE_FRAME;
264 		break;
265 	}
266 
267 	nci_send_cmd(ndev, NCI_OP_RF_DISCOVER_SELECT_CMD,
268 		     sizeof(struct nci_rf_discover_select_cmd), &cmd);
269 }
270 
271 static void nci_rf_deactivate_req(struct nci_dev *ndev, unsigned long opt)
272 {
273 	struct nci_rf_deactivate_cmd cmd;
274 
275 	cmd.type = NCI_DEACTIVATE_TYPE_IDLE_MODE;
276 
277 	nci_send_cmd(ndev, NCI_OP_RF_DEACTIVATE_CMD,
278 		     sizeof(struct nci_rf_deactivate_cmd), &cmd);
279 }
280 
281 static int nci_open_device(struct nci_dev *ndev)
282 {
283 	int rc = 0;
284 
285 	mutex_lock(&ndev->req_lock);
286 
287 	if (test_bit(NCI_UP, &ndev->flags)) {
288 		rc = -EALREADY;
289 		goto done;
290 	}
291 
292 	if (ndev->ops->open(ndev)) {
293 		rc = -EIO;
294 		goto done;
295 	}
296 
297 	atomic_set(&ndev->cmd_cnt, 1);
298 
299 	set_bit(NCI_INIT, &ndev->flags);
300 
301 	rc = __nci_request(ndev, nci_reset_req, 0,
302 			   msecs_to_jiffies(NCI_RESET_TIMEOUT));
303 
304 	if (ndev->ops->setup)
305 		ndev->ops->setup(ndev);
306 
307 	if (!rc) {
308 		rc = __nci_request(ndev, nci_init_req, 0,
309 				   msecs_to_jiffies(NCI_INIT_TIMEOUT));
310 	}
311 
312 	if (!rc) {
313 		rc = __nci_request(ndev, nci_init_complete_req, 0,
314 				   msecs_to_jiffies(NCI_INIT_TIMEOUT));
315 	}
316 
317 	clear_bit(NCI_INIT, &ndev->flags);
318 
319 	if (!rc) {
320 		set_bit(NCI_UP, &ndev->flags);
321 		nci_clear_target_list(ndev);
322 		atomic_set(&ndev->state, NCI_IDLE);
323 	} else {
324 		/* Init failed, cleanup */
325 		skb_queue_purge(&ndev->cmd_q);
326 		skb_queue_purge(&ndev->rx_q);
327 		skb_queue_purge(&ndev->tx_q);
328 
329 		ndev->ops->close(ndev);
330 		ndev->flags = 0;
331 	}
332 
333 done:
334 	mutex_unlock(&ndev->req_lock);
335 	return rc;
336 }
337 
338 static int nci_close_device(struct nci_dev *ndev)
339 {
340 	nci_req_cancel(ndev, ENODEV);
341 	mutex_lock(&ndev->req_lock);
342 
343 	if (!test_and_clear_bit(NCI_UP, &ndev->flags)) {
344 		del_timer_sync(&ndev->cmd_timer);
345 		del_timer_sync(&ndev->data_timer);
346 		mutex_unlock(&ndev->req_lock);
347 		return 0;
348 	}
349 
350 	/* Drop RX and TX queues */
351 	skb_queue_purge(&ndev->rx_q);
352 	skb_queue_purge(&ndev->tx_q);
353 
354 	/* Flush RX and TX wq */
355 	flush_workqueue(ndev->rx_wq);
356 	flush_workqueue(ndev->tx_wq);
357 
358 	/* Reset device */
359 	skb_queue_purge(&ndev->cmd_q);
360 	atomic_set(&ndev->cmd_cnt, 1);
361 
362 	set_bit(NCI_INIT, &ndev->flags);
363 	__nci_request(ndev, nci_reset_req, 0,
364 		      msecs_to_jiffies(NCI_RESET_TIMEOUT));
365 	clear_bit(NCI_INIT, &ndev->flags);
366 
367 	del_timer_sync(&ndev->cmd_timer);
368 
369 	/* Flush cmd wq */
370 	flush_workqueue(ndev->cmd_wq);
371 
372 	/* After this point our queues are empty
373 	 * and no works are scheduled. */
374 	ndev->ops->close(ndev);
375 
376 	/* Clear flags */
377 	ndev->flags = 0;
378 
379 	mutex_unlock(&ndev->req_lock);
380 
381 	return 0;
382 }
383 
384 /* NCI command timer function */
385 static void nci_cmd_timer(unsigned long arg)
386 {
387 	struct nci_dev *ndev = (void *) arg;
388 
389 	atomic_set(&ndev->cmd_cnt, 1);
390 	queue_work(ndev->cmd_wq, &ndev->cmd_work);
391 }
392 
393 /* NCI data exchange timer function */
394 static void nci_data_timer(unsigned long arg)
395 {
396 	struct nci_dev *ndev = (void *) arg;
397 
398 	set_bit(NCI_DATA_EXCHANGE_TO, &ndev->flags);
399 	queue_work(ndev->rx_wq, &ndev->rx_work);
400 }
401 
402 static int nci_dev_up(struct nfc_dev *nfc_dev)
403 {
404 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
405 
406 	return nci_open_device(ndev);
407 }
408 
409 static int nci_dev_down(struct nfc_dev *nfc_dev)
410 {
411 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
412 
413 	return nci_close_device(ndev);
414 }
415 
416 int nci_set_config(struct nci_dev *ndev, __u8 id, size_t len, __u8 *val)
417 {
418 	struct nci_set_config_param param;
419 
420 	if (!val || !len)
421 		return 0;
422 
423 	param.id = id;
424 	param.len = len;
425 	param.val = val;
426 
427 	return __nci_request(ndev, nci_set_config_req, (unsigned long)&param,
428 			     msecs_to_jiffies(NCI_SET_CONFIG_TIMEOUT));
429 }
430 EXPORT_SYMBOL(nci_set_config);
431 
432 static int nci_set_local_general_bytes(struct nfc_dev *nfc_dev)
433 {
434 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
435 	struct nci_set_config_param param;
436 
437 	param.val = nfc_get_local_general_bytes(nfc_dev, &param.len);
438 	if ((param.val == NULL) || (param.len == 0))
439 		return 0;
440 
441 	if (param.len > NFC_MAX_GT_LEN)
442 		return -EINVAL;
443 
444 	param.id = NCI_PN_ATR_REQ_GEN_BYTES;
445 
446 	return nci_request(ndev, nci_set_config_req, (unsigned long)&param,
447 			   msecs_to_jiffies(NCI_SET_CONFIG_TIMEOUT));
448 }
449 
450 static int nci_start_poll(struct nfc_dev *nfc_dev,
451 			  __u32 im_protocols, __u32 tm_protocols)
452 {
453 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
454 	int rc;
455 
456 	if ((atomic_read(&ndev->state) == NCI_DISCOVERY) ||
457 	    (atomic_read(&ndev->state) == NCI_W4_ALL_DISCOVERIES)) {
458 		pr_err("unable to start poll, since poll is already active\n");
459 		return -EBUSY;
460 	}
461 
462 	if (ndev->target_active_prot) {
463 		pr_err("there is an active target\n");
464 		return -EBUSY;
465 	}
466 
467 	if ((atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) ||
468 	    (atomic_read(&ndev->state) == NCI_POLL_ACTIVE)) {
469 		pr_debug("target active or w4 select, implicitly deactivate\n");
470 
471 		rc = nci_request(ndev, nci_rf_deactivate_req, 0,
472 				 msecs_to_jiffies(NCI_RF_DEACTIVATE_TIMEOUT));
473 		if (rc)
474 			return -EBUSY;
475 	}
476 
477 	if (im_protocols & NFC_PROTO_NFC_DEP_MASK) {
478 		rc = nci_set_local_general_bytes(nfc_dev);
479 		if (rc) {
480 			pr_err("failed to set local general bytes\n");
481 			return rc;
482 		}
483 	}
484 
485 	rc = nci_request(ndev, nci_rf_discover_req, im_protocols,
486 			 msecs_to_jiffies(NCI_RF_DISC_TIMEOUT));
487 
488 	if (!rc)
489 		ndev->poll_prots = im_protocols;
490 
491 	return rc;
492 }
493 
494 static void nci_stop_poll(struct nfc_dev *nfc_dev)
495 {
496 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
497 
498 	if ((atomic_read(&ndev->state) != NCI_DISCOVERY) &&
499 	    (atomic_read(&ndev->state) != NCI_W4_ALL_DISCOVERIES)) {
500 		pr_err("unable to stop poll, since poll is not active\n");
501 		return;
502 	}
503 
504 	nci_request(ndev, nci_rf_deactivate_req, 0,
505 		    msecs_to_jiffies(NCI_RF_DEACTIVATE_TIMEOUT));
506 }
507 
508 static int nci_activate_target(struct nfc_dev *nfc_dev,
509 			       struct nfc_target *target, __u32 protocol)
510 {
511 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
512 	struct nci_rf_discover_select_param param;
513 	struct nfc_target *nci_target = NULL;
514 	int i;
515 	int rc = 0;
516 
517 	pr_debug("target_idx %d, protocol 0x%x\n", target->idx, protocol);
518 
519 	if ((atomic_read(&ndev->state) != NCI_W4_HOST_SELECT) &&
520 	    (atomic_read(&ndev->state) != NCI_POLL_ACTIVE)) {
521 		pr_err("there is no available target to activate\n");
522 		return -EINVAL;
523 	}
524 
525 	if (ndev->target_active_prot) {
526 		pr_err("there is already an active target\n");
527 		return -EBUSY;
528 	}
529 
530 	for (i = 0; i < ndev->n_targets; i++) {
531 		if (ndev->targets[i].idx == target->idx) {
532 			nci_target = &ndev->targets[i];
533 			break;
534 		}
535 	}
536 
537 	if (!nci_target) {
538 		pr_err("unable to find the selected target\n");
539 		return -EINVAL;
540 	}
541 
542 	if (!(nci_target->supported_protocols & (1 << protocol))) {
543 		pr_err("target does not support the requested protocol 0x%x\n",
544 		       protocol);
545 		return -EINVAL;
546 	}
547 
548 	if (atomic_read(&ndev->state) == NCI_W4_HOST_SELECT) {
549 		param.rf_discovery_id = nci_target->logical_idx;
550 
551 		if (protocol == NFC_PROTO_JEWEL)
552 			param.rf_protocol = NCI_RF_PROTOCOL_T1T;
553 		else if (protocol == NFC_PROTO_MIFARE)
554 			param.rf_protocol = NCI_RF_PROTOCOL_T2T;
555 		else if (protocol == NFC_PROTO_FELICA)
556 			param.rf_protocol = NCI_RF_PROTOCOL_T3T;
557 		else if (protocol == NFC_PROTO_ISO14443 ||
558 			 protocol == NFC_PROTO_ISO14443_B)
559 			param.rf_protocol = NCI_RF_PROTOCOL_ISO_DEP;
560 		else
561 			param.rf_protocol = NCI_RF_PROTOCOL_NFC_DEP;
562 
563 		rc = nci_request(ndev, nci_rf_discover_select_req,
564 				 (unsigned long)&param,
565 				 msecs_to_jiffies(NCI_RF_DISC_SELECT_TIMEOUT));
566 	}
567 
568 	if (!rc)
569 		ndev->target_active_prot = protocol;
570 
571 	return rc;
572 }
573 
574 static void nci_deactivate_target(struct nfc_dev *nfc_dev,
575 				  struct nfc_target *target)
576 {
577 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
578 
579 	pr_debug("entry\n");
580 
581 	if (!ndev->target_active_prot) {
582 		pr_err("unable to deactivate target, no active target\n");
583 		return;
584 	}
585 
586 	ndev->target_active_prot = 0;
587 
588 	if (atomic_read(&ndev->state) == NCI_POLL_ACTIVE) {
589 		nci_request(ndev, nci_rf_deactivate_req, 0,
590 			    msecs_to_jiffies(NCI_RF_DEACTIVATE_TIMEOUT));
591 	}
592 }
593 
594 static int nci_dep_link_up(struct nfc_dev *nfc_dev, struct nfc_target *target,
595 			   __u8 comm_mode, __u8 *gb, size_t gb_len)
596 {
597 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
598 	int rc;
599 
600 	pr_debug("target_idx %d, comm_mode %d\n", target->idx, comm_mode);
601 
602 	rc = nci_activate_target(nfc_dev, target, NFC_PROTO_NFC_DEP);
603 	if (rc)
604 		return rc;
605 
606 	rc = nfc_set_remote_general_bytes(nfc_dev, ndev->remote_gb,
607 					  ndev->remote_gb_len);
608 	if (!rc)
609 		rc = nfc_dep_link_is_up(nfc_dev, target->idx, NFC_COMM_PASSIVE,
610 					NFC_RF_INITIATOR);
611 
612 	return rc;
613 }
614 
615 static int nci_dep_link_down(struct nfc_dev *nfc_dev)
616 {
617 	pr_debug("entry\n");
618 
619 	nci_deactivate_target(nfc_dev, NULL);
620 
621 	return 0;
622 }
623 
624 
625 static int nci_transceive(struct nfc_dev *nfc_dev, struct nfc_target *target,
626 			  struct sk_buff *skb,
627 			  data_exchange_cb_t cb, void *cb_context)
628 {
629 	struct nci_dev *ndev = nfc_get_drvdata(nfc_dev);
630 	int rc;
631 
632 	pr_debug("target_idx %d, len %d\n", target->idx, skb->len);
633 
634 	if (!ndev->target_active_prot) {
635 		pr_err("unable to exchange data, no active target\n");
636 		return -EINVAL;
637 	}
638 
639 	if (test_and_set_bit(NCI_DATA_EXCHANGE, &ndev->flags))
640 		return -EBUSY;
641 
642 	/* store cb and context to be used on receiving data */
643 	ndev->data_exchange_cb = cb;
644 	ndev->data_exchange_cb_context = cb_context;
645 
646 	rc = nci_send_data(ndev, NCI_STATIC_RF_CONN_ID, skb);
647 	if (rc)
648 		clear_bit(NCI_DATA_EXCHANGE, &ndev->flags);
649 
650 	return rc;
651 }
652 
653 static int nci_enable_se(struct nfc_dev *nfc_dev, u32 se_idx)
654 {
655 	return 0;
656 }
657 
658 static int nci_disable_se(struct nfc_dev *nfc_dev, u32 se_idx)
659 {
660 	return 0;
661 }
662 
663 static int nci_discover_se(struct nfc_dev *nfc_dev)
664 {
665 	return 0;
666 }
667 
668 static struct nfc_ops nci_nfc_ops = {
669 	.dev_up = nci_dev_up,
670 	.dev_down = nci_dev_down,
671 	.start_poll = nci_start_poll,
672 	.stop_poll = nci_stop_poll,
673 	.dep_link_up = nci_dep_link_up,
674 	.dep_link_down = nci_dep_link_down,
675 	.activate_target = nci_activate_target,
676 	.deactivate_target = nci_deactivate_target,
677 	.im_transceive = nci_transceive,
678 	.enable_se = nci_enable_se,
679 	.disable_se = nci_disable_se,
680 	.discover_se = nci_discover_se,
681 };
682 
683 /* ---- Interface to NCI drivers ---- */
684 
685 /**
686  * nci_allocate_device - allocate a new nci device
687  *
688  * @ops: device operations
689  * @supported_protocols: NFC protocols supported by the device
690  */
691 struct nci_dev *nci_allocate_device(struct nci_ops *ops,
692 				    __u32 supported_protocols,
693 				    int tx_headroom, int tx_tailroom)
694 {
695 	struct nci_dev *ndev;
696 
697 	pr_debug("supported_protocols 0x%x\n", supported_protocols);
698 
699 	if (!ops->open || !ops->close || !ops->send)
700 		return NULL;
701 
702 	if (!supported_protocols)
703 		return NULL;
704 
705 	ndev = kzalloc(sizeof(struct nci_dev), GFP_KERNEL);
706 	if (!ndev)
707 		return NULL;
708 
709 	ndev->ops = ops;
710 	ndev->tx_headroom = tx_headroom;
711 	ndev->tx_tailroom = tx_tailroom;
712 	init_completion(&ndev->req_completion);
713 
714 	ndev->nfc_dev = nfc_allocate_device(&nci_nfc_ops,
715 					    supported_protocols,
716 					    tx_headroom + NCI_DATA_HDR_SIZE,
717 					    tx_tailroom);
718 	if (!ndev->nfc_dev)
719 		goto free_exit;
720 
721 	nfc_set_drvdata(ndev->nfc_dev, ndev);
722 
723 	return ndev;
724 
725 free_exit:
726 	kfree(ndev);
727 	return NULL;
728 }
729 EXPORT_SYMBOL(nci_allocate_device);
730 
731 /**
732  * nci_free_device - deallocate nci device
733  *
734  * @ndev: The nci device to deallocate
735  */
736 void nci_free_device(struct nci_dev *ndev)
737 {
738 	nfc_free_device(ndev->nfc_dev);
739 	kfree(ndev);
740 }
741 EXPORT_SYMBOL(nci_free_device);
742 
743 /**
744  * nci_register_device - register a nci device in the nfc subsystem
745  *
746  * @dev: The nci device to register
747  */
748 int nci_register_device(struct nci_dev *ndev)
749 {
750 	int rc;
751 	struct device *dev = &ndev->nfc_dev->dev;
752 	char name[32];
753 
754 	rc = nfc_register_device(ndev->nfc_dev);
755 	if (rc)
756 		goto exit;
757 
758 	ndev->flags = 0;
759 
760 	INIT_WORK(&ndev->cmd_work, nci_cmd_work);
761 	snprintf(name, sizeof(name), "%s_nci_cmd_wq", dev_name(dev));
762 	ndev->cmd_wq = create_singlethread_workqueue(name);
763 	if (!ndev->cmd_wq) {
764 		rc = -ENOMEM;
765 		goto unreg_exit;
766 	}
767 
768 	INIT_WORK(&ndev->rx_work, nci_rx_work);
769 	snprintf(name, sizeof(name), "%s_nci_rx_wq", dev_name(dev));
770 	ndev->rx_wq = create_singlethread_workqueue(name);
771 	if (!ndev->rx_wq) {
772 		rc = -ENOMEM;
773 		goto destroy_cmd_wq_exit;
774 	}
775 
776 	INIT_WORK(&ndev->tx_work, nci_tx_work);
777 	snprintf(name, sizeof(name), "%s_nci_tx_wq", dev_name(dev));
778 	ndev->tx_wq = create_singlethread_workqueue(name);
779 	if (!ndev->tx_wq) {
780 		rc = -ENOMEM;
781 		goto destroy_rx_wq_exit;
782 	}
783 
784 	skb_queue_head_init(&ndev->cmd_q);
785 	skb_queue_head_init(&ndev->rx_q);
786 	skb_queue_head_init(&ndev->tx_q);
787 
788 	setup_timer(&ndev->cmd_timer, nci_cmd_timer,
789 		    (unsigned long) ndev);
790 	setup_timer(&ndev->data_timer, nci_data_timer,
791 		    (unsigned long) ndev);
792 
793 	mutex_init(&ndev->req_lock);
794 
795 	goto exit;
796 
797 destroy_rx_wq_exit:
798 	destroy_workqueue(ndev->rx_wq);
799 
800 destroy_cmd_wq_exit:
801 	destroy_workqueue(ndev->cmd_wq);
802 
803 unreg_exit:
804 	nfc_unregister_device(ndev->nfc_dev);
805 
806 exit:
807 	return rc;
808 }
809 EXPORT_SYMBOL(nci_register_device);
810 
811 /**
812  * nci_unregister_device - unregister a nci device in the nfc subsystem
813  *
814  * @dev: The nci device to unregister
815  */
816 void nci_unregister_device(struct nci_dev *ndev)
817 {
818 	nci_close_device(ndev);
819 
820 	destroy_workqueue(ndev->cmd_wq);
821 	destroy_workqueue(ndev->rx_wq);
822 	destroy_workqueue(ndev->tx_wq);
823 
824 	nfc_unregister_device(ndev->nfc_dev);
825 }
826 EXPORT_SYMBOL(nci_unregister_device);
827 
828 /**
829  * nci_recv_frame - receive frame from NCI drivers
830  *
831  * @ndev: The nci device
832  * @skb: The sk_buff to receive
833  */
834 int nci_recv_frame(struct nci_dev *ndev, struct sk_buff *skb)
835 {
836 	pr_debug("len %d\n", skb->len);
837 
838 	if (!ndev || (!test_bit(NCI_UP, &ndev->flags) &&
839 	    !test_bit(NCI_INIT, &ndev->flags))) {
840 		kfree_skb(skb);
841 		return -ENXIO;
842 	}
843 
844 	/* Queue frame for rx worker thread */
845 	skb_queue_tail(&ndev->rx_q, skb);
846 	queue_work(ndev->rx_wq, &ndev->rx_work);
847 
848 	return 0;
849 }
850 EXPORT_SYMBOL(nci_recv_frame);
851 
852 static int nci_send_frame(struct nci_dev *ndev, struct sk_buff *skb)
853 {
854 	pr_debug("len %d\n", skb->len);
855 
856 	if (!ndev) {
857 		kfree_skb(skb);
858 		return -ENODEV;
859 	}
860 
861 	/* Get rid of skb owner, prior to sending to the driver. */
862 	skb_orphan(skb);
863 
864 	return ndev->ops->send(ndev, skb);
865 }
866 
867 /* Send NCI command */
868 int nci_send_cmd(struct nci_dev *ndev, __u16 opcode, __u8 plen, void *payload)
869 {
870 	struct nci_ctrl_hdr *hdr;
871 	struct sk_buff *skb;
872 
873 	pr_debug("opcode 0x%x, plen %d\n", opcode, plen);
874 
875 	skb = nci_skb_alloc(ndev, (NCI_CTRL_HDR_SIZE + plen), GFP_KERNEL);
876 	if (!skb) {
877 		pr_err("no memory for command\n");
878 		return -ENOMEM;
879 	}
880 
881 	hdr = (struct nci_ctrl_hdr *) skb_put(skb, NCI_CTRL_HDR_SIZE);
882 	hdr->gid = nci_opcode_gid(opcode);
883 	hdr->oid = nci_opcode_oid(opcode);
884 	hdr->plen = plen;
885 
886 	nci_mt_set((__u8 *)hdr, NCI_MT_CMD_PKT);
887 	nci_pbf_set((__u8 *)hdr, NCI_PBF_LAST);
888 
889 	if (plen)
890 		memcpy(skb_put(skb, plen), payload, plen);
891 
892 	skb_queue_tail(&ndev->cmd_q, skb);
893 	queue_work(ndev->cmd_wq, &ndev->cmd_work);
894 
895 	return 0;
896 }
897 
898 /* ---- NCI TX Data worker thread ---- */
899 
900 static void nci_tx_work(struct work_struct *work)
901 {
902 	struct nci_dev *ndev = container_of(work, struct nci_dev, tx_work);
903 	struct sk_buff *skb;
904 
905 	pr_debug("credits_cnt %d\n", atomic_read(&ndev->credits_cnt));
906 
907 	/* Send queued tx data */
908 	while (atomic_read(&ndev->credits_cnt)) {
909 		skb = skb_dequeue(&ndev->tx_q);
910 		if (!skb)
911 			return;
912 
913 		/* Check if data flow control is used */
914 		if (atomic_read(&ndev->credits_cnt) !=
915 		    NCI_DATA_FLOW_CONTROL_NOT_USED)
916 			atomic_dec(&ndev->credits_cnt);
917 
918 		pr_debug("NCI TX: MT=data, PBF=%d, conn_id=%d, plen=%d\n",
919 			 nci_pbf(skb->data),
920 			 nci_conn_id(skb->data),
921 			 nci_plen(skb->data));
922 
923 		nci_send_frame(ndev, skb);
924 
925 		mod_timer(&ndev->data_timer,
926 			  jiffies + msecs_to_jiffies(NCI_DATA_TIMEOUT));
927 	}
928 }
929 
930 /* ----- NCI RX worker thread (data & control) ----- */
931 
932 static void nci_rx_work(struct work_struct *work)
933 {
934 	struct nci_dev *ndev = container_of(work, struct nci_dev, rx_work);
935 	struct sk_buff *skb;
936 
937 	while ((skb = skb_dequeue(&ndev->rx_q))) {
938 		/* Process frame */
939 		switch (nci_mt(skb->data)) {
940 		case NCI_MT_RSP_PKT:
941 			nci_rsp_packet(ndev, skb);
942 			break;
943 
944 		case NCI_MT_NTF_PKT:
945 			nci_ntf_packet(ndev, skb);
946 			break;
947 
948 		case NCI_MT_DATA_PKT:
949 			nci_rx_data_packet(ndev, skb);
950 			break;
951 
952 		default:
953 			pr_err("unknown MT 0x%x\n", nci_mt(skb->data));
954 			kfree_skb(skb);
955 			break;
956 		}
957 	}
958 
959 	/* check if a data exchange timout has occurred */
960 	if (test_bit(NCI_DATA_EXCHANGE_TO, &ndev->flags)) {
961 		/* complete the data exchange transaction, if exists */
962 		if (test_bit(NCI_DATA_EXCHANGE, &ndev->flags))
963 			nci_data_exchange_complete(ndev, NULL, -ETIMEDOUT);
964 
965 		clear_bit(NCI_DATA_EXCHANGE_TO, &ndev->flags);
966 	}
967 }
968 
969 /* ----- NCI TX CMD worker thread ----- */
970 
971 static void nci_cmd_work(struct work_struct *work)
972 {
973 	struct nci_dev *ndev = container_of(work, struct nci_dev, cmd_work);
974 	struct sk_buff *skb;
975 
976 	pr_debug("cmd_cnt %d\n", atomic_read(&ndev->cmd_cnt));
977 
978 	/* Send queued command */
979 	if (atomic_read(&ndev->cmd_cnt)) {
980 		skb = skb_dequeue(&ndev->cmd_q);
981 		if (!skb)
982 			return;
983 
984 		atomic_dec(&ndev->cmd_cnt);
985 
986 		pr_debug("NCI TX: MT=cmd, PBF=%d, GID=0x%x, OID=0x%x, plen=%d\n",
987 			 nci_pbf(skb->data),
988 			 nci_opcode_gid(nci_opcode(skb->data)),
989 			 nci_opcode_oid(nci_opcode(skb->data)),
990 			 nci_plen(skb->data));
991 
992 		nci_send_frame(ndev, skb);
993 
994 		mod_timer(&ndev->cmd_timer,
995 			  jiffies + msecs_to_jiffies(NCI_CMD_TIMEOUT));
996 	}
997 }
998 
999 MODULE_LICENSE("GPL");
1000