xref: /linux/drivers/misc/ibmvmc.c (revision 5d2d4a9f603a47403395408f64b1261ca61f6d50)
1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3  * IBM Power Systems Virtual Management Channel Support.
4  *
5  * Copyright (c) 2004, 2018 IBM Corp.
6  *   Dave Engebretsen engebret@us.ibm.com
7  *   Steven Royer seroyer@linux.vnet.ibm.com
8  *   Adam Reznechek adreznec@linux.vnet.ibm.com
9  *   Bryant G. Ly <bryantly@linux.vnet.ibm.com>
10  */
11 
12 #include <linux/module.h>
13 #include <linux/kernel.h>
14 #include <linux/kthread.h>
15 #include <linux/major.h>
16 #include <linux/string.h>
17 #include <linux/fcntl.h>
18 #include <linux/slab.h>
19 #include <linux/poll.h>
20 #include <linux/init.h>
21 #include <linux/fs.h>
22 #include <linux/interrupt.h>
23 #include <linux/spinlock.h>
24 #include <linux/percpu.h>
25 #include <linux/delay.h>
26 #include <linux/uaccess.h>
27 #include <linux/io.h>
28 #include <linux/miscdevice.h>
29 #include <linux/sched/signal.h>
30 
31 #include <asm/byteorder.h>
32 #include <asm/irq.h>
33 #include <asm/vio.h>
34 
35 #include "ibmvmc.h"
36 
37 #define IBMVMC_DRIVER_VERSION "1.0"
38 
39 /*
40  * Static global variables
41  */
42 static DECLARE_WAIT_QUEUE_HEAD(ibmvmc_read_wait);
43 
44 static const char ibmvmc_driver_name[] = "ibmvmc";
45 
46 static struct ibmvmc_struct ibmvmc;
47 static struct ibmvmc_hmc hmcs[MAX_HMCS];
48 static struct crq_server_adapter ibmvmc_adapter;
49 
50 static int ibmvmc_max_buf_pool_size = DEFAULT_BUF_POOL_SIZE;
51 static int ibmvmc_max_hmcs = DEFAULT_HMCS;
52 static int ibmvmc_max_mtu = DEFAULT_MTU;
53 
54 static inline long h_copy_rdma(s64 length, u64 sliobn, u64 slioba,
55 			       u64 dliobn, u64 dlioba)
56 {
57 	long rc = 0;
58 
59 	/* Ensure all writes to source memory are visible before hcall */
60 	dma_wmb();
61 	pr_debug("ibmvmc: h_copy_rdma(0x%llx, 0x%llx, 0x%llx, 0x%llx, 0x%llx\n",
62 		 length, sliobn, slioba, dliobn, dlioba);
63 	rc = plpar_hcall_norets(H_COPY_RDMA, length, sliobn, slioba,
64 				dliobn, dlioba);
65 	pr_debug("ibmvmc: h_copy_rdma rc = 0x%lx\n", rc);
66 
67 	return rc;
68 }
69 
70 static inline void h_free_crq(uint32_t unit_address)
71 {
72 	long rc = 0;
73 
74 	do {
75 		if (H_IS_LONG_BUSY(rc))
76 			msleep(get_longbusy_msecs(rc));
77 
78 		rc = plpar_hcall_norets(H_FREE_CRQ, unit_address);
79 	} while ((rc == H_BUSY) || (H_IS_LONG_BUSY(rc)));
80 }
81 
82 /**
83  * h_request_vmc: - request a hypervisor virtual management channel device
84  * @vmc_index: drc index of the vmc device created
85  *
86  * Requests the hypervisor create a new virtual management channel device,
87  * allowing this partition to send hypervisor virtualization control
88  * commands.
89  *
90  * Return:
91  *	0 - Success
92  *	Non-zero - Failure
93  */
94 static inline long h_request_vmc(u32 *vmc_index)
95 {
96 	long rc = 0;
97 	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
98 
99 	do {
100 		if (H_IS_LONG_BUSY(rc))
101 			msleep(get_longbusy_msecs(rc));
102 
103 		/* Call to request the VMC device from phyp */
104 		rc = plpar_hcall(H_REQUEST_VMC, retbuf);
105 		pr_debug("ibmvmc: %s rc = 0x%lx\n", __func__, rc);
106 		*vmc_index = retbuf[0];
107 	} while ((rc == H_BUSY) || (H_IS_LONG_BUSY(rc)));
108 
109 	return rc;
110 }
111 
112 /* routines for managing a command/response queue */
113 /**
114  * ibmvmc_handle_event: - Interrupt handler for crq events
115  * @irq:        number of irq to handle, not used
116  * @dev_instance: crq_server_adapter that received interrupt
117  *
118  * Disables interrupts and schedules ibmvmc_task
119  *
120  * Always returns IRQ_HANDLED
121  */
122 static irqreturn_t ibmvmc_handle_event(int irq, void *dev_instance)
123 {
124 	struct crq_server_adapter *adapter =
125 		(struct crq_server_adapter *)dev_instance;
126 
127 	vio_disable_interrupts(to_vio_dev(adapter->dev));
128 	tasklet_schedule(&adapter->work_task);
129 
130 	return IRQ_HANDLED;
131 }
132 
133 /**
134  * ibmvmc_release_crq_queue - Release CRQ Queue
135  *
136  * @adapter:	crq_server_adapter struct
137  *
138  * Return:
139  *	0 - Success
140  *	Non-Zero - Failure
141  */
142 static void ibmvmc_release_crq_queue(struct crq_server_adapter *adapter)
143 {
144 	struct vio_dev *vdev = to_vio_dev(adapter->dev);
145 	struct crq_queue *queue = &adapter->queue;
146 
147 	free_irq(vdev->irq, (void *)adapter);
148 	tasklet_kill(&adapter->work_task);
149 
150 	if (adapter->reset_task)
151 		kthread_stop(adapter->reset_task);
152 
153 	h_free_crq(vdev->unit_address);
154 	dma_unmap_single(adapter->dev,
155 			 queue->msg_token,
156 			 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
157 	free_page((unsigned long)queue->msgs);
158 }
159 
160 /**
161  * ibmvmc_reset_crq_queue - Reset CRQ Queue
162  *
163  * @adapter:	crq_server_adapter struct
164  *
165  * This function calls h_free_crq and then calls H_REG_CRQ and does all the
166  * bookkeeping to get us back to where we can communicate.
167  *
168  * Return:
169  *	0 - Success
170  *	Non-Zero - Failure
171  */
172 static int ibmvmc_reset_crq_queue(struct crq_server_adapter *adapter)
173 {
174 	struct vio_dev *vdev = to_vio_dev(adapter->dev);
175 	struct crq_queue *queue = &adapter->queue;
176 	int rc = 0;
177 
178 	/* Close the CRQ */
179 	h_free_crq(vdev->unit_address);
180 
181 	/* Clean out the queue */
182 	memset(queue->msgs, 0x00, PAGE_SIZE);
183 	queue->cur = 0;
184 
185 	/* And re-open it again */
186 	rc = plpar_hcall_norets(H_REG_CRQ,
187 				vdev->unit_address,
188 				queue->msg_token, PAGE_SIZE);
189 	if (rc == 2)
190 		/* Adapter is good, but other end is not ready */
191 		dev_warn(adapter->dev, "Partner adapter not ready\n");
192 	else if (rc != 0)
193 		dev_err(adapter->dev, "couldn't register crq--rc 0x%x\n", rc);
194 
195 	return rc;
196 }
197 
198 /**
199  * crq_queue_next_crq: - Returns the next entry in message queue
200  * @queue:      crq_queue to use
201  *
202  * Returns pointer to next entry in queue, or NULL if there are no new
203  * entried in the CRQ.
204  */
205 static struct ibmvmc_crq_msg *crq_queue_next_crq(struct crq_queue *queue)
206 {
207 	struct ibmvmc_crq_msg *crq;
208 	unsigned long flags;
209 
210 	spin_lock_irqsave(&queue->lock, flags);
211 	crq = &queue->msgs[queue->cur];
212 	if (crq->valid & 0x80) {
213 		if (++queue->cur == queue->size)
214 			queue->cur = 0;
215 
216 		/* Ensure the read of the valid bit occurs before reading any
217 		 * other bits of the CRQ entry
218 		 */
219 		dma_rmb();
220 	} else {
221 		crq = NULL;
222 	}
223 
224 	spin_unlock_irqrestore(&queue->lock, flags);
225 
226 	return crq;
227 }
228 
229 /**
230  * ibmvmc_send_crq - Send CRQ
231  *
232  * @adapter:	crq_server_adapter struct
233  * @word1:	Word1 Data field
234  * @word2:	Word2 Data field
235  *
236  * Return:
237  *	0 - Success
238  *	Non-Zero - Failure
239  */
240 static long ibmvmc_send_crq(struct crq_server_adapter *adapter,
241 			    u64 word1, u64 word2)
242 {
243 	struct vio_dev *vdev = to_vio_dev(adapter->dev);
244 	long rc = 0;
245 
246 	dev_dbg(adapter->dev, "(0x%x, 0x%016llx, 0x%016llx)\n",
247 		vdev->unit_address, word1, word2);
248 
249 	/*
250 	 * Ensure the command buffer is flushed to memory before handing it
251 	 * over to the other side to prevent it from fetching any stale data.
252 	 */
253 	dma_wmb();
254 	rc = plpar_hcall_norets(H_SEND_CRQ, vdev->unit_address, word1, word2);
255 	dev_dbg(adapter->dev, "rc = 0x%lx\n", rc);
256 
257 	return rc;
258 }
259 
260 /**
261  * alloc_dma_buffer - Create DMA Buffer
262  *
263  * @vdev:	vio_dev struct
264  * @size:	Size field
265  * @dma_handle:	DMA address field
266  *
267  * Allocates memory for the command queue and maps remote memory into an
268  * ioba.
269  *
270  * Returns a pointer to the buffer
271  */
272 static void *alloc_dma_buffer(struct vio_dev *vdev, size_t size,
273 			      dma_addr_t *dma_handle)
274 {
275 	/* allocate memory */
276 	void *buffer = kzalloc(size, GFP_ATOMIC);
277 
278 	if (!buffer) {
279 		*dma_handle = 0;
280 		return NULL;
281 	}
282 
283 	/* DMA map */
284 	*dma_handle = dma_map_single(&vdev->dev, buffer, size,
285 				     DMA_BIDIRECTIONAL);
286 
287 	if (dma_mapping_error(&vdev->dev, *dma_handle)) {
288 		*dma_handle = 0;
289 		kfree_sensitive(buffer);
290 		return NULL;
291 	}
292 
293 	return buffer;
294 }
295 
296 /**
297  * free_dma_buffer - Free DMA Buffer
298  *
299  * @vdev:	vio_dev struct
300  * @size:	Size field
301  * @vaddr:	Address field
302  * @dma_handle:	DMA address field
303  *
304  * Releases memory for a command queue and unmaps mapped remote memory.
305  */
306 static void free_dma_buffer(struct vio_dev *vdev, size_t size, void *vaddr,
307 			    dma_addr_t dma_handle)
308 {
309 	/* DMA unmap */
310 	dma_unmap_single(&vdev->dev, dma_handle, size, DMA_BIDIRECTIONAL);
311 
312 	/* deallocate memory */
313 	kfree_sensitive(vaddr);
314 }
315 
316 /**
317  * ibmvmc_get_valid_hmc_buffer - Retrieve Valid HMC Buffer
318  *
319  * @hmc_index:	HMC Index Field
320  *
321  * Return:
322  *	Pointer to ibmvmc_buffer
323  */
324 static struct ibmvmc_buffer *ibmvmc_get_valid_hmc_buffer(u8 hmc_index)
325 {
326 	struct ibmvmc_buffer *buffer;
327 	struct ibmvmc_buffer *ret_buf = NULL;
328 	unsigned long i;
329 
330 	if (hmc_index > ibmvmc.max_hmc_index)
331 		return NULL;
332 
333 	buffer = hmcs[hmc_index].buffer;
334 
335 	for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
336 		if (buffer[i].valid && buffer[i].free &&
337 		    buffer[i].owner == VMC_BUF_OWNER_ALPHA) {
338 			buffer[i].free = 0;
339 			ret_buf = &buffer[i];
340 			break;
341 		}
342 	}
343 
344 	return ret_buf;
345 }
346 
347 /**
348  * ibmvmc_get_free_hmc_buffer - Get Free HMC Buffer
349  *
350  * @adapter:	crq_server_adapter struct
351  * @hmc_index:	Hmc Index field
352  *
353  * Return:
354  *	Pointer to ibmvmc_buffer
355  */
356 static struct ibmvmc_buffer *ibmvmc_get_free_hmc_buffer(struct crq_server_adapter *adapter,
357 							u8 hmc_index)
358 {
359 	struct ibmvmc_buffer *buffer;
360 	struct ibmvmc_buffer *ret_buf = NULL;
361 	unsigned long i;
362 
363 	if (hmc_index > ibmvmc.max_hmc_index) {
364 		dev_info(adapter->dev, "get_free_hmc_buffer: invalid hmc_index=0x%x\n",
365 			 hmc_index);
366 		return NULL;
367 	}
368 
369 	buffer = hmcs[hmc_index].buffer;
370 
371 	for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
372 		if (buffer[i].free &&
373 		    buffer[i].owner == VMC_BUF_OWNER_ALPHA) {
374 			buffer[i].free = 0;
375 			ret_buf = &buffer[i];
376 			break;
377 		}
378 	}
379 
380 	return ret_buf;
381 }
382 
383 /**
384  * ibmvmc_free_hmc_buffer - Free an HMC Buffer
385  *
386  * @hmc:	ibmvmc_hmc struct
387  * @buffer:	ibmvmc_buffer struct
388  *
389  */
390 static void ibmvmc_free_hmc_buffer(struct ibmvmc_hmc *hmc,
391 				   struct ibmvmc_buffer *buffer)
392 {
393 	unsigned long flags;
394 
395 	spin_lock_irqsave(&hmc->lock, flags);
396 	buffer->free = 1;
397 	spin_unlock_irqrestore(&hmc->lock, flags);
398 }
399 
400 /**
401  * ibmvmc_count_hmc_buffers - Count HMC Buffers
402  *
403  * @hmc_index:	HMC Index field
404  * @valid:	Valid number of buffers field
405  * @free:	Free number of buffers field
406  *
407  */
408 static void ibmvmc_count_hmc_buffers(u8 hmc_index, unsigned int *valid,
409 				     unsigned int *free)
410 {
411 	struct ibmvmc_buffer *buffer;
412 	unsigned long i;
413 	unsigned long flags;
414 
415 	if (hmc_index > ibmvmc.max_hmc_index)
416 		return;
417 
418 	if (!valid || !free)
419 		return;
420 
421 	*valid = 0; *free = 0;
422 
423 	buffer = hmcs[hmc_index].buffer;
424 	spin_lock_irqsave(&hmcs[hmc_index].lock, flags);
425 
426 	for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
427 		if (buffer[i].valid) {
428 			*valid = *valid + 1;
429 			if (buffer[i].free)
430 				*free = *free + 1;
431 		}
432 	}
433 
434 	spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
435 }
436 
437 /**
438  * ibmvmc_get_free_hmc - Get Free HMC
439  *
440  * Return:
441  *	Pointer to an available HMC Connection
442  *	Null otherwise
443  */
444 static struct ibmvmc_hmc *ibmvmc_get_free_hmc(void)
445 {
446 	unsigned long i;
447 	unsigned long flags;
448 
449 	/*
450 	 * Find an available HMC connection.
451 	 */
452 	for (i = 0; i <= ibmvmc.max_hmc_index; i++) {
453 		spin_lock_irqsave(&hmcs[i].lock, flags);
454 		if (hmcs[i].state == ibmhmc_state_free) {
455 			hmcs[i].index = i;
456 			hmcs[i].state = ibmhmc_state_initial;
457 			spin_unlock_irqrestore(&hmcs[i].lock, flags);
458 			return &hmcs[i];
459 		}
460 		spin_unlock_irqrestore(&hmcs[i].lock, flags);
461 	}
462 
463 	return NULL;
464 }
465 
466 /**
467  * ibmvmc_return_hmc - Return an HMC Connection
468  *
469  * @hmc:		ibmvmc_hmc struct
470  * @release_readers:	Number of readers connected to session
471  *
472  * This function releases the HMC connections back into the pool.
473  *
474  * Return:
475  *	0 - Success
476  *	Non-zero - Failure
477  */
478 static int ibmvmc_return_hmc(struct ibmvmc_hmc *hmc, bool release_readers)
479 {
480 	struct ibmvmc_buffer *buffer;
481 	struct crq_server_adapter *adapter;
482 	struct vio_dev *vdev;
483 	unsigned long i;
484 	unsigned long flags;
485 
486 	if (!hmc || !hmc->adapter)
487 		return -EIO;
488 
489 	if (release_readers) {
490 		if (hmc->file_session) {
491 			struct ibmvmc_file_session *session = hmc->file_session;
492 
493 			session->valid = 0;
494 			wake_up_interruptible(&ibmvmc_read_wait);
495 		}
496 	}
497 
498 	adapter = hmc->adapter;
499 	vdev = to_vio_dev(adapter->dev);
500 
501 	spin_lock_irqsave(&hmc->lock, flags);
502 	hmc->index = 0;
503 	hmc->state = ibmhmc_state_free;
504 	hmc->queue_head = 0;
505 	hmc->queue_tail = 0;
506 	buffer = hmc->buffer;
507 	for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
508 		if (buffer[i].valid) {
509 			free_dma_buffer(vdev,
510 					ibmvmc.max_mtu,
511 					buffer[i].real_addr_local,
512 					buffer[i].dma_addr_local);
513 			dev_dbg(adapter->dev, "Forgot buffer id 0x%lx\n", i);
514 		}
515 		memset(&buffer[i], 0, sizeof(struct ibmvmc_buffer));
516 
517 		hmc->queue_outbound_msgs[i] = VMC_INVALID_BUFFER_ID;
518 	}
519 
520 	spin_unlock_irqrestore(&hmc->lock, flags);
521 
522 	return 0;
523 }
524 
525 /**
526  * ibmvmc_send_open - Interface Open
527  * @buffer: Pointer to ibmvmc_buffer struct
528  * @hmc: Pointer to ibmvmc_hmc struct
529  *
530  * This command is sent by the management partition as the result of a
531  * management partition device request. It causes the hypervisor to
532  * prepare a set of data buffers for the management application connection
533  * indicated HMC idx. A unique HMC Idx would be used if multiple management
534  * applications running concurrently were desired. Before responding to this
535  * command, the hypervisor must provide the management partition with at
536  * least one of these new buffers via the Add Buffer. This indicates whether
537  * the messages are inbound or outbound from the hypervisor.
538  *
539  * Return:
540  *	0 - Success
541  *	Non-zero - Failure
542  */
543 static int ibmvmc_send_open(struct ibmvmc_buffer *buffer,
544 			    struct ibmvmc_hmc *hmc)
545 {
546 	struct ibmvmc_crq_msg crq_msg;
547 	struct crq_server_adapter *adapter;
548 	__be64 *crq_as_u64 = (__be64 *)&crq_msg;
549 	int rc = 0;
550 
551 	if (!hmc || !hmc->adapter)
552 		return -EIO;
553 
554 	adapter = hmc->adapter;
555 
556 	dev_dbg(adapter->dev, "send_open: 0x%lx 0x%lx 0x%lx 0x%lx 0x%lx\n",
557 		(unsigned long)buffer->size, (unsigned long)adapter->liobn,
558 		(unsigned long)buffer->dma_addr_local,
559 		(unsigned long)adapter->riobn,
560 		(unsigned long)buffer->dma_addr_remote);
561 
562 	rc = h_copy_rdma(buffer->size,
563 			 adapter->liobn,
564 			 buffer->dma_addr_local,
565 			 adapter->riobn,
566 			 buffer->dma_addr_remote);
567 	if (rc) {
568 		dev_err(adapter->dev, "Error: In send_open, h_copy_rdma rc 0x%x\n",
569 			rc);
570 		return -EIO;
571 	}
572 
573 	hmc->state = ibmhmc_state_opening;
574 
575 	crq_msg.valid = 0x80;
576 	crq_msg.type = VMC_MSG_OPEN;
577 	crq_msg.status = 0;
578 	crq_msg.var1.rsvd = 0;
579 	crq_msg.hmc_session = hmc->session;
580 	crq_msg.hmc_index = hmc->index;
581 	crq_msg.var2.buffer_id = cpu_to_be16(buffer->id);
582 	crq_msg.rsvd = 0;
583 	crq_msg.var3.rsvd = 0;
584 
585 	ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
586 			be64_to_cpu(crq_as_u64[1]));
587 
588 	return rc;
589 }
590 
591 /**
592  * ibmvmc_send_close - Interface Close
593  * @hmc: Pointer to ibmvmc_hmc struct
594  *
595  * This command is sent by the management partition to terminate a
596  * management application to hypervisor connection. When this command is
597  * sent, the management partition has quiesced all I/O operations to all
598  * buffers associated with this management application connection, and
599  * has freed any storage for these buffers.
600  *
601  * Return:
602  *	0 - Success
603  *	Non-zero - Failure
604  */
605 static int ibmvmc_send_close(struct ibmvmc_hmc *hmc)
606 {
607 	struct ibmvmc_crq_msg crq_msg;
608 	struct crq_server_adapter *adapter;
609 	__be64 *crq_as_u64 = (__be64 *)&crq_msg;
610 	int rc = 0;
611 
612 	if (!hmc || !hmc->adapter)
613 		return -EIO;
614 
615 	adapter = hmc->adapter;
616 
617 	dev_info(adapter->dev, "CRQ send: close\n");
618 
619 	crq_msg.valid = 0x80;
620 	crq_msg.type = VMC_MSG_CLOSE;
621 	crq_msg.status = 0;
622 	crq_msg.var1.rsvd = 0;
623 	crq_msg.hmc_session = hmc->session;
624 	crq_msg.hmc_index = hmc->index;
625 	crq_msg.var2.rsvd = 0;
626 	crq_msg.rsvd = 0;
627 	crq_msg.var3.rsvd = 0;
628 
629 	ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
630 			be64_to_cpu(crq_as_u64[1]));
631 
632 	return rc;
633 }
634 
635 /**
636  * ibmvmc_send_capabilities - Send VMC Capabilities
637  *
638  * @adapter:	crq_server_adapter struct
639  *
640  * The capabilities message is an administrative message sent after the CRQ
641  * initialization sequence of messages and is used to exchange VMC capabilities
642  * between the management partition and the hypervisor. The management
643  * partition must send this message and the hypervisor must respond with VMC
644  * capabilities Response message before HMC interface message can begin. Any
645  * HMC interface messages received before the exchange of capabilities has
646  * complete are dropped.
647  *
648  * Return:
649  *	0 - Success
650  */
651 static int ibmvmc_send_capabilities(struct crq_server_adapter *adapter)
652 {
653 	struct ibmvmc_admin_crq_msg crq_msg;
654 	__be64 *crq_as_u64 = (__be64 *)&crq_msg;
655 
656 	dev_dbg(adapter->dev, "ibmvmc: CRQ send: capabilities\n");
657 	crq_msg.valid = 0x80;
658 	crq_msg.type = VMC_MSG_CAP;
659 	crq_msg.status = 0;
660 	crq_msg.rsvd[0] = 0;
661 	crq_msg.rsvd[1] = 0;
662 	crq_msg.max_hmc = ibmvmc_max_hmcs;
663 	crq_msg.max_mtu = cpu_to_be32(ibmvmc_max_mtu);
664 	crq_msg.pool_size = cpu_to_be16(ibmvmc_max_buf_pool_size);
665 	crq_msg.crq_size = cpu_to_be16(adapter->queue.size);
666 	crq_msg.version = cpu_to_be16(IBMVMC_PROTOCOL_VERSION);
667 
668 	ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
669 			be64_to_cpu(crq_as_u64[1]));
670 
671 	ibmvmc.state = ibmvmc_state_capabilities;
672 
673 	return 0;
674 }
675 
676 /**
677  * ibmvmc_send_add_buffer_resp - Add Buffer Response
678  *
679  * @adapter:	crq_server_adapter struct
680  * @status:	Status field
681  * @hmc_session: HMC Session field
682  * @hmc_index:	HMC Index field
683  * @buffer_id:	Buffer Id field
684  *
685  * This command is sent by the management partition to the hypervisor in
686  * response to the Add Buffer message. The Status field indicates the result of
687  * the command.
688  *
689  * Return:
690  *	0 - Success
691  */
692 static int ibmvmc_send_add_buffer_resp(struct crq_server_adapter *adapter,
693 				       u8 status, u8 hmc_session,
694 				       u8 hmc_index, u16 buffer_id)
695 {
696 	struct ibmvmc_crq_msg crq_msg;
697 	__be64 *crq_as_u64 = (__be64 *)&crq_msg;
698 
699 	dev_dbg(adapter->dev, "CRQ send: add_buffer_resp\n");
700 	crq_msg.valid = 0x80;
701 	crq_msg.type = VMC_MSG_ADD_BUF_RESP;
702 	crq_msg.status = status;
703 	crq_msg.var1.rsvd = 0;
704 	crq_msg.hmc_session = hmc_session;
705 	crq_msg.hmc_index = hmc_index;
706 	crq_msg.var2.buffer_id = cpu_to_be16(buffer_id);
707 	crq_msg.rsvd = 0;
708 	crq_msg.var3.rsvd = 0;
709 
710 	ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
711 			be64_to_cpu(crq_as_u64[1]));
712 
713 	return 0;
714 }
715 
716 /**
717  * ibmvmc_send_rem_buffer_resp - Remove Buffer Response
718  *
719  * @adapter:	crq_server_adapter struct
720  * @status:	Status field
721  * @hmc_session: HMC Session field
722  * @hmc_index:	HMC Index field
723  * @buffer_id:	Buffer Id field
724  *
725  * This command is sent by the management partition to the hypervisor in
726  * response to the Remove Buffer message. The Buffer ID field indicates
727  * which buffer the management partition selected to remove. The Status
728  * field indicates the result of the command.
729  *
730  * Return:
731  *	0 - Success
732  */
733 static int ibmvmc_send_rem_buffer_resp(struct crq_server_adapter *adapter,
734 				       u8 status, u8 hmc_session,
735 				       u8 hmc_index, u16 buffer_id)
736 {
737 	struct ibmvmc_crq_msg crq_msg;
738 	__be64 *crq_as_u64 = (__be64 *)&crq_msg;
739 
740 	dev_dbg(adapter->dev, "CRQ send: rem_buffer_resp\n");
741 	crq_msg.valid = 0x80;
742 	crq_msg.type = VMC_MSG_REM_BUF_RESP;
743 	crq_msg.status = status;
744 	crq_msg.var1.rsvd = 0;
745 	crq_msg.hmc_session = hmc_session;
746 	crq_msg.hmc_index = hmc_index;
747 	crq_msg.var2.buffer_id = cpu_to_be16(buffer_id);
748 	crq_msg.rsvd = 0;
749 	crq_msg.var3.rsvd = 0;
750 
751 	ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
752 			be64_to_cpu(crq_as_u64[1]));
753 
754 	return 0;
755 }
756 
757 /**
758  * ibmvmc_send_msg - Signal Message
759  *
760  * @adapter:	crq_server_adapter struct
761  * @buffer:	ibmvmc_buffer struct
762  * @hmc:	ibmvmc_hmc struct
763  * @msg_len:	message length field
764  *
765  * This command is sent between the management partition and the hypervisor
766  * in order to signal the arrival of an HMC protocol message. The command
767  * can be sent by both the management partition and the hypervisor. It is
768  * used for all traffic between the management application and the hypervisor,
769  * regardless of who initiated the communication.
770  *
771  * There is no response to this message.
772  *
773  * Return:
774  *	0 - Success
775  *	Non-zero - Failure
776  */
777 static int ibmvmc_send_msg(struct crq_server_adapter *adapter,
778 			   struct ibmvmc_buffer *buffer,
779 			   struct ibmvmc_hmc *hmc, int msg_len)
780 {
781 	struct ibmvmc_crq_msg crq_msg;
782 	__be64 *crq_as_u64 = (__be64 *)&crq_msg;
783 	int rc = 0;
784 
785 	dev_dbg(adapter->dev, "CRQ send: rdma to HV\n");
786 	rc = h_copy_rdma(msg_len,
787 			 adapter->liobn,
788 			 buffer->dma_addr_local,
789 			 adapter->riobn,
790 			 buffer->dma_addr_remote);
791 	if (rc) {
792 		dev_err(adapter->dev, "Error in send_msg, h_copy_rdma rc 0x%x\n",
793 			rc);
794 		return rc;
795 	}
796 
797 	crq_msg.valid = 0x80;
798 	crq_msg.type = VMC_MSG_SIGNAL;
799 	crq_msg.status = 0;
800 	crq_msg.var1.rsvd = 0;
801 	crq_msg.hmc_session = hmc->session;
802 	crq_msg.hmc_index = hmc->index;
803 	crq_msg.var2.buffer_id = cpu_to_be16(buffer->id);
804 	crq_msg.var3.msg_len = cpu_to_be32(msg_len);
805 	dev_dbg(adapter->dev, "CRQ send: msg to HV 0x%llx 0x%llx\n",
806 		be64_to_cpu(crq_as_u64[0]), be64_to_cpu(crq_as_u64[1]));
807 
808 	buffer->owner = VMC_BUF_OWNER_HV;
809 	ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
810 			be64_to_cpu(crq_as_u64[1]));
811 
812 	return rc;
813 }
814 
815 /**
816  * ibmvmc_open - Open Session
817  *
818  * @inode:	inode struct
819  * @file:	file struct
820  *
821  * Return:
822  *	0 - Success
823  *	Non-zero - Failure
824  */
825 static int ibmvmc_open(struct inode *inode, struct file *file)
826 {
827 	struct ibmvmc_file_session *session;
828 
829 	pr_debug("%s: inode = 0x%lx, file = 0x%lx, state = 0x%x\n", __func__,
830 		 (unsigned long)inode, (unsigned long)file,
831 		 ibmvmc.state);
832 
833 	session = kzalloc(sizeof(*session), GFP_KERNEL);
834 	if (!session)
835 		return -ENOMEM;
836 
837 	session->file = file;
838 	file->private_data = session;
839 
840 	return 0;
841 }
842 
843 /**
844  * ibmvmc_close - Close Session
845  *
846  * @inode:	inode struct
847  * @file:	file struct
848  *
849  * Return:
850  *	0 - Success
851  *	Non-zero - Failure
852  */
853 static int ibmvmc_close(struct inode *inode, struct file *file)
854 {
855 	struct ibmvmc_file_session *session;
856 	struct ibmvmc_hmc *hmc;
857 	int rc = 0;
858 	unsigned long flags;
859 
860 	pr_debug("%s: file = 0x%lx, state = 0x%x\n", __func__,
861 		 (unsigned long)file, ibmvmc.state);
862 
863 	session = file->private_data;
864 	if (!session)
865 		return -EIO;
866 
867 	hmc = session->hmc;
868 	if (hmc) {
869 		if (!hmc->adapter)
870 			return -EIO;
871 
872 		if (ibmvmc.state == ibmvmc_state_failed) {
873 			dev_warn(hmc->adapter->dev, "close: state_failed\n");
874 			return -EIO;
875 		}
876 
877 		spin_lock_irqsave(&hmc->lock, flags);
878 		if (hmc->state >= ibmhmc_state_opening) {
879 			rc = ibmvmc_send_close(hmc);
880 			if (rc)
881 				dev_warn(hmc->adapter->dev, "close: send_close failed.\n");
882 		}
883 		spin_unlock_irqrestore(&hmc->lock, flags);
884 	}
885 
886 	kfree_sensitive(session);
887 
888 	return rc;
889 }
890 
891 /**
892  * ibmvmc_read - Read
893  *
894  * @file:	file struct
895  * @buf:	Character buffer
896  * @nbytes:	Size in bytes
897  * @ppos:	Offset
898  *
899  * Return:
900  *	0 - Success
901  *	Non-zero - Failure
902  */
903 static ssize_t ibmvmc_read(struct file *file, char *buf, size_t nbytes,
904 			   loff_t *ppos)
905 {
906 	struct ibmvmc_file_session *session;
907 	struct ibmvmc_hmc *hmc;
908 	struct crq_server_adapter *adapter;
909 	struct ibmvmc_buffer *buffer;
910 	ssize_t n;
911 	ssize_t retval = 0;
912 	unsigned long flags;
913 	DEFINE_WAIT(wait);
914 
915 	pr_debug("ibmvmc: read: file = 0x%lx, buf = 0x%lx, nbytes = 0x%lx\n",
916 		 (unsigned long)file, (unsigned long)buf,
917 		 (unsigned long)nbytes);
918 
919 	if (nbytes == 0)
920 		return 0;
921 
922 	if (nbytes > ibmvmc.max_mtu) {
923 		pr_warn("ibmvmc: read: nbytes invalid 0x%x\n",
924 			(unsigned int)nbytes);
925 		return -EINVAL;
926 	}
927 
928 	session = file->private_data;
929 	if (!session) {
930 		pr_warn("ibmvmc: read: no session\n");
931 		return -EIO;
932 	}
933 
934 	hmc = session->hmc;
935 	if (!hmc) {
936 		pr_warn("ibmvmc: read: no hmc\n");
937 		return -EIO;
938 	}
939 
940 	adapter = hmc->adapter;
941 	if (!adapter) {
942 		pr_warn("ibmvmc: read: no adapter\n");
943 		return -EIO;
944 	}
945 
946 	do {
947 		prepare_to_wait(&ibmvmc_read_wait, &wait, TASK_INTERRUPTIBLE);
948 
949 		spin_lock_irqsave(&hmc->lock, flags);
950 		if (hmc->queue_tail != hmc->queue_head)
951 			/* Data is available */
952 			break;
953 
954 		spin_unlock_irqrestore(&hmc->lock, flags);
955 
956 		if (!session->valid) {
957 			retval = -EBADFD;
958 			goto out;
959 		}
960 		if (file->f_flags & O_NONBLOCK) {
961 			retval = -EAGAIN;
962 			goto out;
963 		}
964 
965 		schedule();
966 
967 		if (signal_pending(current)) {
968 			retval = -ERESTARTSYS;
969 			goto out;
970 		}
971 	} while (1);
972 
973 	buffer = &(hmc->buffer[hmc->queue_outbound_msgs[hmc->queue_tail]]);
974 	hmc->queue_tail++;
975 	if (hmc->queue_tail == ibmvmc_max_buf_pool_size)
976 		hmc->queue_tail = 0;
977 	spin_unlock_irqrestore(&hmc->lock, flags);
978 
979 	nbytes = min_t(size_t, nbytes, buffer->msg_len);
980 	n = copy_to_user((void *)buf, buffer->real_addr_local, nbytes);
981 	dev_dbg(adapter->dev, "read: copy to user nbytes = 0x%lx.\n", nbytes);
982 	ibmvmc_free_hmc_buffer(hmc, buffer);
983 	retval = nbytes;
984 
985 	if (n) {
986 		dev_warn(adapter->dev, "read: copy to user failed.\n");
987 		retval = -EFAULT;
988 	}
989 
990  out:
991 	finish_wait(&ibmvmc_read_wait, &wait);
992 	dev_dbg(adapter->dev, "read: out %ld\n", retval);
993 	return retval;
994 }
995 
996 /**
997  * ibmvmc_poll - Poll
998  *
999  * @file:	file struct
1000  * @wait:	Poll Table
1001  *
1002  * Return:
1003  *	poll.h return values
1004  */
1005 static unsigned int ibmvmc_poll(struct file *file, poll_table *wait)
1006 {
1007 	struct ibmvmc_file_session *session;
1008 	struct ibmvmc_hmc *hmc;
1009 	unsigned int mask = 0;
1010 
1011 	session = file->private_data;
1012 	if (!session)
1013 		return 0;
1014 
1015 	hmc = session->hmc;
1016 	if (!hmc)
1017 		return 0;
1018 
1019 	poll_wait(file, &ibmvmc_read_wait, wait);
1020 
1021 	if (hmc->queue_head != hmc->queue_tail)
1022 		mask |= POLLIN | POLLRDNORM;
1023 
1024 	return mask;
1025 }
1026 
1027 /**
1028  * ibmvmc_write - Write
1029  *
1030  * @file:	file struct
1031  * @buffer:	Character buffer
1032  * @count:	Count field
1033  * @ppos:	Offset
1034  *
1035  * Return:
1036  *	0 - Success
1037  *	Non-zero - Failure
1038  */
1039 static ssize_t ibmvmc_write(struct file *file, const char *buffer,
1040 			    size_t count, loff_t *ppos)
1041 {
1042 	struct inode *inode;
1043 	struct ibmvmc_buffer *vmc_buffer;
1044 	struct ibmvmc_file_session *session;
1045 	struct crq_server_adapter *adapter;
1046 	struct ibmvmc_hmc *hmc;
1047 	unsigned char *buf;
1048 	unsigned long flags;
1049 	size_t bytes;
1050 	const char *p = buffer;
1051 	size_t c = count;
1052 	int ret = 0;
1053 
1054 	session = file->private_data;
1055 	if (!session)
1056 		return -EIO;
1057 
1058 	hmc = session->hmc;
1059 	if (!hmc)
1060 		return -EIO;
1061 
1062 	spin_lock_irqsave(&hmc->lock, flags);
1063 	if (hmc->state == ibmhmc_state_free) {
1064 		/* HMC connection is not valid (possibly was reset under us). */
1065 		ret = -EIO;
1066 		goto out;
1067 	}
1068 
1069 	adapter = hmc->adapter;
1070 	if (!adapter) {
1071 		ret = -EIO;
1072 		goto out;
1073 	}
1074 
1075 	if (count > ibmvmc.max_mtu) {
1076 		dev_warn(adapter->dev, "invalid buffer size 0x%lx\n",
1077 			 (unsigned long)count);
1078 		ret = -EIO;
1079 		goto out;
1080 	}
1081 
1082 	/* Waiting for the open resp message to the ioctl(1) - retry */
1083 	if (hmc->state == ibmhmc_state_opening) {
1084 		ret = -EBUSY;
1085 		goto out;
1086 	}
1087 
1088 	/* Make sure the ioctl() was called & the open msg sent, and that
1089 	 * the HMC connection has not failed.
1090 	 */
1091 	if (hmc->state != ibmhmc_state_ready) {
1092 		ret = -EIO;
1093 		goto out;
1094 	}
1095 
1096 	vmc_buffer = ibmvmc_get_valid_hmc_buffer(hmc->index);
1097 	if (!vmc_buffer) {
1098 		/* No buffer available for the msg send, or we have not yet
1099 		 * completed the open/open_resp sequence.  Retry until this is
1100 		 * complete.
1101 		 */
1102 		ret = -EBUSY;
1103 		goto out;
1104 	}
1105 	if (!vmc_buffer->real_addr_local) {
1106 		dev_err(adapter->dev, "no buffer storage assigned\n");
1107 		ret = -EIO;
1108 		goto out;
1109 	}
1110 	buf = vmc_buffer->real_addr_local;
1111 
1112 	while (c > 0) {
1113 		bytes = min_t(size_t, c, vmc_buffer->size);
1114 
1115 		bytes -= copy_from_user(buf, p, bytes);
1116 		if (!bytes) {
1117 			ret = -EFAULT;
1118 			goto out;
1119 		}
1120 		c -= bytes;
1121 		p += bytes;
1122 	}
1123 	if (p == buffer)
1124 		goto out;
1125 
1126 	inode = file_inode(file);
1127 	inode_set_mtime_to_ts(inode, inode_set_ctime_current(inode));
1128 	mark_inode_dirty(inode);
1129 
1130 	dev_dbg(adapter->dev, "write: file = 0x%lx, count = 0x%lx\n",
1131 		(unsigned long)file, (unsigned long)count);
1132 
1133 	ibmvmc_send_msg(adapter, vmc_buffer, hmc, count);
1134 	ret = p - buffer;
1135  out:
1136 	spin_unlock_irqrestore(&hmc->lock, flags);
1137 	return (ssize_t)(ret);
1138 }
1139 
1140 /**
1141  * ibmvmc_setup_hmc - Setup the HMC
1142  *
1143  * @session:	ibmvmc_file_session struct
1144  *
1145  * Return:
1146  *	0 - Success
1147  *	Non-zero - Failure
1148  */
1149 static long ibmvmc_setup_hmc(struct ibmvmc_file_session *session)
1150 {
1151 	struct ibmvmc_hmc *hmc;
1152 	unsigned int valid, free, index;
1153 
1154 	if (ibmvmc.state == ibmvmc_state_failed) {
1155 		pr_warn("ibmvmc: Reserve HMC: state_failed\n");
1156 		return -EIO;
1157 	}
1158 
1159 	if (ibmvmc.state < ibmvmc_state_ready) {
1160 		pr_warn("ibmvmc: Reserve HMC: not state_ready\n");
1161 		return -EAGAIN;
1162 	}
1163 
1164 	/* Device is busy until capabilities have been exchanged and we
1165 	 * have a generic buffer for each possible HMC connection.
1166 	 */
1167 	for (index = 0; index <= ibmvmc.max_hmc_index; index++) {
1168 		valid = 0;
1169 		ibmvmc_count_hmc_buffers(index, &valid, &free);
1170 		if (valid == 0) {
1171 			pr_warn("ibmvmc: buffers not ready for index %d\n",
1172 				index);
1173 			return -ENOBUFS;
1174 		}
1175 	}
1176 
1177 	/* Get an hmc object, and transition to ibmhmc_state_initial */
1178 	hmc = ibmvmc_get_free_hmc();
1179 	if (!hmc) {
1180 		pr_warn("%s: free hmc not found\n", __func__);
1181 		return -EBUSY;
1182 	}
1183 
1184 	hmc->session = hmc->session + 1;
1185 	if (hmc->session == 0xff)
1186 		hmc->session = 1;
1187 
1188 	session->hmc = hmc;
1189 	hmc->adapter = &ibmvmc_adapter;
1190 	hmc->file_session = session;
1191 	session->valid = 1;
1192 
1193 	return 0;
1194 }
1195 
1196 /**
1197  * ibmvmc_ioctl_sethmcid - IOCTL Set HMC ID
1198  *
1199  * @session:	ibmvmc_file_session struct
1200  * @new_hmc_id:	HMC id field
1201  *
1202  * IOCTL command to setup the hmc id
1203  *
1204  * Return:
1205  *	0 - Success
1206  *	Non-zero - Failure
1207  */
1208 static long ibmvmc_ioctl_sethmcid(struct ibmvmc_file_session *session,
1209 				  unsigned char __user *new_hmc_id)
1210 {
1211 	struct ibmvmc_hmc *hmc;
1212 	struct ibmvmc_buffer *buffer;
1213 	size_t bytes;
1214 	char print_buffer[HMC_ID_LEN + 1];
1215 	unsigned long flags;
1216 	long rc = 0;
1217 
1218 	/* Reserve HMC session */
1219 	hmc = session->hmc;
1220 	if (!hmc) {
1221 		rc = ibmvmc_setup_hmc(session);
1222 		if (rc)
1223 			return rc;
1224 
1225 		hmc = session->hmc;
1226 		if (!hmc) {
1227 			pr_err("ibmvmc: setup_hmc success but no hmc\n");
1228 			return -EIO;
1229 		}
1230 	}
1231 
1232 	if (hmc->state != ibmhmc_state_initial) {
1233 		pr_warn("ibmvmc: sethmcid: invalid state to send open 0x%x\n",
1234 			hmc->state);
1235 		return -EIO;
1236 	}
1237 
1238 	bytes = copy_from_user(hmc->hmc_id, new_hmc_id, HMC_ID_LEN);
1239 	if (bytes)
1240 		return -EFAULT;
1241 
1242 	/* Send Open Session command */
1243 	spin_lock_irqsave(&hmc->lock, flags);
1244 	buffer = ibmvmc_get_valid_hmc_buffer(hmc->index);
1245 	spin_unlock_irqrestore(&hmc->lock, flags);
1246 
1247 	if (!buffer || !buffer->real_addr_local) {
1248 		pr_warn("ibmvmc: sethmcid: no buffer available\n");
1249 		return -EIO;
1250 	}
1251 
1252 	strscpy(print_buffer, hmc->hmc_id, sizeof(print_buffer));
1253 	pr_info("ibmvmc: sethmcid: Set HMC ID: \"%s\"\n", print_buffer);
1254 
1255 	memcpy(buffer->real_addr_local, hmc->hmc_id, HMC_ID_LEN);
1256 	/* RDMA over ID, send open msg, change state to ibmhmc_state_opening */
1257 	rc = ibmvmc_send_open(buffer, hmc);
1258 
1259 	return rc;
1260 }
1261 
1262 /**
1263  * ibmvmc_ioctl_query - IOCTL Query
1264  *
1265  * @session:	ibmvmc_file_session struct
1266  * @ret_struct:	ibmvmc_query_struct
1267  *
1268  * Return:
1269  *	0 - Success
1270  *	Non-zero - Failure
1271  */
1272 static long ibmvmc_ioctl_query(struct ibmvmc_file_session *session,
1273 			       struct ibmvmc_query_struct __user *ret_struct)
1274 {
1275 	struct ibmvmc_query_struct query_struct;
1276 	size_t bytes;
1277 
1278 	memset(&query_struct, 0, sizeof(query_struct));
1279 	query_struct.have_vmc = (ibmvmc.state > ibmvmc_state_initial);
1280 	query_struct.state = ibmvmc.state;
1281 	query_struct.vmc_drc_index = ibmvmc.vmc_drc_index;
1282 
1283 	bytes = copy_to_user(ret_struct, &query_struct,
1284 			     sizeof(query_struct));
1285 	if (bytes)
1286 		return -EFAULT;
1287 
1288 	return 0;
1289 }
1290 
1291 /**
1292  * ibmvmc_ioctl_requestvmc - IOCTL Request VMC
1293  *
1294  * @session:	ibmvmc_file_session struct
1295  * @ret_vmc_index:	VMC Index
1296  *
1297  * Return:
1298  *	0 - Success
1299  *	Non-zero - Failure
1300  */
1301 static long ibmvmc_ioctl_requestvmc(struct ibmvmc_file_session *session,
1302 				    u32 __user *ret_vmc_index)
1303 {
1304 	/* TODO: (adreznec) Add locking to control multiple process access */
1305 	size_t bytes;
1306 	long rc;
1307 	u32 vmc_drc_index;
1308 
1309 	/* Call to request the VMC device from phyp*/
1310 	rc = h_request_vmc(&vmc_drc_index);
1311 	pr_debug("ibmvmc: requestvmc: H_REQUEST_VMC rc = 0x%lx\n", rc);
1312 
1313 	if (rc == H_SUCCESS) {
1314 		rc = 0;
1315 	} else if (rc == H_FUNCTION) {
1316 		pr_err("ibmvmc: requestvmc: h_request_vmc not supported\n");
1317 		return -EPERM;
1318 	} else if (rc == H_AUTHORITY) {
1319 		pr_err("ibmvmc: requestvmc: hypervisor denied vmc request\n");
1320 		return -EPERM;
1321 	} else if (rc == H_HARDWARE) {
1322 		pr_err("ibmvmc: requestvmc: hypervisor hardware fault\n");
1323 		return -EIO;
1324 	} else if (rc == H_RESOURCE) {
1325 		pr_err("ibmvmc: requestvmc: vmc resource unavailable\n");
1326 		return -ENODEV;
1327 	} else if (rc == H_NOT_AVAILABLE) {
1328 		pr_err("ibmvmc: requestvmc: system cannot be vmc managed\n");
1329 		return -EPERM;
1330 	} else if (rc == H_PARAMETER) {
1331 		pr_err("ibmvmc: requestvmc: invalid parameter\n");
1332 		return -EINVAL;
1333 	}
1334 
1335 	/* Success, set the vmc index in global struct */
1336 	ibmvmc.vmc_drc_index = vmc_drc_index;
1337 
1338 	bytes = copy_to_user(ret_vmc_index, &vmc_drc_index,
1339 			     sizeof(*ret_vmc_index));
1340 	if (bytes) {
1341 		pr_warn("ibmvmc: requestvmc: copy to user failed.\n");
1342 		return -EFAULT;
1343 	}
1344 	return rc;
1345 }
1346 
1347 /**
1348  * ibmvmc_ioctl - IOCTL
1349  *
1350  * @file:	file information
1351  * @cmd:	cmd field
1352  * @arg:	Argument field
1353  *
1354  * Return:
1355  *	0 - Success
1356  *	Non-zero - Failure
1357  */
1358 static long ibmvmc_ioctl(struct file *file,
1359 			 unsigned int cmd, unsigned long arg)
1360 {
1361 	struct ibmvmc_file_session *session = file->private_data;
1362 
1363 	pr_debug("ibmvmc: ioctl file=0x%lx, cmd=0x%x, arg=0x%lx, ses=0x%lx\n",
1364 		 (unsigned long)file, cmd, arg,
1365 		 (unsigned long)session);
1366 
1367 	if (!session) {
1368 		pr_warn("ibmvmc: ioctl: no session\n");
1369 		return -EIO;
1370 	}
1371 
1372 	switch (cmd) {
1373 	case VMC_IOCTL_SETHMCID:
1374 		return ibmvmc_ioctl_sethmcid(session,
1375 			(unsigned char __user *)arg);
1376 	case VMC_IOCTL_QUERY:
1377 		return ibmvmc_ioctl_query(session,
1378 			(struct ibmvmc_query_struct __user *)arg);
1379 	case VMC_IOCTL_REQUESTVMC:
1380 		return ibmvmc_ioctl_requestvmc(session,
1381 			(unsigned int __user *)arg);
1382 	default:
1383 		pr_warn("ibmvmc: unknown ioctl 0x%x\n", cmd);
1384 		return -EINVAL;
1385 	}
1386 }
1387 
1388 static const struct file_operations ibmvmc_fops = {
1389 	.owner		= THIS_MODULE,
1390 	.read		= ibmvmc_read,
1391 	.write		= ibmvmc_write,
1392 	.poll		= ibmvmc_poll,
1393 	.unlocked_ioctl	= ibmvmc_ioctl,
1394 	.open           = ibmvmc_open,
1395 	.release        = ibmvmc_close,
1396 };
1397 
1398 /**
1399  * ibmvmc_add_buffer - Add Buffer
1400  *
1401  * @adapter: crq_server_adapter struct
1402  * @crq:	ibmvmc_crq_msg struct
1403  *
1404  * This message transfers a buffer from hypervisor ownership to management
1405  * partition ownership. The LIOBA is obtained from the virtual TCE table
1406  * associated with the hypervisor side of the VMC device, and points to a
1407  * buffer of size MTU (as established in the capabilities exchange).
1408  *
1409  * Typical flow for ading buffers:
1410  * 1. A new management application connection is opened by the management
1411  *	partition.
1412  * 2. The hypervisor assigns new buffers for the traffic associated with
1413  *	that connection.
1414  * 3. The hypervisor sends VMC Add Buffer messages to the management
1415  *	partition, informing it of the new buffers.
1416  * 4. The hypervisor sends an HMC protocol message (to the management
1417  *	application) notifying it of the new buffers. This informs the
1418  *	application that it has buffers available for sending HMC
1419  *	commands.
1420  *
1421  * Return:
1422  *	0 - Success
1423  *	Non-zero - Failure
1424  */
1425 static int ibmvmc_add_buffer(struct crq_server_adapter *adapter,
1426 			     struct ibmvmc_crq_msg *crq)
1427 {
1428 	struct ibmvmc_buffer *buffer;
1429 	u8 hmc_index;
1430 	u8 hmc_session;
1431 	u16 buffer_id;
1432 	unsigned long flags;
1433 	int rc = 0;
1434 
1435 	if (!crq)
1436 		return -1;
1437 
1438 	hmc_session = crq->hmc_session;
1439 	hmc_index = crq->hmc_index;
1440 	buffer_id = be16_to_cpu(crq->var2.buffer_id);
1441 
1442 	if (hmc_index > ibmvmc.max_hmc_index) {
1443 		dev_err(adapter->dev, "add_buffer: invalid hmc_index = 0x%x\n",
1444 			hmc_index);
1445 		ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_HMC_INDEX,
1446 					    hmc_session, hmc_index, buffer_id);
1447 		return -1;
1448 	}
1449 
1450 	if (buffer_id >= ibmvmc.max_buffer_pool_size) {
1451 		dev_err(adapter->dev, "add_buffer: invalid buffer_id = 0x%x\n",
1452 			buffer_id);
1453 		ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_BUFFER_ID,
1454 					    hmc_session, hmc_index, buffer_id);
1455 		return -1;
1456 	}
1457 
1458 	spin_lock_irqsave(&hmcs[hmc_index].lock, flags);
1459 	buffer = &hmcs[hmc_index].buffer[buffer_id];
1460 
1461 	if (buffer->real_addr_local || buffer->dma_addr_local) {
1462 		dev_warn(adapter->dev, "add_buffer: already allocated id = 0x%lx\n",
1463 			 (unsigned long)buffer_id);
1464 		spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
1465 		ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_BUFFER_ID,
1466 					    hmc_session, hmc_index, buffer_id);
1467 		return -1;
1468 	}
1469 
1470 	buffer->real_addr_local = alloc_dma_buffer(to_vio_dev(adapter->dev),
1471 						   ibmvmc.max_mtu,
1472 						   &buffer->dma_addr_local);
1473 
1474 	if (!buffer->real_addr_local) {
1475 		dev_err(adapter->dev, "add_buffer: alloc_dma_buffer failed.\n");
1476 		spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
1477 		ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INTERFACE_FAILURE,
1478 					    hmc_session, hmc_index, buffer_id);
1479 		return -1;
1480 	}
1481 
1482 	buffer->dma_addr_remote = be32_to_cpu(crq->var3.lioba);
1483 	buffer->size = ibmvmc.max_mtu;
1484 	buffer->owner = crq->var1.owner;
1485 	buffer->free = 1;
1486 	/* Must ensure valid==1 is observable only after all other fields are */
1487 	dma_wmb();
1488 	buffer->valid = 1;
1489 	buffer->id = buffer_id;
1490 
1491 	dev_dbg(adapter->dev, "add_buffer: successfully added a buffer:\n");
1492 	dev_dbg(adapter->dev, "   index: %d, session: %d, buffer: 0x%x, owner: %d\n",
1493 		hmc_index, hmc_session, buffer_id, buffer->owner);
1494 	dev_dbg(adapter->dev, "   local: 0x%x, remote: 0x%x\n",
1495 		(u32)buffer->dma_addr_local,
1496 		(u32)buffer->dma_addr_remote);
1497 	spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
1498 
1499 	ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_SUCCESS, hmc_session,
1500 				    hmc_index, buffer_id);
1501 
1502 	return rc;
1503 }
1504 
1505 /**
1506  * ibmvmc_rem_buffer - Remove Buffer
1507  *
1508  * @adapter: crq_server_adapter struct
1509  * @crq:	ibmvmc_crq_msg struct
1510  *
1511  * This message requests an HMC buffer to be transferred from management
1512  * partition ownership to hypervisor ownership. The management partition may
1513  * not be able to satisfy the request at a particular point in time if all its
1514  * buffers are in use. The management partition requires a depth of at least
1515  * one inbound buffer to allow management application commands to flow to the
1516  * hypervisor. It is, therefore, an interface error for the hypervisor to
1517  * attempt to remove the management partition's last buffer.
1518  *
1519  * The hypervisor is expected to manage buffer usage with the management
1520  * application directly and inform the management partition when buffers may be
1521  * removed. The typical flow for removing buffers:
1522  *
1523  * 1. The management application no longer needs a communication path to a
1524  *	particular hypervisor function. That function is closed.
1525  * 2. The hypervisor and the management application quiesce all traffic to that
1526  *	function. The hypervisor requests a reduction in buffer pool size.
1527  * 3. The management application acknowledges the reduction in buffer pool size.
1528  * 4. The hypervisor sends a Remove Buffer message to the management partition,
1529  *	informing it of the reduction in buffers.
1530  * 5. The management partition verifies it can remove the buffer. This is
1531  *	possible if buffers have been quiesced.
1532  *
1533  * Return:
1534  *	0 - Success
1535  *	Non-zero - Failure
1536  */
1537 /*
1538  * The hypervisor requested that we pick an unused buffer, and return it.
1539  * Before sending the buffer back, we free any storage associated with the
1540  * buffer.
1541  */
1542 static int ibmvmc_rem_buffer(struct crq_server_adapter *adapter,
1543 			     struct ibmvmc_crq_msg *crq)
1544 {
1545 	struct ibmvmc_buffer *buffer;
1546 	u8 hmc_index;
1547 	u8 hmc_session;
1548 	u16 buffer_id = 0;
1549 	unsigned long flags;
1550 	int rc = 0;
1551 
1552 	if (!crq)
1553 		return -1;
1554 
1555 	hmc_session = crq->hmc_session;
1556 	hmc_index = crq->hmc_index;
1557 
1558 	if (hmc_index > ibmvmc.max_hmc_index) {
1559 		dev_warn(adapter->dev, "rem_buffer: invalid hmc_index = 0x%x\n",
1560 			 hmc_index);
1561 		ibmvmc_send_rem_buffer_resp(adapter, VMC_MSG_INVALID_HMC_INDEX,
1562 					    hmc_session, hmc_index, buffer_id);
1563 		return -1;
1564 	}
1565 
1566 	spin_lock_irqsave(&hmcs[hmc_index].lock, flags);
1567 	buffer = ibmvmc_get_free_hmc_buffer(adapter, hmc_index);
1568 	if (!buffer) {
1569 		dev_info(adapter->dev, "rem_buffer: no buffer to remove\n");
1570 		spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
1571 		ibmvmc_send_rem_buffer_resp(adapter, VMC_MSG_NO_BUFFER,
1572 					    hmc_session, hmc_index,
1573 					    VMC_INVALID_BUFFER_ID);
1574 		return -1;
1575 	}
1576 
1577 	buffer_id = buffer->id;
1578 
1579 	if (buffer->valid)
1580 		free_dma_buffer(to_vio_dev(adapter->dev),
1581 				ibmvmc.max_mtu,
1582 				buffer->real_addr_local,
1583 				buffer->dma_addr_local);
1584 
1585 	memset(buffer, 0, sizeof(struct ibmvmc_buffer));
1586 	spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
1587 
1588 	dev_dbg(adapter->dev, "rem_buffer: removed buffer 0x%x.\n", buffer_id);
1589 	ibmvmc_send_rem_buffer_resp(adapter, VMC_MSG_SUCCESS, hmc_session,
1590 				    hmc_index, buffer_id);
1591 
1592 	return rc;
1593 }
1594 
1595 static int ibmvmc_recv_msg(struct crq_server_adapter *adapter,
1596 			   struct ibmvmc_crq_msg *crq)
1597 {
1598 	struct ibmvmc_buffer *buffer;
1599 	struct ibmvmc_hmc *hmc;
1600 	unsigned long msg_len;
1601 	u8 hmc_index;
1602 	u8 hmc_session;
1603 	u16 buffer_id;
1604 	unsigned long flags;
1605 	int rc = 0;
1606 
1607 	if (!crq)
1608 		return -1;
1609 
1610 	/* Hypervisor writes CRQs directly into our memory in big endian */
1611 	dev_dbg(adapter->dev, "Recv_msg: msg from HV 0x%016llx 0x%016llx\n",
1612 		be64_to_cpu(*((unsigned long *)crq)),
1613 		be64_to_cpu(*(((unsigned long *)crq) + 1)));
1614 
1615 	hmc_session = crq->hmc_session;
1616 	hmc_index = crq->hmc_index;
1617 	buffer_id = be16_to_cpu(crq->var2.buffer_id);
1618 	msg_len = be32_to_cpu(crq->var3.msg_len);
1619 
1620 	if (hmc_index > ibmvmc.max_hmc_index) {
1621 		dev_err(adapter->dev, "Recv_msg: invalid hmc_index = 0x%x\n",
1622 			hmc_index);
1623 		ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_HMC_INDEX,
1624 					    hmc_session, hmc_index, buffer_id);
1625 		return -1;
1626 	}
1627 
1628 	if (buffer_id >= ibmvmc.max_buffer_pool_size) {
1629 		dev_err(adapter->dev, "Recv_msg: invalid buffer_id = 0x%x\n",
1630 			buffer_id);
1631 		ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_BUFFER_ID,
1632 					    hmc_session, hmc_index, buffer_id);
1633 		return -1;
1634 	}
1635 
1636 	hmc = &hmcs[hmc_index];
1637 	spin_lock_irqsave(&hmc->lock, flags);
1638 
1639 	if (hmc->state == ibmhmc_state_free) {
1640 		dev_err(adapter->dev, "Recv_msg: invalid hmc state = 0x%x\n",
1641 			hmc->state);
1642 		/* HMC connection is not valid (possibly was reset under us). */
1643 		spin_unlock_irqrestore(&hmc->lock, flags);
1644 		return -1;
1645 	}
1646 
1647 	buffer = &hmc->buffer[buffer_id];
1648 
1649 	if (buffer->valid == 0 || buffer->owner == VMC_BUF_OWNER_ALPHA) {
1650 		dev_err(adapter->dev, "Recv_msg: not valid, or not HV.  0x%x 0x%x\n",
1651 			buffer->valid, buffer->owner);
1652 		spin_unlock_irqrestore(&hmc->lock, flags);
1653 		return -1;
1654 	}
1655 
1656 	/* RDMA the data into the partition. */
1657 	rc = h_copy_rdma(msg_len,
1658 			 adapter->riobn,
1659 			 buffer->dma_addr_remote,
1660 			 adapter->liobn,
1661 			 buffer->dma_addr_local);
1662 
1663 	dev_dbg(adapter->dev, "Recv_msg: msg_len = 0x%x, buffer_id = 0x%x, queue_head = 0x%x, hmc_idx = 0x%x\n",
1664 		(unsigned int)msg_len, (unsigned int)buffer_id,
1665 		(unsigned int)hmc->queue_head, (unsigned int)hmc_index);
1666 	buffer->msg_len = msg_len;
1667 	buffer->free = 0;
1668 	buffer->owner = VMC_BUF_OWNER_ALPHA;
1669 
1670 	if (rc) {
1671 		dev_err(adapter->dev, "Failure in recv_msg: h_copy_rdma = 0x%x\n",
1672 			rc);
1673 		spin_unlock_irqrestore(&hmc->lock, flags);
1674 		return -1;
1675 	}
1676 
1677 	/* Must be locked because read operates on the same data */
1678 	hmc->queue_outbound_msgs[hmc->queue_head] = buffer_id;
1679 	hmc->queue_head++;
1680 	if (hmc->queue_head == ibmvmc_max_buf_pool_size)
1681 		hmc->queue_head = 0;
1682 
1683 	if (hmc->queue_head == hmc->queue_tail)
1684 		dev_err(adapter->dev, "outbound buffer queue wrapped.\n");
1685 
1686 	spin_unlock_irqrestore(&hmc->lock, flags);
1687 
1688 	wake_up_interruptible(&ibmvmc_read_wait);
1689 
1690 	return 0;
1691 }
1692 
1693 /**
1694  * ibmvmc_process_capabilities - Process Capabilities
1695  *
1696  * @adapter:	crq_server_adapter struct
1697  * @crqp:	ibmvmc_crq_msg struct
1698  *
1699  */
1700 static void ibmvmc_process_capabilities(struct crq_server_adapter *adapter,
1701 					struct ibmvmc_crq_msg *crqp)
1702 {
1703 	struct ibmvmc_admin_crq_msg *crq = (struct ibmvmc_admin_crq_msg *)crqp;
1704 
1705 	if ((be16_to_cpu(crq->version) >> 8) !=
1706 			(IBMVMC_PROTOCOL_VERSION >> 8)) {
1707 		dev_err(adapter->dev, "init failed, incompatible versions 0x%x 0x%x\n",
1708 			be16_to_cpu(crq->version),
1709 			IBMVMC_PROTOCOL_VERSION);
1710 		ibmvmc.state = ibmvmc_state_failed;
1711 		return;
1712 	}
1713 
1714 	ibmvmc.max_mtu = min_t(u32, ibmvmc_max_mtu, be32_to_cpu(crq->max_mtu));
1715 	ibmvmc.max_buffer_pool_size = min_t(u16, ibmvmc_max_buf_pool_size,
1716 					    be16_to_cpu(crq->pool_size));
1717 	ibmvmc.max_hmc_index = min_t(u8, ibmvmc_max_hmcs, crq->max_hmc) - 1;
1718 	ibmvmc.state = ibmvmc_state_ready;
1719 
1720 	dev_info(adapter->dev, "Capabilities: mtu=0x%x, pool_size=0x%x, max_hmc=0x%x\n",
1721 		 ibmvmc.max_mtu, ibmvmc.max_buffer_pool_size,
1722 		 ibmvmc.max_hmc_index);
1723 }
1724 
1725 /**
1726  * ibmvmc_validate_hmc_session - Validate HMC Session
1727  *
1728  * @adapter:	crq_server_adapter struct
1729  * @crq:	ibmvmc_crq_msg struct
1730  *
1731  * Return:
1732  *	0 - Success
1733  *	Non-zero - Failure
1734  */
1735 static int ibmvmc_validate_hmc_session(struct crq_server_adapter *adapter,
1736 				       struct ibmvmc_crq_msg *crq)
1737 {
1738 	unsigned char hmc_index;
1739 
1740 	hmc_index = crq->hmc_index;
1741 
1742 	if (crq->hmc_session == 0)
1743 		return 0;
1744 
1745 	if (hmc_index > ibmvmc.max_hmc_index)
1746 		return -1;
1747 
1748 	if (hmcs[hmc_index].session != crq->hmc_session) {
1749 		dev_warn(adapter->dev, "Drop, bad session: expected 0x%x, recv 0x%x\n",
1750 			 hmcs[hmc_index].session, crq->hmc_session);
1751 		return -1;
1752 	}
1753 
1754 	return 0;
1755 }
1756 
1757 /**
1758  * ibmvmc_reset - Reset
1759  *
1760  * @adapter:	crq_server_adapter struct
1761  * @xport_event:	export_event field
1762  *
1763  * Closes all HMC sessions and conditionally schedules a CRQ reset.
1764  * @xport_event: If true, the partner closed their CRQ; we don't need to reset.
1765  *               If false, we need to schedule a CRQ reset.
1766  */
1767 static void ibmvmc_reset(struct crq_server_adapter *adapter, bool xport_event)
1768 {
1769 	int i;
1770 
1771 	if (ibmvmc.state != ibmvmc_state_sched_reset) {
1772 		dev_info(adapter->dev, "*** Reset to initial state.\n");
1773 		for (i = 0; i < ibmvmc_max_hmcs; i++)
1774 			ibmvmc_return_hmc(&hmcs[i], xport_event);
1775 
1776 		if (xport_event) {
1777 			/* CRQ was closed by the partner.  We don't need to do
1778 			 * anything except set ourself to the correct state to
1779 			 * handle init msgs.
1780 			 */
1781 			ibmvmc.state = ibmvmc_state_crqinit;
1782 		} else {
1783 			/* The partner did not close their CRQ - instead, we're
1784 			 * closing the CRQ on our end. Need to schedule this
1785 			 * for process context, because CRQ reset may require a
1786 			 * sleep.
1787 			 *
1788 			 * Setting ibmvmc.state here immediately prevents
1789 			 * ibmvmc_open from completing until the reset
1790 			 * completes in process context.
1791 			 */
1792 			ibmvmc.state = ibmvmc_state_sched_reset;
1793 			dev_dbg(adapter->dev, "Device reset scheduled");
1794 			wake_up_interruptible(&adapter->reset_wait_queue);
1795 		}
1796 	}
1797 }
1798 
1799 /**
1800  * ibmvmc_reset_task - Reset Task
1801  *
1802  * @data:	Data field
1803  *
1804  * Performs a CRQ reset of the VMC device in process context.
1805  * NOTE: This function should not be called directly, use ibmvmc_reset.
1806  */
1807 static int ibmvmc_reset_task(void *data)
1808 {
1809 	struct crq_server_adapter *adapter = data;
1810 	int rc;
1811 
1812 	set_user_nice(current, -20);
1813 
1814 	while (!kthread_should_stop()) {
1815 		wait_event_interruptible(adapter->reset_wait_queue,
1816 			(ibmvmc.state == ibmvmc_state_sched_reset) ||
1817 			kthread_should_stop());
1818 
1819 		if (kthread_should_stop())
1820 			break;
1821 
1822 		dev_dbg(adapter->dev, "CRQ resetting in process context");
1823 		tasklet_disable(&adapter->work_task);
1824 
1825 		rc = ibmvmc_reset_crq_queue(adapter);
1826 
1827 		if (rc != H_SUCCESS && rc != H_RESOURCE) {
1828 			dev_err(adapter->dev, "Error initializing CRQ.  rc = 0x%x\n",
1829 				rc);
1830 			ibmvmc.state = ibmvmc_state_failed;
1831 		} else {
1832 			ibmvmc.state = ibmvmc_state_crqinit;
1833 
1834 			if (ibmvmc_send_crq(adapter, 0xC001000000000000LL, 0)
1835 			    != 0 && rc != H_RESOURCE)
1836 				dev_warn(adapter->dev, "Failed to send initialize CRQ message\n");
1837 		}
1838 
1839 		vio_enable_interrupts(to_vio_dev(adapter->dev));
1840 		tasklet_enable(&adapter->work_task);
1841 	}
1842 
1843 	return 0;
1844 }
1845 
1846 /**
1847  * ibmvmc_process_open_resp - Process Open Response
1848  *
1849  * @crq: ibmvmc_crq_msg struct
1850  * @adapter:    crq_server_adapter struct
1851  *
1852  * This command is sent by the hypervisor in response to the Interface
1853  * Open message. When this message is received, the indicated buffer is
1854  * again available for management partition use.
1855  */
1856 static void ibmvmc_process_open_resp(struct ibmvmc_crq_msg *crq,
1857 				     struct crq_server_adapter *adapter)
1858 {
1859 	unsigned char hmc_index;
1860 	unsigned short buffer_id;
1861 
1862 	hmc_index = crq->hmc_index;
1863 	if (hmc_index > ibmvmc.max_hmc_index) {
1864 		/* Why would PHYP give an index > max negotiated? */
1865 		ibmvmc_reset(adapter, false);
1866 		return;
1867 	}
1868 
1869 	if (crq->status) {
1870 		dev_warn(adapter->dev, "open_resp: failed - status 0x%x\n",
1871 			 crq->status);
1872 		ibmvmc_return_hmc(&hmcs[hmc_index], false);
1873 		return;
1874 	}
1875 
1876 	if (hmcs[hmc_index].state == ibmhmc_state_opening) {
1877 		buffer_id = be16_to_cpu(crq->var2.buffer_id);
1878 		if (buffer_id >= ibmvmc.max_buffer_pool_size) {
1879 			dev_err(adapter->dev, "open_resp: invalid buffer_id = 0x%x\n",
1880 				buffer_id);
1881 			hmcs[hmc_index].state = ibmhmc_state_failed;
1882 		} else {
1883 			ibmvmc_free_hmc_buffer(&hmcs[hmc_index],
1884 					       &hmcs[hmc_index].buffer[buffer_id]);
1885 			hmcs[hmc_index].state = ibmhmc_state_ready;
1886 			dev_dbg(adapter->dev, "open_resp: set hmc state = ready\n");
1887 		}
1888 	} else {
1889 		dev_warn(adapter->dev, "open_resp: invalid hmc state (0x%x)\n",
1890 			 hmcs[hmc_index].state);
1891 	}
1892 }
1893 
1894 /**
1895  * ibmvmc_process_close_resp - Process Close Response
1896  *
1897  * @crq: ibmvmc_crq_msg struct
1898  * @adapter:    crq_server_adapter struct
1899  *
1900  * This command is sent by the hypervisor in response to the managemant
1901  * application Interface Close message.
1902  *
1903  * If the close fails, simply reset the entire driver as the state of the VMC
1904  * must be in tough shape.
1905  */
1906 static void ibmvmc_process_close_resp(struct ibmvmc_crq_msg *crq,
1907 				      struct crq_server_adapter *adapter)
1908 {
1909 	unsigned char hmc_index;
1910 
1911 	hmc_index = crq->hmc_index;
1912 	if (hmc_index > ibmvmc.max_hmc_index) {
1913 		ibmvmc_reset(adapter, false);
1914 		return;
1915 	}
1916 
1917 	if (crq->status) {
1918 		dev_warn(adapter->dev, "close_resp: failed - status 0x%x\n",
1919 			 crq->status);
1920 		ibmvmc_reset(adapter, false);
1921 		return;
1922 	}
1923 
1924 	ibmvmc_return_hmc(&hmcs[hmc_index], false);
1925 }
1926 
1927 /**
1928  * ibmvmc_crq_process - Process CRQ
1929  *
1930  * @adapter:    crq_server_adapter struct
1931  * @crq:	ibmvmc_crq_msg struct
1932  *
1933  * Process the CRQ message based upon the type of message received.
1934  *
1935  */
1936 static void ibmvmc_crq_process(struct crq_server_adapter *adapter,
1937 			       struct ibmvmc_crq_msg *crq)
1938 {
1939 	switch (crq->type) {
1940 	case VMC_MSG_CAP_RESP:
1941 		dev_dbg(adapter->dev, "CRQ recv: capabilities resp (0x%x)\n",
1942 			crq->type);
1943 		if (ibmvmc.state == ibmvmc_state_capabilities)
1944 			ibmvmc_process_capabilities(adapter, crq);
1945 		else
1946 			dev_warn(adapter->dev, "caps msg invalid in state 0x%x\n",
1947 				 ibmvmc.state);
1948 		break;
1949 	case VMC_MSG_OPEN_RESP:
1950 		dev_dbg(adapter->dev, "CRQ recv: open resp (0x%x)\n",
1951 			crq->type);
1952 		if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
1953 			ibmvmc_process_open_resp(crq, adapter);
1954 		break;
1955 	case VMC_MSG_ADD_BUF:
1956 		dev_dbg(adapter->dev, "CRQ recv: add buf (0x%x)\n",
1957 			crq->type);
1958 		if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
1959 			ibmvmc_add_buffer(adapter, crq);
1960 		break;
1961 	case VMC_MSG_REM_BUF:
1962 		dev_dbg(adapter->dev, "CRQ recv: rem buf (0x%x)\n",
1963 			crq->type);
1964 		if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
1965 			ibmvmc_rem_buffer(adapter, crq);
1966 		break;
1967 	case VMC_MSG_SIGNAL:
1968 		dev_dbg(adapter->dev, "CRQ recv: signal msg (0x%x)\n",
1969 			crq->type);
1970 		if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
1971 			ibmvmc_recv_msg(adapter, crq);
1972 		break;
1973 	case VMC_MSG_CLOSE_RESP:
1974 		dev_dbg(adapter->dev, "CRQ recv: close resp (0x%x)\n",
1975 			crq->type);
1976 		if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
1977 			ibmvmc_process_close_resp(crq, adapter);
1978 		break;
1979 	case VMC_MSG_CAP:
1980 	case VMC_MSG_OPEN:
1981 	case VMC_MSG_CLOSE:
1982 	case VMC_MSG_ADD_BUF_RESP:
1983 	case VMC_MSG_REM_BUF_RESP:
1984 		dev_warn(adapter->dev, "CRQ recv: unexpected msg (0x%x)\n",
1985 			 crq->type);
1986 		break;
1987 	default:
1988 		dev_warn(adapter->dev, "CRQ recv: unknown msg (0x%x)\n",
1989 			 crq->type);
1990 		break;
1991 	}
1992 }
1993 
1994 /**
1995  * ibmvmc_handle_crq_init - Handle CRQ Init
1996  *
1997  * @crq:	ibmvmc_crq_msg struct
1998  * @adapter:	crq_server_adapter struct
1999  *
2000  * Handle the type of crq initialization based on whether
2001  * it is a message or a response.
2002  *
2003  */
2004 static void ibmvmc_handle_crq_init(struct ibmvmc_crq_msg *crq,
2005 				   struct crq_server_adapter *adapter)
2006 {
2007 	switch (crq->type) {
2008 	case 0x01:	/* Initialization message */
2009 		dev_dbg(adapter->dev, "CRQ recv: CRQ init msg - state 0x%x\n",
2010 			ibmvmc.state);
2011 		if (ibmvmc.state == ibmvmc_state_crqinit) {
2012 			/* Send back a response */
2013 			if (ibmvmc_send_crq(adapter, 0xC002000000000000,
2014 					    0) == 0)
2015 				ibmvmc_send_capabilities(adapter);
2016 			else
2017 				dev_err(adapter->dev, " Unable to send init rsp\n");
2018 		} else {
2019 			dev_err(adapter->dev, "Invalid state 0x%x mtu = 0x%x\n",
2020 				ibmvmc.state, ibmvmc.max_mtu);
2021 		}
2022 
2023 		break;
2024 	case 0x02:	/* Initialization response */
2025 		dev_dbg(adapter->dev, "CRQ recv: initialization resp msg - state 0x%x\n",
2026 			ibmvmc.state);
2027 		if (ibmvmc.state == ibmvmc_state_crqinit)
2028 			ibmvmc_send_capabilities(adapter);
2029 		break;
2030 	default:
2031 		dev_warn(adapter->dev, "Unknown crq message type 0x%lx\n",
2032 			 (unsigned long)crq->type);
2033 	}
2034 }
2035 
2036 /**
2037  * ibmvmc_handle_crq - Handle CRQ
2038  *
2039  * @crq:	ibmvmc_crq_msg struct
2040  * @adapter:	crq_server_adapter struct
2041  *
2042  * Read the command elements from the command queue and execute the
2043  * requests based upon the type of crq message.
2044  *
2045  */
2046 static void ibmvmc_handle_crq(struct ibmvmc_crq_msg *crq,
2047 			      struct crq_server_adapter *adapter)
2048 {
2049 	switch (crq->valid) {
2050 	case 0xC0:		/* initialization */
2051 		ibmvmc_handle_crq_init(crq, adapter);
2052 		break;
2053 	case 0xFF:	/* Hypervisor telling us the connection is closed */
2054 		dev_warn(adapter->dev, "CRQ recv: virtual adapter failed - resetting.\n");
2055 		ibmvmc_reset(adapter, true);
2056 		break;
2057 	case 0x80:	/* real payload */
2058 		ibmvmc_crq_process(adapter, crq);
2059 		break;
2060 	default:
2061 		dev_warn(adapter->dev, "CRQ recv: unknown msg 0x%02x.\n",
2062 			 crq->valid);
2063 		break;
2064 	}
2065 }
2066 
2067 static void ibmvmc_task(unsigned long data)
2068 {
2069 	struct crq_server_adapter *adapter =
2070 		(struct crq_server_adapter *)data;
2071 	struct vio_dev *vdev = to_vio_dev(adapter->dev);
2072 	struct ibmvmc_crq_msg *crq;
2073 	int done = 0;
2074 
2075 	while (!done) {
2076 		/* Pull all the valid messages off the CRQ */
2077 		while ((crq = crq_queue_next_crq(&adapter->queue)) != NULL) {
2078 			ibmvmc_handle_crq(crq, adapter);
2079 			crq->valid = 0x00;
2080 			/* CRQ reset was requested, stop processing CRQs.
2081 			 * Interrupts will be re-enabled by the reset task.
2082 			 */
2083 			if (ibmvmc.state == ibmvmc_state_sched_reset)
2084 				return;
2085 		}
2086 
2087 		vio_enable_interrupts(vdev);
2088 		crq = crq_queue_next_crq(&adapter->queue);
2089 		if (crq) {
2090 			vio_disable_interrupts(vdev);
2091 			ibmvmc_handle_crq(crq, adapter);
2092 			crq->valid = 0x00;
2093 			/* CRQ reset was requested, stop processing CRQs.
2094 			 * Interrupts will be re-enabled by the reset task.
2095 			 */
2096 			if (ibmvmc.state == ibmvmc_state_sched_reset)
2097 				return;
2098 		} else {
2099 			done = 1;
2100 		}
2101 	}
2102 }
2103 
2104 /**
2105  * ibmvmc_init_crq_queue - Init CRQ Queue
2106  *
2107  * @adapter:	crq_server_adapter struct
2108  *
2109  * Return:
2110  *	0 - Success
2111  *	Non-zero - Failure
2112  */
2113 static int ibmvmc_init_crq_queue(struct crq_server_adapter *adapter)
2114 {
2115 	struct vio_dev *vdev = to_vio_dev(adapter->dev);
2116 	struct crq_queue *queue = &adapter->queue;
2117 	int rc = 0;
2118 	int retrc = 0;
2119 
2120 	queue->msgs = (struct ibmvmc_crq_msg *)get_zeroed_page(GFP_KERNEL);
2121 
2122 	if (!queue->msgs)
2123 		goto malloc_failed;
2124 
2125 	queue->size = PAGE_SIZE / sizeof(*queue->msgs);
2126 
2127 	queue->msg_token = dma_map_single(adapter->dev, queue->msgs,
2128 					  queue->size * sizeof(*queue->msgs),
2129 					  DMA_BIDIRECTIONAL);
2130 
2131 	if (dma_mapping_error(adapter->dev, queue->msg_token))
2132 		goto map_failed;
2133 
2134 	retrc = plpar_hcall_norets(H_REG_CRQ,
2135 				   vdev->unit_address,
2136 				   queue->msg_token, PAGE_SIZE);
2137 	rc = retrc;
2138 
2139 	if (rc == H_RESOURCE)
2140 		rc = ibmvmc_reset_crq_queue(adapter);
2141 
2142 	if (rc == 2) {
2143 		dev_warn(adapter->dev, "Partner adapter not ready\n");
2144 		retrc = 0;
2145 	} else if (rc != 0) {
2146 		dev_err(adapter->dev, "Error %d opening adapter\n", rc);
2147 		goto reg_crq_failed;
2148 	}
2149 
2150 	queue->cur = 0;
2151 	spin_lock_init(&queue->lock);
2152 
2153 	tasklet_init(&adapter->work_task, ibmvmc_task, (unsigned long)adapter);
2154 
2155 	if (request_irq(vdev->irq,
2156 			ibmvmc_handle_event,
2157 			0, "ibmvmc", (void *)adapter) != 0) {
2158 		dev_err(adapter->dev, "couldn't register irq 0x%x\n",
2159 			vdev->irq);
2160 		goto req_irq_failed;
2161 	}
2162 
2163 	rc = vio_enable_interrupts(vdev);
2164 	if (rc != 0) {
2165 		dev_err(adapter->dev, "Error %d enabling interrupts!!!\n", rc);
2166 		goto req_irq_failed;
2167 	}
2168 
2169 	return retrc;
2170 
2171 req_irq_failed:
2172 	/* Cannot have any work since we either never got our IRQ registered,
2173 	 * or never got interrupts enabled
2174 	 */
2175 	tasklet_kill(&adapter->work_task);
2176 	h_free_crq(vdev->unit_address);
2177 reg_crq_failed:
2178 	dma_unmap_single(adapter->dev,
2179 			 queue->msg_token,
2180 			 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
2181 map_failed:
2182 	free_page((unsigned long)queue->msgs);
2183 malloc_failed:
2184 	return -ENOMEM;
2185 }
2186 
2187 /* Fill in the liobn and riobn fields on the adapter */
2188 static int read_dma_window(struct vio_dev *vdev,
2189 			   struct crq_server_adapter *adapter)
2190 {
2191 	const __be32 *dma_window;
2192 	const __be32 *prop;
2193 
2194 	/* TODO Using of_parse_dma_window would be better, but it doesn't give
2195 	 * a way to read multiple windows without already knowing the size of
2196 	 * a window or the number of windows
2197 	 */
2198 	dma_window =
2199 		(const __be32 *)vio_get_attribute(vdev, "ibm,my-dma-window",
2200 						NULL);
2201 	if (!dma_window) {
2202 		dev_warn(adapter->dev, "Couldn't find ibm,my-dma-window property\n");
2203 		return -1;
2204 	}
2205 
2206 	adapter->liobn = be32_to_cpu(*dma_window);
2207 	dma_window++;
2208 
2209 	prop = (const __be32 *)vio_get_attribute(vdev, "ibm,#dma-address-cells",
2210 						NULL);
2211 	if (!prop) {
2212 		dev_warn(adapter->dev, "Couldn't find ibm,#dma-address-cells property\n");
2213 		dma_window++;
2214 	} else {
2215 		dma_window += be32_to_cpu(*prop);
2216 	}
2217 
2218 	prop = (const __be32 *)vio_get_attribute(vdev, "ibm,#dma-size-cells",
2219 						NULL);
2220 	if (!prop) {
2221 		dev_warn(adapter->dev, "Couldn't find ibm,#dma-size-cells property\n");
2222 		dma_window++;
2223 	} else {
2224 		dma_window += be32_to_cpu(*prop);
2225 	}
2226 
2227 	/* dma_window should point to the second window now */
2228 	adapter->riobn = be32_to_cpu(*dma_window);
2229 
2230 	return 0;
2231 }
2232 
2233 static int ibmvmc_probe(struct vio_dev *vdev, const struct vio_device_id *id)
2234 {
2235 	struct crq_server_adapter *adapter = &ibmvmc_adapter;
2236 	int rc;
2237 
2238 	dev_set_drvdata(&vdev->dev, NULL);
2239 	memset(adapter, 0, sizeof(*adapter));
2240 	adapter->dev = &vdev->dev;
2241 
2242 	dev_info(adapter->dev, "Probe for UA 0x%x\n", vdev->unit_address);
2243 
2244 	rc = read_dma_window(vdev, adapter);
2245 	if (rc != 0) {
2246 		ibmvmc.state = ibmvmc_state_failed;
2247 		return -1;
2248 	}
2249 
2250 	dev_dbg(adapter->dev, "Probe: liobn 0x%x, riobn 0x%x\n",
2251 		adapter->liobn, adapter->riobn);
2252 
2253 	init_waitqueue_head(&adapter->reset_wait_queue);
2254 	adapter->reset_task = kthread_run(ibmvmc_reset_task, adapter, "ibmvmc");
2255 	if (IS_ERR(adapter->reset_task)) {
2256 		dev_err(adapter->dev, "Failed to start reset thread\n");
2257 		ibmvmc.state = ibmvmc_state_failed;
2258 		rc = PTR_ERR(adapter->reset_task);
2259 		adapter->reset_task = NULL;
2260 		return rc;
2261 	}
2262 
2263 	rc = ibmvmc_init_crq_queue(adapter);
2264 	if (rc != 0 && rc != H_RESOURCE) {
2265 		dev_err(adapter->dev, "Error initializing CRQ.  rc = 0x%x\n",
2266 			rc);
2267 		ibmvmc.state = ibmvmc_state_failed;
2268 		goto crq_failed;
2269 	}
2270 
2271 	ibmvmc.state = ibmvmc_state_crqinit;
2272 
2273 	/* Try to send an initialization message.  Note that this is allowed
2274 	 * to fail if the other end is not acive.  In that case we just wait
2275 	 * for the other side to initialize.
2276 	 */
2277 	if (ibmvmc_send_crq(adapter, 0xC001000000000000LL, 0) != 0 &&
2278 	    rc != H_RESOURCE)
2279 		dev_warn(adapter->dev, "Failed to send initialize CRQ message\n");
2280 
2281 	dev_set_drvdata(&vdev->dev, adapter);
2282 
2283 	return 0;
2284 
2285 crq_failed:
2286 	kthread_stop(adapter->reset_task);
2287 	adapter->reset_task = NULL;
2288 	return -EPERM;
2289 }
2290 
2291 static void ibmvmc_remove(struct vio_dev *vdev)
2292 {
2293 	struct crq_server_adapter *adapter = dev_get_drvdata(&vdev->dev);
2294 
2295 	dev_info(adapter->dev, "Entering remove for UA 0x%x\n",
2296 		 vdev->unit_address);
2297 	ibmvmc_release_crq_queue(adapter);
2298 }
2299 
2300 static struct vio_device_id ibmvmc_device_table[] = {
2301 	{ "ibm,vmc", "IBM,vmc" },
2302 	{ "", "" }
2303 };
2304 MODULE_DEVICE_TABLE(vio, ibmvmc_device_table);
2305 
2306 static struct vio_driver ibmvmc_driver = {
2307 	.name        = ibmvmc_driver_name,
2308 	.id_table    = ibmvmc_device_table,
2309 	.probe       = ibmvmc_probe,
2310 	.remove      = ibmvmc_remove,
2311 };
2312 
2313 static void __init ibmvmc_scrub_module_parms(void)
2314 {
2315 	if (ibmvmc_max_mtu > MAX_MTU) {
2316 		pr_warn("ibmvmc: Max MTU reduced to %d\n", MAX_MTU);
2317 		ibmvmc_max_mtu = MAX_MTU;
2318 	} else if (ibmvmc_max_mtu < MIN_MTU) {
2319 		pr_warn("ibmvmc: Max MTU increased to %d\n", MIN_MTU);
2320 		ibmvmc_max_mtu = MIN_MTU;
2321 	}
2322 
2323 	if (ibmvmc_max_buf_pool_size > MAX_BUF_POOL_SIZE) {
2324 		pr_warn("ibmvmc: Max buffer pool size reduced to %d\n",
2325 			MAX_BUF_POOL_SIZE);
2326 		ibmvmc_max_buf_pool_size = MAX_BUF_POOL_SIZE;
2327 	} else if (ibmvmc_max_buf_pool_size < MIN_BUF_POOL_SIZE) {
2328 		pr_warn("ibmvmc: Max buffer pool size increased to %d\n",
2329 			MIN_BUF_POOL_SIZE);
2330 		ibmvmc_max_buf_pool_size = MIN_BUF_POOL_SIZE;
2331 	}
2332 
2333 	if (ibmvmc_max_hmcs > MAX_HMCS) {
2334 		pr_warn("ibmvmc: Max HMCs reduced to %d\n", MAX_HMCS);
2335 		ibmvmc_max_hmcs = MAX_HMCS;
2336 	} else if (ibmvmc_max_hmcs < MIN_HMCS) {
2337 		pr_warn("ibmvmc: Max HMCs increased to %d\n", MIN_HMCS);
2338 		ibmvmc_max_hmcs = MIN_HMCS;
2339 	}
2340 }
2341 
2342 static struct miscdevice ibmvmc_miscdev = {
2343 	.name = ibmvmc_driver_name,
2344 	.minor = MISC_DYNAMIC_MINOR,
2345 	.fops = &ibmvmc_fops,
2346 };
2347 
2348 static int __init ibmvmc_module_init(void)
2349 {
2350 	int rc, i, j;
2351 
2352 	ibmvmc.state = ibmvmc_state_initial;
2353 	pr_info("ibmvmc: version %s\n", IBMVMC_DRIVER_VERSION);
2354 
2355 	rc = misc_register(&ibmvmc_miscdev);
2356 	if (rc) {
2357 		pr_err("ibmvmc: misc registration failed\n");
2358 		goto misc_register_failed;
2359 	}
2360 	pr_info("ibmvmc: node %d:%d\n", MISC_MAJOR,
2361 		ibmvmc_miscdev.minor);
2362 
2363 	/* Initialize data structures */
2364 	memset(hmcs, 0, sizeof(struct ibmvmc_hmc) * MAX_HMCS);
2365 	for (i = 0; i < MAX_HMCS; i++) {
2366 		spin_lock_init(&hmcs[i].lock);
2367 		hmcs[i].state = ibmhmc_state_free;
2368 		for (j = 0; j < MAX_BUF_POOL_SIZE; j++)
2369 			hmcs[i].queue_outbound_msgs[j] = VMC_INVALID_BUFFER_ID;
2370 	}
2371 
2372 	/* Sanity check module parms */
2373 	ibmvmc_scrub_module_parms();
2374 
2375 	/*
2376 	 * Initialize some reasonable values.  Might be negotiated smaller
2377 	 * values during the capabilities exchange.
2378 	 */
2379 	ibmvmc.max_mtu = ibmvmc_max_mtu;
2380 	ibmvmc.max_buffer_pool_size = ibmvmc_max_buf_pool_size;
2381 	ibmvmc.max_hmc_index = ibmvmc_max_hmcs - 1;
2382 
2383 	rc = vio_register_driver(&ibmvmc_driver);
2384 
2385 	if (rc) {
2386 		pr_err("ibmvmc: rc %d from vio_register_driver\n", rc);
2387 		goto vio_reg_failed;
2388 	}
2389 
2390 	return 0;
2391 
2392 vio_reg_failed:
2393 	misc_deregister(&ibmvmc_miscdev);
2394 misc_register_failed:
2395 	return rc;
2396 }
2397 
2398 static void __exit ibmvmc_module_exit(void)
2399 {
2400 	pr_info("ibmvmc: module exit\n");
2401 	vio_unregister_driver(&ibmvmc_driver);
2402 	misc_deregister(&ibmvmc_miscdev);
2403 }
2404 
2405 module_init(ibmvmc_module_init);
2406 module_exit(ibmvmc_module_exit);
2407 
2408 module_param_named(buf_pool_size, ibmvmc_max_buf_pool_size,
2409 		   int, 0644);
2410 MODULE_PARM_DESC(buf_pool_size, "Buffer pool size");
2411 module_param_named(max_hmcs, ibmvmc_max_hmcs, int, 0644);
2412 MODULE_PARM_DESC(max_hmcs, "Max HMCs");
2413 module_param_named(max_mtu, ibmvmc_max_mtu, int, 0644);
2414 MODULE_PARM_DESC(max_mtu, "Max MTU");
2415 
2416 MODULE_AUTHOR("Steven Royer <seroyer@linux.vnet.ibm.com>");
2417 MODULE_DESCRIPTION("IBM VMC");
2418 MODULE_VERSION(IBMVMC_DRIVER_VERSION);
2419 MODULE_LICENSE("GPL v2");
2420