xref: /linux/block/bsg.c (revision bf80eef2212a1e8451df13b52533f4bc31bb4f8e)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * bsg.c - block layer implementation of the sg v4 interface
4  */
5 #include <linux/module.h>
6 #include <linux/init.h>
7 #include <linux/file.h>
8 #include <linux/blkdev.h>
9 #include <linux/cdev.h>
10 #include <linux/jiffies.h>
11 #include <linux/percpu.h>
12 #include <linux/idr.h>
13 #include <linux/bsg.h>
14 #include <linux/slab.h>
15 
16 #include <scsi/scsi.h>
17 #include <scsi/scsi_ioctl.h>
18 #include <scsi/sg.h>
19 
20 #define BSG_DESCRIPTION	"Block layer SCSI generic (bsg) driver"
21 #define BSG_VERSION	"0.4"
22 
23 struct bsg_device {
24 	struct request_queue *queue;
25 	struct device device;
26 	struct cdev cdev;
27 	int max_queue;
28 	unsigned int timeout;
29 	unsigned int reserved_size;
30 	bsg_sg_io_fn *sg_io_fn;
31 };
32 
33 static inline struct bsg_device *to_bsg_device(struct inode *inode)
34 {
35 	return container_of(inode->i_cdev, struct bsg_device, cdev);
36 }
37 
38 #define BSG_DEFAULT_CMDS	64
39 #define BSG_MAX_DEVS		32768
40 
41 static DEFINE_IDA(bsg_minor_ida);
42 static struct class *bsg_class;
43 static int bsg_major;
44 
45 static unsigned int bsg_timeout(struct bsg_device *bd, struct sg_io_v4 *hdr)
46 {
47 	unsigned int timeout = BLK_DEFAULT_SG_TIMEOUT;
48 
49 	if (hdr->timeout)
50 		timeout = msecs_to_jiffies(hdr->timeout);
51 	else if (bd->timeout)
52 		timeout = bd->timeout;
53 
54 	return max_t(unsigned int, timeout, BLK_MIN_SG_TIMEOUT);
55 }
56 
57 static int bsg_sg_io(struct bsg_device *bd, fmode_t mode, void __user *uarg)
58 {
59 	struct sg_io_v4 hdr;
60 	int ret;
61 
62 	if (copy_from_user(&hdr, uarg, sizeof(hdr)))
63 		return -EFAULT;
64 	if (hdr.guard != 'Q')
65 		return -EINVAL;
66 	ret = bd->sg_io_fn(bd->queue, &hdr, mode, bsg_timeout(bd, &hdr));
67 	if (!ret && copy_to_user(uarg, &hdr, sizeof(hdr)))
68 		return -EFAULT;
69 	return ret;
70 }
71 
72 static int bsg_open(struct inode *inode, struct file *file)
73 {
74 	if (!blk_get_queue(to_bsg_device(inode)->queue))
75 		return -ENXIO;
76 	return 0;
77 }
78 
79 static int bsg_release(struct inode *inode, struct file *file)
80 {
81 	blk_put_queue(to_bsg_device(inode)->queue);
82 	return 0;
83 }
84 
85 static int bsg_get_command_q(struct bsg_device *bd, int __user *uarg)
86 {
87 	return put_user(READ_ONCE(bd->max_queue), uarg);
88 }
89 
90 static int bsg_set_command_q(struct bsg_device *bd, int __user *uarg)
91 {
92 	int max_queue;
93 
94 	if (get_user(max_queue, uarg))
95 		return -EFAULT;
96 	if (max_queue < 1)
97 		return -EINVAL;
98 	WRITE_ONCE(bd->max_queue, max_queue);
99 	return 0;
100 }
101 
102 static long bsg_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
103 {
104 	struct bsg_device *bd = to_bsg_device(file_inode(file));
105 	struct request_queue *q = bd->queue;
106 	void __user *uarg = (void __user *) arg;
107 	int __user *intp = uarg;
108 	int val;
109 
110 	switch (cmd) {
111 	/*
112 	 * Our own ioctls
113 	 */
114 	case SG_GET_COMMAND_Q:
115 		return bsg_get_command_q(bd, uarg);
116 	case SG_SET_COMMAND_Q:
117 		return bsg_set_command_q(bd, uarg);
118 
119 	/*
120 	 * SCSI/sg ioctls
121 	 */
122 	case SG_GET_VERSION_NUM:
123 		return put_user(30527, intp);
124 	case SCSI_IOCTL_GET_IDLUN:
125 		return put_user(0, intp);
126 	case SCSI_IOCTL_GET_BUS_NUMBER:
127 		return put_user(0, intp);
128 	case SG_SET_TIMEOUT:
129 		if (get_user(val, intp))
130 			return -EFAULT;
131 		bd->timeout = clock_t_to_jiffies(val);
132 		return 0;
133 	case SG_GET_TIMEOUT:
134 		return jiffies_to_clock_t(bd->timeout);
135 	case SG_GET_RESERVED_SIZE:
136 		return put_user(min(bd->reserved_size, queue_max_bytes(q)),
137 				intp);
138 	case SG_SET_RESERVED_SIZE:
139 		if (get_user(val, intp))
140 			return -EFAULT;
141 		if (val < 0)
142 			return -EINVAL;
143 		bd->reserved_size =
144 			min_t(unsigned int, val, queue_max_bytes(q));
145 		return 0;
146 	case SG_EMULATED_HOST:
147 		return put_user(1, intp);
148 	case SG_IO:
149 		return bsg_sg_io(bd, file->f_mode, uarg);
150 	case SCSI_IOCTL_SEND_COMMAND:
151 		pr_warn_ratelimited("%s: calling unsupported SCSI_IOCTL_SEND_COMMAND\n",
152 				current->comm);
153 		return -EINVAL;
154 	default:
155 		return -ENOTTY;
156 	}
157 }
158 
159 static const struct file_operations bsg_fops = {
160 	.open		=	bsg_open,
161 	.release	=	bsg_release,
162 	.unlocked_ioctl	=	bsg_ioctl,
163 	.compat_ioctl	=	compat_ptr_ioctl,
164 	.owner		=	THIS_MODULE,
165 	.llseek		=	default_llseek,
166 };
167 
168 static void bsg_device_release(struct device *dev)
169 {
170 	struct bsg_device *bd = container_of(dev, struct bsg_device, device);
171 
172 	ida_free(&bsg_minor_ida, MINOR(bd->device.devt));
173 	kfree(bd);
174 }
175 
176 void bsg_unregister_queue(struct bsg_device *bd)
177 {
178 	if (bd->queue->kobj.sd)
179 		sysfs_remove_link(&bd->queue->kobj, "bsg");
180 	cdev_device_del(&bd->cdev, &bd->device);
181 	put_device(&bd->device);
182 }
183 EXPORT_SYMBOL_GPL(bsg_unregister_queue);
184 
185 struct bsg_device *bsg_register_queue(struct request_queue *q,
186 		struct device *parent, const char *name, bsg_sg_io_fn *sg_io_fn)
187 {
188 	struct bsg_device *bd;
189 	int ret;
190 
191 	bd = kzalloc(sizeof(*bd), GFP_KERNEL);
192 	if (!bd)
193 		return ERR_PTR(-ENOMEM);
194 	bd->max_queue = BSG_DEFAULT_CMDS;
195 	bd->reserved_size = INT_MAX;
196 	bd->queue = q;
197 	bd->sg_io_fn = sg_io_fn;
198 
199 	ret = ida_alloc_max(&bsg_minor_ida, BSG_MAX_DEVS - 1, GFP_KERNEL);
200 	if (ret < 0) {
201 		if (ret == -ENOSPC)
202 			dev_err(parent, "bsg: too many bsg devices\n");
203 		kfree(bd);
204 		return ERR_PTR(ret);
205 	}
206 	bd->device.devt = MKDEV(bsg_major, ret);
207 	bd->device.class = bsg_class;
208 	bd->device.parent = parent;
209 	bd->device.release = bsg_device_release;
210 	dev_set_name(&bd->device, "%s", name);
211 	device_initialize(&bd->device);
212 
213 	cdev_init(&bd->cdev, &bsg_fops);
214 	bd->cdev.owner = THIS_MODULE;
215 	ret = cdev_device_add(&bd->cdev, &bd->device);
216 	if (ret)
217 		goto out_put_device;
218 
219 	if (q->kobj.sd) {
220 		ret = sysfs_create_link(&q->kobj, &bd->device.kobj, "bsg");
221 		if (ret)
222 			goto out_device_del;
223 	}
224 
225 	return bd;
226 
227 out_device_del:
228 	cdev_device_del(&bd->cdev, &bd->device);
229 out_put_device:
230 	put_device(&bd->device);
231 	return ERR_PTR(ret);
232 }
233 EXPORT_SYMBOL_GPL(bsg_register_queue);
234 
235 static char *bsg_devnode(struct device *dev, umode_t *mode)
236 {
237 	return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev));
238 }
239 
240 static int __init bsg_init(void)
241 {
242 	dev_t devid;
243 	int ret;
244 
245 	bsg_class = class_create(THIS_MODULE, "bsg");
246 	if (IS_ERR(bsg_class))
247 		return PTR_ERR(bsg_class);
248 	bsg_class->devnode = bsg_devnode;
249 
250 	ret = alloc_chrdev_region(&devid, 0, BSG_MAX_DEVS, "bsg");
251 	if (ret)
252 		goto destroy_bsg_class;
253 	bsg_major = MAJOR(devid);
254 
255 	printk(KERN_INFO BSG_DESCRIPTION " version " BSG_VERSION
256 	       " loaded (major %d)\n", bsg_major);
257 	return 0;
258 
259 destroy_bsg_class:
260 	class_destroy(bsg_class);
261 	return ret;
262 }
263 
264 MODULE_AUTHOR("Jens Axboe");
265 MODULE_DESCRIPTION(BSG_DESCRIPTION);
266 MODULE_LICENSE("GPL");
267 
268 device_initcall(bsg_init);
269