xref: /linux/block/bsg.c (revision ab52c59103002b49f2455371e4b9c56ba3ef1781)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * bsg.c - block layer implementation of the sg v4 interface
4  */
5 #include <linux/module.h>
6 #include <linux/init.h>
7 #include <linux/file.h>
8 #include <linux/blkdev.h>
9 #include <linux/cdev.h>
10 #include <linux/jiffies.h>
11 #include <linux/percpu.h>
12 #include <linux/idr.h>
13 #include <linux/bsg.h>
14 #include <linux/slab.h>
15 
16 #include <scsi/scsi.h>
17 #include <scsi/scsi_ioctl.h>
18 #include <scsi/sg.h>
19 
20 #define BSG_DESCRIPTION	"Block layer SCSI generic (bsg) driver"
21 #define BSG_VERSION	"0.4"
22 
23 struct bsg_device {
24 	struct request_queue *queue;
25 	struct device device;
26 	struct cdev cdev;
27 	int max_queue;
28 	unsigned int timeout;
29 	unsigned int reserved_size;
30 	bsg_sg_io_fn *sg_io_fn;
31 };
32 
33 static inline struct bsg_device *to_bsg_device(struct inode *inode)
34 {
35 	return container_of(inode->i_cdev, struct bsg_device, cdev);
36 }
37 
38 #define BSG_DEFAULT_CMDS	64
39 #define BSG_MAX_DEVS		(1 << MINORBITS)
40 
41 static DEFINE_IDA(bsg_minor_ida);
42 static const struct class bsg_class;
43 static int bsg_major;
44 
45 static unsigned int bsg_timeout(struct bsg_device *bd, struct sg_io_v4 *hdr)
46 {
47 	unsigned int timeout = BLK_DEFAULT_SG_TIMEOUT;
48 
49 	if (hdr->timeout)
50 		timeout = msecs_to_jiffies(hdr->timeout);
51 	else if (bd->timeout)
52 		timeout = bd->timeout;
53 
54 	return max_t(unsigned int, timeout, BLK_MIN_SG_TIMEOUT);
55 }
56 
57 static int bsg_sg_io(struct bsg_device *bd, bool open_for_write,
58 		     void __user *uarg)
59 {
60 	struct sg_io_v4 hdr;
61 	int ret;
62 
63 	if (copy_from_user(&hdr, uarg, sizeof(hdr)))
64 		return -EFAULT;
65 	if (hdr.guard != 'Q')
66 		return -EINVAL;
67 	ret = bd->sg_io_fn(bd->queue, &hdr, open_for_write,
68 			   bsg_timeout(bd, &hdr));
69 	if (!ret && copy_to_user(uarg, &hdr, sizeof(hdr)))
70 		return -EFAULT;
71 	return ret;
72 }
73 
74 static int bsg_open(struct inode *inode, struct file *file)
75 {
76 	if (!blk_get_queue(to_bsg_device(inode)->queue))
77 		return -ENXIO;
78 	return 0;
79 }
80 
81 static int bsg_release(struct inode *inode, struct file *file)
82 {
83 	blk_put_queue(to_bsg_device(inode)->queue);
84 	return 0;
85 }
86 
87 static int bsg_get_command_q(struct bsg_device *bd, int __user *uarg)
88 {
89 	return put_user(READ_ONCE(bd->max_queue), uarg);
90 }
91 
92 static int bsg_set_command_q(struct bsg_device *bd, int __user *uarg)
93 {
94 	int max_queue;
95 
96 	if (get_user(max_queue, uarg))
97 		return -EFAULT;
98 	if (max_queue < 1)
99 		return -EINVAL;
100 	WRITE_ONCE(bd->max_queue, max_queue);
101 	return 0;
102 }
103 
104 static long bsg_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
105 {
106 	struct bsg_device *bd = to_bsg_device(file_inode(file));
107 	struct request_queue *q = bd->queue;
108 	void __user *uarg = (void __user *) arg;
109 	int __user *intp = uarg;
110 	int val;
111 
112 	switch (cmd) {
113 	/*
114 	 * Our own ioctls
115 	 */
116 	case SG_GET_COMMAND_Q:
117 		return bsg_get_command_q(bd, uarg);
118 	case SG_SET_COMMAND_Q:
119 		return bsg_set_command_q(bd, uarg);
120 
121 	/*
122 	 * SCSI/sg ioctls
123 	 */
124 	case SG_GET_VERSION_NUM:
125 		return put_user(30527, intp);
126 	case SCSI_IOCTL_GET_IDLUN:
127 		return put_user(0, intp);
128 	case SCSI_IOCTL_GET_BUS_NUMBER:
129 		return put_user(0, intp);
130 	case SG_SET_TIMEOUT:
131 		if (get_user(val, intp))
132 			return -EFAULT;
133 		bd->timeout = clock_t_to_jiffies(val);
134 		return 0;
135 	case SG_GET_TIMEOUT:
136 		return jiffies_to_clock_t(bd->timeout);
137 	case SG_GET_RESERVED_SIZE:
138 		return put_user(min(bd->reserved_size, queue_max_bytes(q)),
139 				intp);
140 	case SG_SET_RESERVED_SIZE:
141 		if (get_user(val, intp))
142 			return -EFAULT;
143 		if (val < 0)
144 			return -EINVAL;
145 		bd->reserved_size =
146 			min_t(unsigned int, val, queue_max_bytes(q));
147 		return 0;
148 	case SG_EMULATED_HOST:
149 		return put_user(1, intp);
150 	case SG_IO:
151 		return bsg_sg_io(bd, file->f_mode & FMODE_WRITE, uarg);
152 	case SCSI_IOCTL_SEND_COMMAND:
153 		pr_warn_ratelimited("%s: calling unsupported SCSI_IOCTL_SEND_COMMAND\n",
154 				current->comm);
155 		return -EINVAL;
156 	default:
157 		return -ENOTTY;
158 	}
159 }
160 
161 static const struct file_operations bsg_fops = {
162 	.open		=	bsg_open,
163 	.release	=	bsg_release,
164 	.unlocked_ioctl	=	bsg_ioctl,
165 	.compat_ioctl	=	compat_ptr_ioctl,
166 	.owner		=	THIS_MODULE,
167 	.llseek		=	default_llseek,
168 };
169 
170 static void bsg_device_release(struct device *dev)
171 {
172 	struct bsg_device *bd = container_of(dev, struct bsg_device, device);
173 
174 	ida_free(&bsg_minor_ida, MINOR(bd->device.devt));
175 	kfree(bd);
176 }
177 
178 void bsg_unregister_queue(struct bsg_device *bd)
179 {
180 	struct gendisk *disk = bd->queue->disk;
181 
182 	if (disk && disk->queue_kobj.sd)
183 		sysfs_remove_link(&disk->queue_kobj, "bsg");
184 	cdev_device_del(&bd->cdev, &bd->device);
185 	put_device(&bd->device);
186 }
187 EXPORT_SYMBOL_GPL(bsg_unregister_queue);
188 
189 struct bsg_device *bsg_register_queue(struct request_queue *q,
190 		struct device *parent, const char *name, bsg_sg_io_fn *sg_io_fn)
191 {
192 	struct bsg_device *bd;
193 	int ret;
194 
195 	bd = kzalloc(sizeof(*bd), GFP_KERNEL);
196 	if (!bd)
197 		return ERR_PTR(-ENOMEM);
198 	bd->max_queue = BSG_DEFAULT_CMDS;
199 	bd->reserved_size = INT_MAX;
200 	bd->queue = q;
201 	bd->sg_io_fn = sg_io_fn;
202 
203 	ret = ida_alloc_max(&bsg_minor_ida, BSG_MAX_DEVS - 1, GFP_KERNEL);
204 	if (ret < 0) {
205 		if (ret == -ENOSPC)
206 			dev_err(parent, "bsg: too many bsg devices\n");
207 		kfree(bd);
208 		return ERR_PTR(ret);
209 	}
210 	bd->device.devt = MKDEV(bsg_major, ret);
211 	bd->device.class = &bsg_class;
212 	bd->device.parent = parent;
213 	bd->device.release = bsg_device_release;
214 	dev_set_name(&bd->device, "%s", name);
215 	device_initialize(&bd->device);
216 
217 	cdev_init(&bd->cdev, &bsg_fops);
218 	bd->cdev.owner = THIS_MODULE;
219 	ret = cdev_device_add(&bd->cdev, &bd->device);
220 	if (ret)
221 		goto out_put_device;
222 
223 	if (q->disk && q->disk->queue_kobj.sd) {
224 		ret = sysfs_create_link(&q->disk->queue_kobj, &bd->device.kobj,
225 					"bsg");
226 		if (ret)
227 			goto out_device_del;
228 	}
229 
230 	return bd;
231 
232 out_device_del:
233 	cdev_device_del(&bd->cdev, &bd->device);
234 out_put_device:
235 	put_device(&bd->device);
236 	return ERR_PTR(ret);
237 }
238 EXPORT_SYMBOL_GPL(bsg_register_queue);
239 
240 static char *bsg_devnode(const struct device *dev, umode_t *mode)
241 {
242 	return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev));
243 }
244 
245 static const struct class bsg_class = {
246 	.name		= "bsg",
247 	.devnode	= bsg_devnode,
248 };
249 
250 static int __init bsg_init(void)
251 {
252 	dev_t devid;
253 	int ret;
254 
255 	ret = class_register(&bsg_class);
256 	if (ret)
257 		return ret;
258 
259 	ret = alloc_chrdev_region(&devid, 0, BSG_MAX_DEVS, "bsg");
260 	if (ret)
261 		goto destroy_bsg_class;
262 	bsg_major = MAJOR(devid);
263 
264 	printk(KERN_INFO BSG_DESCRIPTION " version " BSG_VERSION
265 	       " loaded (major %d)\n", bsg_major);
266 	return 0;
267 
268 destroy_bsg_class:
269 	class_unregister(&bsg_class);
270 	return ret;
271 }
272 
273 MODULE_AUTHOR("Jens Axboe");
274 MODULE_DESCRIPTION(BSG_DESCRIPTION);
275 MODULE_LICENSE("GPL");
276 
277 device_initcall(bsg_init);
278