xref: /linux/drivers/cxl/core/port.c (revision 01ecadbe09b6c685de413ada8ba6688e9467c4b3)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright(c) 2020 Intel Corporation. All rights reserved. */
3 #include <linux/platform_device.h>
4 #include <linux/memregion.h>
5 #include <linux/workqueue.h>
6 #include <linux/debugfs.h>
7 #include <linux/device.h>
8 #include <linux/module.h>
9 #include <linux/pci.h>
10 #include <linux/slab.h>
11 #include <linux/idr.h>
12 #include <linux/node.h>
13 #include <cxl/einj.h>
14 #include <cxlmem.h>
15 #include <cxlpci.h>
16 #include <cxl.h>
17 #include "core.h"
18 
19 /**
20  * DOC: cxl core
21  *
22  * The CXL core provides a set of interfaces that can be consumed by CXL aware
23  * drivers. The interfaces allow for creation, modification, and destruction of
24  * regions, memory devices, ports, and decoders. CXL aware drivers must register
25  * with the CXL core via these interfaces in order to be able to participate in
26  * cross-device interleave coordination. The CXL core also establishes and
27  * maintains the bridge to the nvdimm subsystem.
28  *
29  * CXL core introduces sysfs hierarchy to control the devices that are
30  * instantiated by the core.
31  */
32 
33 /*
34  * All changes to the interleave configuration occur with this lock held
35  * for write.
36  */
37 DECLARE_RWSEM(cxl_region_rwsem);
38 
39 static DEFINE_IDA(cxl_port_ida);
40 static DEFINE_XARRAY(cxl_root_buses);
41 
cxl_num_decoders_committed(struct cxl_port * port)42 int cxl_num_decoders_committed(struct cxl_port *port)
43 {
44 	lockdep_assert_held(&cxl_region_rwsem);
45 
46 	return port->commit_end + 1;
47 }
48 
devtype_show(struct device * dev,struct device_attribute * attr,char * buf)49 static ssize_t devtype_show(struct device *dev, struct device_attribute *attr,
50 			    char *buf)
51 {
52 	return sysfs_emit(buf, "%s\n", dev->type->name);
53 }
54 static DEVICE_ATTR_RO(devtype);
55 
cxl_device_id(const struct device * dev)56 static int cxl_device_id(const struct device *dev)
57 {
58 	if (dev->type == &cxl_nvdimm_bridge_type)
59 		return CXL_DEVICE_NVDIMM_BRIDGE;
60 	if (dev->type == &cxl_nvdimm_type)
61 		return CXL_DEVICE_NVDIMM;
62 	if (dev->type == CXL_PMEM_REGION_TYPE())
63 		return CXL_DEVICE_PMEM_REGION;
64 	if (dev->type == CXL_DAX_REGION_TYPE())
65 		return CXL_DEVICE_DAX_REGION;
66 	if (is_cxl_port(dev)) {
67 		if (is_cxl_root(to_cxl_port(dev)))
68 			return CXL_DEVICE_ROOT;
69 		return CXL_DEVICE_PORT;
70 	}
71 	if (is_cxl_memdev(dev))
72 		return CXL_DEVICE_MEMORY_EXPANDER;
73 	if (dev->type == CXL_REGION_TYPE())
74 		return CXL_DEVICE_REGION;
75 	if (dev->type == &cxl_pmu_type)
76 		return CXL_DEVICE_PMU;
77 	return 0;
78 }
79 
modalias_show(struct device * dev,struct device_attribute * attr,char * buf)80 static ssize_t modalias_show(struct device *dev, struct device_attribute *attr,
81 			     char *buf)
82 {
83 	return sysfs_emit(buf, CXL_MODALIAS_FMT "\n", cxl_device_id(dev));
84 }
85 static DEVICE_ATTR_RO(modalias);
86 
87 static struct attribute *cxl_base_attributes[] = {
88 	&dev_attr_devtype.attr,
89 	&dev_attr_modalias.attr,
90 	NULL,
91 };
92 
93 struct attribute_group cxl_base_attribute_group = {
94 	.attrs = cxl_base_attributes,
95 };
96 
start_show(struct device * dev,struct device_attribute * attr,char * buf)97 static ssize_t start_show(struct device *dev, struct device_attribute *attr,
98 			  char *buf)
99 {
100 	struct cxl_decoder *cxld = to_cxl_decoder(dev);
101 
102 	return sysfs_emit(buf, "%#llx\n", cxld->hpa_range.start);
103 }
104 static DEVICE_ATTR_ADMIN_RO(start);
105 
size_show(struct device * dev,struct device_attribute * attr,char * buf)106 static ssize_t size_show(struct device *dev, struct device_attribute *attr,
107 			char *buf)
108 {
109 	struct cxl_decoder *cxld = to_cxl_decoder(dev);
110 
111 	return sysfs_emit(buf, "%#llx\n", range_len(&cxld->hpa_range));
112 }
113 static DEVICE_ATTR_RO(size);
114 
115 #define CXL_DECODER_FLAG_ATTR(name, flag)                            \
116 static ssize_t name##_show(struct device *dev,                       \
117 			   struct device_attribute *attr, char *buf) \
118 {                                                                    \
119 	struct cxl_decoder *cxld = to_cxl_decoder(dev);              \
120                                                                      \
121 	return sysfs_emit(buf, "%s\n",                               \
122 			  (cxld->flags & (flag)) ? "1" : "0");       \
123 }                                                                    \
124 static DEVICE_ATTR_RO(name)
125 
126 CXL_DECODER_FLAG_ATTR(cap_pmem, CXL_DECODER_F_PMEM);
127 CXL_DECODER_FLAG_ATTR(cap_ram, CXL_DECODER_F_RAM);
128 CXL_DECODER_FLAG_ATTR(cap_type2, CXL_DECODER_F_TYPE2);
129 CXL_DECODER_FLAG_ATTR(cap_type3, CXL_DECODER_F_TYPE3);
130 CXL_DECODER_FLAG_ATTR(locked, CXL_DECODER_F_LOCK);
131 
target_type_show(struct device * dev,struct device_attribute * attr,char * buf)132 static ssize_t target_type_show(struct device *dev,
133 				struct device_attribute *attr, char *buf)
134 {
135 	struct cxl_decoder *cxld = to_cxl_decoder(dev);
136 
137 	switch (cxld->target_type) {
138 	case CXL_DECODER_DEVMEM:
139 		return sysfs_emit(buf, "accelerator\n");
140 	case CXL_DECODER_HOSTONLYMEM:
141 		return sysfs_emit(buf, "expander\n");
142 	}
143 	return -ENXIO;
144 }
145 static DEVICE_ATTR_RO(target_type);
146 
emit_target_list(struct cxl_switch_decoder * cxlsd,char * buf)147 static ssize_t emit_target_list(struct cxl_switch_decoder *cxlsd, char *buf)
148 {
149 	struct cxl_decoder *cxld = &cxlsd->cxld;
150 	ssize_t offset = 0;
151 	int i, rc = 0;
152 
153 	for (i = 0; i < cxld->interleave_ways; i++) {
154 		struct cxl_dport *dport = cxlsd->target[i];
155 		struct cxl_dport *next = NULL;
156 
157 		if (!dport)
158 			break;
159 
160 		if (i + 1 < cxld->interleave_ways)
161 			next = cxlsd->target[i + 1];
162 		rc = sysfs_emit_at(buf, offset, "%d%s", dport->port_id,
163 				   next ? "," : "");
164 		if (rc < 0)
165 			return rc;
166 		offset += rc;
167 	}
168 
169 	return offset;
170 }
171 
target_list_show(struct device * dev,struct device_attribute * attr,char * buf)172 static ssize_t target_list_show(struct device *dev,
173 				struct device_attribute *attr, char *buf)
174 {
175 	struct cxl_switch_decoder *cxlsd = to_cxl_switch_decoder(dev);
176 	ssize_t offset;
177 	int rc;
178 
179 	guard(rwsem_read)(&cxl_region_rwsem);
180 	rc = emit_target_list(cxlsd, buf);
181 	if (rc < 0)
182 		return rc;
183 	offset = rc;
184 
185 	rc = sysfs_emit_at(buf, offset, "\n");
186 	if (rc < 0)
187 		return rc;
188 
189 	return offset + rc;
190 }
191 static DEVICE_ATTR_RO(target_list);
192 
mode_show(struct device * dev,struct device_attribute * attr,char * buf)193 static ssize_t mode_show(struct device *dev, struct device_attribute *attr,
194 			 char *buf)
195 {
196 	struct cxl_endpoint_decoder *cxled = to_cxl_endpoint_decoder(dev);
197 	struct cxl_memdev *cxlmd = cxled_to_memdev(cxled);
198 	struct cxl_dev_state *cxlds = cxlmd->cxlds;
199 	/* without @cxl_dpa_rwsem, make sure @part is not reloaded */
200 	int part = READ_ONCE(cxled->part);
201 	const char *desc;
202 
203 	if (part < 0)
204 		desc = "none";
205 	else
206 		desc = cxlds->part[part].res.name;
207 
208 	return sysfs_emit(buf, "%s\n", desc);
209 }
210 
mode_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t len)211 static ssize_t mode_store(struct device *dev, struct device_attribute *attr,
212 			  const char *buf, size_t len)
213 {
214 	struct cxl_endpoint_decoder *cxled = to_cxl_endpoint_decoder(dev);
215 	enum cxl_partition_mode mode;
216 	ssize_t rc;
217 
218 	if (sysfs_streq(buf, "pmem"))
219 		mode = CXL_PARTMODE_PMEM;
220 	else if (sysfs_streq(buf, "ram"))
221 		mode = CXL_PARTMODE_RAM;
222 	else
223 		return -EINVAL;
224 
225 	rc = cxl_dpa_set_part(cxled, mode);
226 	if (rc)
227 		return rc;
228 
229 	return len;
230 }
231 static DEVICE_ATTR_RW(mode);
232 
dpa_resource_show(struct device * dev,struct device_attribute * attr,char * buf)233 static ssize_t dpa_resource_show(struct device *dev, struct device_attribute *attr,
234 			    char *buf)
235 {
236 	struct cxl_endpoint_decoder *cxled = to_cxl_endpoint_decoder(dev);
237 
238 	guard(rwsem_read)(&cxl_dpa_rwsem);
239 	return sysfs_emit(buf, "%#llx\n", (u64)cxl_dpa_resource_start(cxled));
240 }
241 static DEVICE_ATTR_RO(dpa_resource);
242 
dpa_size_show(struct device * dev,struct device_attribute * attr,char * buf)243 static ssize_t dpa_size_show(struct device *dev, struct device_attribute *attr,
244 			     char *buf)
245 {
246 	struct cxl_endpoint_decoder *cxled = to_cxl_endpoint_decoder(dev);
247 	resource_size_t size = cxl_dpa_size(cxled);
248 
249 	return sysfs_emit(buf, "%pa\n", &size);
250 }
251 
dpa_size_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t len)252 static ssize_t dpa_size_store(struct device *dev, struct device_attribute *attr,
253 			      const char *buf, size_t len)
254 {
255 	struct cxl_endpoint_decoder *cxled = to_cxl_endpoint_decoder(dev);
256 	unsigned long long size;
257 	ssize_t rc;
258 
259 	rc = kstrtoull(buf, 0, &size);
260 	if (rc)
261 		return rc;
262 
263 	if (!IS_ALIGNED(size, SZ_256M))
264 		return -EINVAL;
265 
266 	rc = cxl_dpa_free(cxled);
267 	if (rc)
268 		return rc;
269 
270 	if (size == 0)
271 		return len;
272 
273 	rc = cxl_dpa_alloc(cxled, size);
274 	if (rc)
275 		return rc;
276 
277 	return len;
278 }
279 static DEVICE_ATTR_RW(dpa_size);
280 
interleave_granularity_show(struct device * dev,struct device_attribute * attr,char * buf)281 static ssize_t interleave_granularity_show(struct device *dev,
282 					   struct device_attribute *attr,
283 					   char *buf)
284 {
285 	struct cxl_decoder *cxld = to_cxl_decoder(dev);
286 
287 	return sysfs_emit(buf, "%d\n", cxld->interleave_granularity);
288 }
289 
290 static DEVICE_ATTR_RO(interleave_granularity);
291 
interleave_ways_show(struct device * dev,struct device_attribute * attr,char * buf)292 static ssize_t interleave_ways_show(struct device *dev,
293 				    struct device_attribute *attr, char *buf)
294 {
295 	struct cxl_decoder *cxld = to_cxl_decoder(dev);
296 
297 	return sysfs_emit(buf, "%d\n", cxld->interleave_ways);
298 }
299 
300 static DEVICE_ATTR_RO(interleave_ways);
301 
qos_class_show(struct device * dev,struct device_attribute * attr,char * buf)302 static ssize_t qos_class_show(struct device *dev,
303 			      struct device_attribute *attr, char *buf)
304 {
305 	struct cxl_root_decoder *cxlrd = to_cxl_root_decoder(dev);
306 
307 	return sysfs_emit(buf, "%d\n", cxlrd->qos_class);
308 }
309 static DEVICE_ATTR_RO(qos_class);
310 
311 static struct attribute *cxl_decoder_base_attrs[] = {
312 	&dev_attr_start.attr,
313 	&dev_attr_size.attr,
314 	&dev_attr_locked.attr,
315 	&dev_attr_interleave_granularity.attr,
316 	&dev_attr_interleave_ways.attr,
317 	NULL,
318 };
319 
320 static struct attribute_group cxl_decoder_base_attribute_group = {
321 	.attrs = cxl_decoder_base_attrs,
322 };
323 
324 static struct attribute *cxl_decoder_root_attrs[] = {
325 	&dev_attr_cap_pmem.attr,
326 	&dev_attr_cap_ram.attr,
327 	&dev_attr_cap_type2.attr,
328 	&dev_attr_cap_type3.attr,
329 	&dev_attr_target_list.attr,
330 	&dev_attr_qos_class.attr,
331 	SET_CXL_REGION_ATTR(create_pmem_region)
332 	SET_CXL_REGION_ATTR(create_ram_region)
333 	SET_CXL_REGION_ATTR(delete_region)
334 	NULL,
335 };
336 
can_create_pmem(struct cxl_root_decoder * cxlrd)337 static bool can_create_pmem(struct cxl_root_decoder *cxlrd)
338 {
339 	unsigned long flags = CXL_DECODER_F_TYPE3 | CXL_DECODER_F_PMEM;
340 
341 	return (cxlrd->cxlsd.cxld.flags & flags) == flags;
342 }
343 
can_create_ram(struct cxl_root_decoder * cxlrd)344 static bool can_create_ram(struct cxl_root_decoder *cxlrd)
345 {
346 	unsigned long flags = CXL_DECODER_F_TYPE3 | CXL_DECODER_F_RAM;
347 
348 	return (cxlrd->cxlsd.cxld.flags & flags) == flags;
349 }
350 
cxl_root_decoder_visible(struct kobject * kobj,struct attribute * a,int n)351 static umode_t cxl_root_decoder_visible(struct kobject *kobj, struct attribute *a, int n)
352 {
353 	struct device *dev = kobj_to_dev(kobj);
354 	struct cxl_root_decoder *cxlrd = to_cxl_root_decoder(dev);
355 
356 	if (a == CXL_REGION_ATTR(create_pmem_region) && !can_create_pmem(cxlrd))
357 		return 0;
358 
359 	if (a == CXL_REGION_ATTR(create_ram_region) && !can_create_ram(cxlrd))
360 		return 0;
361 
362 	if (a == CXL_REGION_ATTR(delete_region) &&
363 	    !(can_create_pmem(cxlrd) || can_create_ram(cxlrd)))
364 		return 0;
365 
366 	return a->mode;
367 }
368 
369 static struct attribute_group cxl_decoder_root_attribute_group = {
370 	.attrs = cxl_decoder_root_attrs,
371 	.is_visible = cxl_root_decoder_visible,
372 };
373 
374 static const struct attribute_group *cxl_decoder_root_attribute_groups[] = {
375 	&cxl_decoder_root_attribute_group,
376 	&cxl_decoder_base_attribute_group,
377 	&cxl_base_attribute_group,
378 	NULL,
379 };
380 
381 static struct attribute *cxl_decoder_switch_attrs[] = {
382 	&dev_attr_target_type.attr,
383 	&dev_attr_target_list.attr,
384 	SET_CXL_REGION_ATTR(region)
385 	NULL,
386 };
387 
388 static struct attribute_group cxl_decoder_switch_attribute_group = {
389 	.attrs = cxl_decoder_switch_attrs,
390 };
391 
392 static const struct attribute_group *cxl_decoder_switch_attribute_groups[] = {
393 	&cxl_decoder_switch_attribute_group,
394 	&cxl_decoder_base_attribute_group,
395 	&cxl_base_attribute_group,
396 	NULL,
397 };
398 
399 static struct attribute *cxl_decoder_endpoint_attrs[] = {
400 	&dev_attr_target_type.attr,
401 	&dev_attr_mode.attr,
402 	&dev_attr_dpa_size.attr,
403 	&dev_attr_dpa_resource.attr,
404 	SET_CXL_REGION_ATTR(region)
405 	NULL,
406 };
407 
408 static struct attribute_group cxl_decoder_endpoint_attribute_group = {
409 	.attrs = cxl_decoder_endpoint_attrs,
410 };
411 
412 static const struct attribute_group *cxl_decoder_endpoint_attribute_groups[] = {
413 	&cxl_decoder_base_attribute_group,
414 	&cxl_decoder_endpoint_attribute_group,
415 	&cxl_base_attribute_group,
416 	NULL,
417 };
418 
__cxl_decoder_release(struct cxl_decoder * cxld)419 static void __cxl_decoder_release(struct cxl_decoder *cxld)
420 {
421 	struct cxl_port *port = to_cxl_port(cxld->dev.parent);
422 
423 	ida_free(&port->decoder_ida, cxld->id);
424 	put_device(&port->dev);
425 }
426 
cxl_endpoint_decoder_release(struct device * dev)427 static void cxl_endpoint_decoder_release(struct device *dev)
428 {
429 	struct cxl_endpoint_decoder *cxled = to_cxl_endpoint_decoder(dev);
430 
431 	__cxl_decoder_release(&cxled->cxld);
432 	kfree(cxled);
433 }
434 
cxl_switch_decoder_release(struct device * dev)435 static void cxl_switch_decoder_release(struct device *dev)
436 {
437 	struct cxl_switch_decoder *cxlsd = to_cxl_switch_decoder(dev);
438 
439 	__cxl_decoder_release(&cxlsd->cxld);
440 	kfree(cxlsd);
441 }
442 
to_cxl_root_decoder(struct device * dev)443 struct cxl_root_decoder *to_cxl_root_decoder(struct device *dev)
444 {
445 	if (dev_WARN_ONCE(dev, !is_root_decoder(dev),
446 			  "not a cxl_root_decoder device\n"))
447 		return NULL;
448 	return container_of(dev, struct cxl_root_decoder, cxlsd.cxld.dev);
449 }
450 EXPORT_SYMBOL_NS_GPL(to_cxl_root_decoder, "CXL");
451 
cxl_root_decoder_release(struct device * dev)452 static void cxl_root_decoder_release(struct device *dev)
453 {
454 	struct cxl_root_decoder *cxlrd = to_cxl_root_decoder(dev);
455 
456 	if (atomic_read(&cxlrd->region_id) >= 0)
457 		memregion_free(atomic_read(&cxlrd->region_id));
458 	__cxl_decoder_release(&cxlrd->cxlsd.cxld);
459 	kfree(cxlrd);
460 }
461 
462 static const struct device_type cxl_decoder_endpoint_type = {
463 	.name = "cxl_decoder_endpoint",
464 	.release = cxl_endpoint_decoder_release,
465 	.groups = cxl_decoder_endpoint_attribute_groups,
466 };
467 
468 static const struct device_type cxl_decoder_switch_type = {
469 	.name = "cxl_decoder_switch",
470 	.release = cxl_switch_decoder_release,
471 	.groups = cxl_decoder_switch_attribute_groups,
472 };
473 
474 static const struct device_type cxl_decoder_root_type = {
475 	.name = "cxl_decoder_root",
476 	.release = cxl_root_decoder_release,
477 	.groups = cxl_decoder_root_attribute_groups,
478 };
479 
is_endpoint_decoder(struct device * dev)480 bool is_endpoint_decoder(struct device *dev)
481 {
482 	return dev->type == &cxl_decoder_endpoint_type;
483 }
484 EXPORT_SYMBOL_NS_GPL(is_endpoint_decoder, "CXL");
485 
is_root_decoder(struct device * dev)486 bool is_root_decoder(struct device *dev)
487 {
488 	return dev->type == &cxl_decoder_root_type;
489 }
490 EXPORT_SYMBOL_NS_GPL(is_root_decoder, "CXL");
491 
is_switch_decoder(struct device * dev)492 bool is_switch_decoder(struct device *dev)
493 {
494 	return is_root_decoder(dev) || dev->type == &cxl_decoder_switch_type;
495 }
496 EXPORT_SYMBOL_NS_GPL(is_switch_decoder, "CXL");
497 
to_cxl_decoder(struct device * dev)498 struct cxl_decoder *to_cxl_decoder(struct device *dev)
499 {
500 	if (dev_WARN_ONCE(dev,
501 			  !is_switch_decoder(dev) && !is_endpoint_decoder(dev),
502 			  "not a cxl_decoder device\n"))
503 		return NULL;
504 	return container_of(dev, struct cxl_decoder, dev);
505 }
506 EXPORT_SYMBOL_NS_GPL(to_cxl_decoder, "CXL");
507 
to_cxl_endpoint_decoder(struct device * dev)508 struct cxl_endpoint_decoder *to_cxl_endpoint_decoder(struct device *dev)
509 {
510 	if (dev_WARN_ONCE(dev, !is_endpoint_decoder(dev),
511 			  "not a cxl_endpoint_decoder device\n"))
512 		return NULL;
513 	return container_of(dev, struct cxl_endpoint_decoder, cxld.dev);
514 }
515 EXPORT_SYMBOL_NS_GPL(to_cxl_endpoint_decoder, "CXL");
516 
to_cxl_switch_decoder(struct device * dev)517 struct cxl_switch_decoder *to_cxl_switch_decoder(struct device *dev)
518 {
519 	if (dev_WARN_ONCE(dev, !is_switch_decoder(dev),
520 			  "not a cxl_switch_decoder device\n"))
521 		return NULL;
522 	return container_of(dev, struct cxl_switch_decoder, cxld.dev);
523 }
524 EXPORT_SYMBOL_NS_GPL(to_cxl_switch_decoder, "CXL");
525 
cxl_ep_release(struct cxl_ep * ep)526 static void cxl_ep_release(struct cxl_ep *ep)
527 {
528 	put_device(ep->ep);
529 	kfree(ep);
530 }
531 
cxl_ep_remove(struct cxl_port * port,struct cxl_ep * ep)532 static void cxl_ep_remove(struct cxl_port *port, struct cxl_ep *ep)
533 {
534 	if (!ep)
535 		return;
536 	xa_erase(&port->endpoints, (unsigned long) ep->ep);
537 	cxl_ep_release(ep);
538 }
539 
cxl_port_release(struct device * dev)540 static void cxl_port_release(struct device *dev)
541 {
542 	struct cxl_port *port = to_cxl_port(dev);
543 	unsigned long index;
544 	struct cxl_ep *ep;
545 
546 	xa_for_each(&port->endpoints, index, ep)
547 		cxl_ep_remove(port, ep);
548 	xa_destroy(&port->endpoints);
549 	xa_destroy(&port->dports);
550 	xa_destroy(&port->regions);
551 	ida_free(&cxl_port_ida, port->id);
552 	if (is_cxl_root(port))
553 		kfree(to_cxl_root(port));
554 	else
555 		kfree(port);
556 }
557 
decoders_committed_show(struct device * dev,struct device_attribute * attr,char * buf)558 static ssize_t decoders_committed_show(struct device *dev,
559 				       struct device_attribute *attr, char *buf)
560 {
561 	struct cxl_port *port = to_cxl_port(dev);
562 
563 	guard(rwsem_read)(&cxl_region_rwsem);
564 	return sysfs_emit(buf, "%d\n", cxl_num_decoders_committed(port));
565 }
566 
567 static DEVICE_ATTR_RO(decoders_committed);
568 
569 static struct attribute *cxl_port_attrs[] = {
570 	&dev_attr_decoders_committed.attr,
571 	NULL,
572 };
573 
574 static struct attribute_group cxl_port_attribute_group = {
575 	.attrs = cxl_port_attrs,
576 };
577 
578 static const struct attribute_group *cxl_port_attribute_groups[] = {
579 	&cxl_base_attribute_group,
580 	&cxl_port_attribute_group,
581 	NULL,
582 };
583 
584 static const struct device_type cxl_port_type = {
585 	.name = "cxl_port",
586 	.release = cxl_port_release,
587 	.groups = cxl_port_attribute_groups,
588 };
589 
is_cxl_port(const struct device * dev)590 bool is_cxl_port(const struct device *dev)
591 {
592 	return dev->type == &cxl_port_type;
593 }
594 EXPORT_SYMBOL_NS_GPL(is_cxl_port, "CXL");
595 
to_cxl_port(const struct device * dev)596 struct cxl_port *to_cxl_port(const struct device *dev)
597 {
598 	if (dev_WARN_ONCE(dev, dev->type != &cxl_port_type,
599 			  "not a cxl_port device\n"))
600 		return NULL;
601 	return container_of(dev, struct cxl_port, dev);
602 }
603 EXPORT_SYMBOL_NS_GPL(to_cxl_port, "CXL");
604 
unregister_port(void * _port)605 static void unregister_port(void *_port)
606 {
607 	struct cxl_port *port = _port;
608 	struct cxl_port *parent;
609 	struct device *lock_dev;
610 
611 	if (is_cxl_root(port))
612 		parent = NULL;
613 	else
614 		parent = to_cxl_port(port->dev.parent);
615 
616 	/*
617 	 * CXL root port's and the first level of ports are unregistered
618 	 * under the platform firmware device lock, all other ports are
619 	 * unregistered while holding their parent port lock.
620 	 */
621 	if (!parent)
622 		lock_dev = port->uport_dev;
623 	else if (is_cxl_root(parent))
624 		lock_dev = parent->uport_dev;
625 	else
626 		lock_dev = &parent->dev;
627 
628 	device_lock_assert(lock_dev);
629 	port->dead = true;
630 	device_unregister(&port->dev);
631 }
632 
cxl_unlink_uport(void * _port)633 static void cxl_unlink_uport(void *_port)
634 {
635 	struct cxl_port *port = _port;
636 
637 	sysfs_remove_link(&port->dev.kobj, "uport");
638 }
639 
devm_cxl_link_uport(struct device * host,struct cxl_port * port)640 static int devm_cxl_link_uport(struct device *host, struct cxl_port *port)
641 {
642 	int rc;
643 
644 	rc = sysfs_create_link(&port->dev.kobj, &port->uport_dev->kobj,
645 			       "uport");
646 	if (rc)
647 		return rc;
648 	return devm_add_action_or_reset(host, cxl_unlink_uport, port);
649 }
650 
cxl_unlink_parent_dport(void * _port)651 static void cxl_unlink_parent_dport(void *_port)
652 {
653 	struct cxl_port *port = _port;
654 
655 	sysfs_remove_link(&port->dev.kobj, "parent_dport");
656 }
657 
devm_cxl_link_parent_dport(struct device * host,struct cxl_port * port,struct cxl_dport * parent_dport)658 static int devm_cxl_link_parent_dport(struct device *host,
659 				      struct cxl_port *port,
660 				      struct cxl_dport *parent_dport)
661 {
662 	int rc;
663 
664 	if (!parent_dport)
665 		return 0;
666 
667 	rc = sysfs_create_link(&port->dev.kobj, &parent_dport->dport_dev->kobj,
668 			       "parent_dport");
669 	if (rc)
670 		return rc;
671 	return devm_add_action_or_reset(host, cxl_unlink_parent_dport, port);
672 }
673 
674 static struct lock_class_key cxl_port_key;
675 
cxl_port_alloc(struct device * uport_dev,struct cxl_dport * parent_dport)676 static struct cxl_port *cxl_port_alloc(struct device *uport_dev,
677 				       struct cxl_dport *parent_dport)
678 {
679 	struct cxl_root *cxl_root __free(kfree) = NULL;
680 	struct cxl_port *port, *_port __free(kfree) = NULL;
681 	struct device *dev;
682 	int rc;
683 
684 	/* No parent_dport, root cxl_port */
685 	if (!parent_dport) {
686 		cxl_root = kzalloc(sizeof(*cxl_root), GFP_KERNEL);
687 		if (!cxl_root)
688 			return ERR_PTR(-ENOMEM);
689 	} else {
690 		_port = kzalloc(sizeof(*port), GFP_KERNEL);
691 		if (!_port)
692 			return ERR_PTR(-ENOMEM);
693 	}
694 
695 	rc = ida_alloc(&cxl_port_ida, GFP_KERNEL);
696 	if (rc < 0)
697 		return ERR_PTR(rc);
698 
699 	if (cxl_root)
700 		port = &no_free_ptr(cxl_root)->port;
701 	else
702 		port = no_free_ptr(_port);
703 
704 	port->id = rc;
705 	port->uport_dev = uport_dev;
706 
707 	/*
708 	 * The top-level cxl_port "cxl_root" does not have a cxl_port as
709 	 * its parent and it does not have any corresponding component
710 	 * registers as its decode is described by a fixed platform
711 	 * description.
712 	 */
713 	dev = &port->dev;
714 	if (parent_dport) {
715 		struct cxl_port *parent_port = parent_dport->port;
716 		struct cxl_port *iter;
717 
718 		dev->parent = &parent_port->dev;
719 		port->depth = parent_port->depth + 1;
720 		port->parent_dport = parent_dport;
721 
722 		/*
723 		 * walk to the host bridge, or the first ancestor that knows
724 		 * the host bridge
725 		 */
726 		iter = port;
727 		while (!iter->host_bridge &&
728 		       !is_cxl_root(to_cxl_port(iter->dev.parent)))
729 			iter = to_cxl_port(iter->dev.parent);
730 		if (iter->host_bridge)
731 			port->host_bridge = iter->host_bridge;
732 		else if (parent_dport->rch)
733 			port->host_bridge = parent_dport->dport_dev;
734 		else
735 			port->host_bridge = iter->uport_dev;
736 		dev_dbg(uport_dev, "host-bridge: %s\n",
737 			dev_name(port->host_bridge));
738 	} else
739 		dev->parent = uport_dev;
740 
741 	ida_init(&port->decoder_ida);
742 	port->hdm_end = -1;
743 	port->commit_end = -1;
744 	xa_init(&port->dports);
745 	xa_init(&port->endpoints);
746 	xa_init(&port->regions);
747 
748 	device_initialize(dev);
749 	lockdep_set_class_and_subclass(&dev->mutex, &cxl_port_key, port->depth);
750 	device_set_pm_not_required(dev);
751 	dev->bus = &cxl_bus_type;
752 	dev->type = &cxl_port_type;
753 
754 	return port;
755 }
756 
cxl_setup_comp_regs(struct device * host,struct cxl_register_map * map,resource_size_t component_reg_phys)757 static int cxl_setup_comp_regs(struct device *host, struct cxl_register_map *map,
758 			       resource_size_t component_reg_phys)
759 {
760 	*map = (struct cxl_register_map) {
761 		.host = host,
762 		.reg_type = CXL_REGLOC_RBI_EMPTY,
763 		.resource = component_reg_phys,
764 	};
765 
766 	if (component_reg_phys == CXL_RESOURCE_NONE)
767 		return 0;
768 
769 	map->reg_type = CXL_REGLOC_RBI_COMPONENT;
770 	map->max_size = CXL_COMPONENT_REG_BLOCK_SIZE;
771 
772 	return cxl_setup_regs(map);
773 }
774 
cxl_port_setup_regs(struct cxl_port * port,resource_size_t component_reg_phys)775 static int cxl_port_setup_regs(struct cxl_port *port,
776 			resource_size_t component_reg_phys)
777 {
778 	if (dev_is_platform(port->uport_dev))
779 		return 0;
780 	return cxl_setup_comp_regs(&port->dev, &port->reg_map,
781 				   component_reg_phys);
782 }
783 
cxl_dport_setup_regs(struct device * host,struct cxl_dport * dport,resource_size_t component_reg_phys)784 static int cxl_dport_setup_regs(struct device *host, struct cxl_dport *dport,
785 				resource_size_t component_reg_phys)
786 {
787 	int rc;
788 
789 	if (dev_is_platform(dport->dport_dev))
790 		return 0;
791 
792 	/*
793 	 * use @dport->dport_dev for the context for error messages during
794 	 * register probing, and fixup @host after the fact, since @host may be
795 	 * NULL.
796 	 */
797 	rc = cxl_setup_comp_regs(dport->dport_dev, &dport->reg_map,
798 				 component_reg_phys);
799 	dport->reg_map.host = host;
800 	return rc;
801 }
802 
803 DEFINE_SHOW_ATTRIBUTE(einj_cxl_available_error_type);
804 
cxl_einj_inject(void * data,u64 type)805 static int cxl_einj_inject(void *data, u64 type)
806 {
807 	struct cxl_dport *dport = data;
808 
809 	if (dport->rch)
810 		return einj_cxl_inject_rch_error(dport->rcrb.base, type);
811 
812 	return einj_cxl_inject_error(to_pci_dev(dport->dport_dev), type);
813 }
814 DEFINE_DEBUGFS_ATTRIBUTE(cxl_einj_inject_fops, NULL, cxl_einj_inject,
815 			 "0x%llx\n");
816 
cxl_debugfs_create_dport_dir(struct cxl_dport * dport)817 static void cxl_debugfs_create_dport_dir(struct cxl_dport *dport)
818 {
819 	struct dentry *dir;
820 
821 	if (!einj_cxl_is_initialized())
822 		return;
823 
824 	/*
825 	 * dport_dev needs to be a PCIe port for CXL 2.0+ ports because
826 	 * EINJ expects a dport SBDF to be specified for 2.0 error injection.
827 	 */
828 	if (!dport->rch && !dev_is_pci(dport->dport_dev))
829 		return;
830 
831 	dir = cxl_debugfs_create_dir(dev_name(dport->dport_dev));
832 
833 	debugfs_create_file("einj_inject", 0200, dir, dport,
834 			    &cxl_einj_inject_fops);
835 }
836 
cxl_port_add(struct cxl_port * port,resource_size_t component_reg_phys,struct cxl_dport * parent_dport)837 static int cxl_port_add(struct cxl_port *port,
838 			resource_size_t component_reg_phys,
839 			struct cxl_dport *parent_dport)
840 {
841 	struct device *dev __free(put_device) = &port->dev;
842 	int rc;
843 
844 	if (is_cxl_memdev(port->uport_dev)) {
845 		struct cxl_memdev *cxlmd = to_cxl_memdev(port->uport_dev);
846 		struct cxl_dev_state *cxlds = cxlmd->cxlds;
847 
848 		rc = dev_set_name(dev, "endpoint%d", port->id);
849 		if (rc)
850 			return rc;
851 
852 		/*
853 		 * The endpoint driver already enumerated the component and RAS
854 		 * registers. Reuse that enumeration while prepping them to be
855 		 * mapped by the cxl_port driver.
856 		 */
857 		port->reg_map = cxlds->reg_map;
858 		port->reg_map.host = &port->dev;
859 		cxlmd->endpoint = port;
860 	} else if (parent_dport) {
861 		rc = dev_set_name(dev, "port%d", port->id);
862 		if (rc)
863 			return rc;
864 
865 		rc = cxl_port_setup_regs(port, component_reg_phys);
866 		if (rc)
867 			return rc;
868 	} else {
869 		rc = dev_set_name(dev, "root%d", port->id);
870 		if (rc)
871 			return rc;
872 	}
873 
874 	rc = device_add(dev);
875 	if (rc)
876 		return rc;
877 
878 	/* Inhibit the cleanup function invoked */
879 	dev = NULL;
880 	return 0;
881 }
882 
__devm_cxl_add_port(struct device * host,struct device * uport_dev,resource_size_t component_reg_phys,struct cxl_dport * parent_dport)883 static struct cxl_port *__devm_cxl_add_port(struct device *host,
884 					    struct device *uport_dev,
885 					    resource_size_t component_reg_phys,
886 					    struct cxl_dport *parent_dport)
887 {
888 	struct cxl_port *port;
889 	int rc;
890 
891 	port = cxl_port_alloc(uport_dev, parent_dport);
892 	if (IS_ERR(port))
893 		return port;
894 
895 	rc = cxl_port_add(port, component_reg_phys, parent_dport);
896 	if (rc)
897 		return ERR_PTR(rc);
898 
899 	rc = devm_add_action_or_reset(host, unregister_port, port);
900 	if (rc)
901 		return ERR_PTR(rc);
902 
903 	rc = devm_cxl_link_uport(host, port);
904 	if (rc)
905 		return ERR_PTR(rc);
906 
907 	rc = devm_cxl_link_parent_dport(host, port, parent_dport);
908 	if (rc)
909 		return ERR_PTR(rc);
910 
911 	if (parent_dport && dev_is_pci(uport_dev))
912 		port->pci_latency = cxl_pci_get_latency(to_pci_dev(uport_dev));
913 
914 	return port;
915 }
916 
917 /**
918  * devm_cxl_add_port - register a cxl_port in CXL memory decode hierarchy
919  * @host: host device for devm operations
920  * @uport_dev: "physical" device implementing this upstream port
921  * @component_reg_phys: (optional) for configurable cxl_port instances
922  * @parent_dport: next hop up in the CXL memory decode hierarchy
923  */
devm_cxl_add_port(struct device * host,struct device * uport_dev,resource_size_t component_reg_phys,struct cxl_dport * parent_dport)924 struct cxl_port *devm_cxl_add_port(struct device *host,
925 				   struct device *uport_dev,
926 				   resource_size_t component_reg_phys,
927 				   struct cxl_dport *parent_dport)
928 {
929 	struct cxl_port *port, *parent_port;
930 
931 	port = __devm_cxl_add_port(host, uport_dev, component_reg_phys,
932 				   parent_dport);
933 
934 	parent_port = parent_dport ? parent_dport->port : NULL;
935 	if (IS_ERR(port)) {
936 		dev_dbg(uport_dev, "Failed to add%s%s%s: %ld\n",
937 			parent_port ? " port to " : "",
938 			parent_port ? dev_name(&parent_port->dev) : "",
939 			parent_port ? "" : " root port",
940 			PTR_ERR(port));
941 	} else {
942 		dev_dbg(uport_dev, "%s added%s%s%s\n",
943 			dev_name(&port->dev),
944 			parent_port ? " to " : "",
945 			parent_port ? dev_name(&parent_port->dev) : "",
946 			parent_port ? "" : " (root port)");
947 	}
948 
949 	return port;
950 }
951 EXPORT_SYMBOL_NS_GPL(devm_cxl_add_port, "CXL");
952 
devm_cxl_add_root(struct device * host,const struct cxl_root_ops * ops)953 struct cxl_root *devm_cxl_add_root(struct device *host,
954 				   const struct cxl_root_ops *ops)
955 {
956 	struct cxl_root *cxl_root;
957 	struct cxl_port *port;
958 
959 	port = devm_cxl_add_port(host, host, CXL_RESOURCE_NONE, NULL);
960 	if (IS_ERR(port))
961 		return ERR_CAST(port);
962 
963 	cxl_root = to_cxl_root(port);
964 	cxl_root->ops = ops;
965 	return cxl_root;
966 }
967 EXPORT_SYMBOL_NS_GPL(devm_cxl_add_root, "CXL");
968 
cxl_port_to_pci_bus(struct cxl_port * port)969 struct pci_bus *cxl_port_to_pci_bus(struct cxl_port *port)
970 {
971 	/* There is no pci_bus associated with a CXL platform-root port */
972 	if (is_cxl_root(port))
973 		return NULL;
974 
975 	if (dev_is_pci(port->uport_dev)) {
976 		struct pci_dev *pdev = to_pci_dev(port->uport_dev);
977 
978 		return pdev->subordinate;
979 	}
980 
981 	return xa_load(&cxl_root_buses, (unsigned long)port->uport_dev);
982 }
983 EXPORT_SYMBOL_NS_GPL(cxl_port_to_pci_bus, "CXL");
984 
unregister_pci_bus(void * uport_dev)985 static void unregister_pci_bus(void *uport_dev)
986 {
987 	xa_erase(&cxl_root_buses, (unsigned long)uport_dev);
988 }
989 
devm_cxl_register_pci_bus(struct device * host,struct device * uport_dev,struct pci_bus * bus)990 int devm_cxl_register_pci_bus(struct device *host, struct device *uport_dev,
991 			      struct pci_bus *bus)
992 {
993 	int rc;
994 
995 	if (dev_is_pci(uport_dev))
996 		return -EINVAL;
997 
998 	rc = xa_insert(&cxl_root_buses, (unsigned long)uport_dev, bus,
999 		       GFP_KERNEL);
1000 	if (rc)
1001 		return rc;
1002 	return devm_add_action_or_reset(host, unregister_pci_bus, uport_dev);
1003 }
1004 EXPORT_SYMBOL_NS_GPL(devm_cxl_register_pci_bus, "CXL");
1005 
dev_is_cxl_root_child(struct device * dev)1006 static bool dev_is_cxl_root_child(struct device *dev)
1007 {
1008 	struct cxl_port *port, *parent;
1009 
1010 	if (!is_cxl_port(dev))
1011 		return false;
1012 
1013 	port = to_cxl_port(dev);
1014 	if (is_cxl_root(port))
1015 		return false;
1016 
1017 	parent = to_cxl_port(port->dev.parent);
1018 	if (is_cxl_root(parent))
1019 		return true;
1020 
1021 	return false;
1022 }
1023 
find_cxl_root(struct cxl_port * port)1024 struct cxl_root *find_cxl_root(struct cxl_port *port)
1025 {
1026 	struct cxl_port *iter = port;
1027 
1028 	while (iter && !is_cxl_root(iter))
1029 		iter = to_cxl_port(iter->dev.parent);
1030 
1031 	if (!iter)
1032 		return NULL;
1033 	get_device(&iter->dev);
1034 	return to_cxl_root(iter);
1035 }
1036 EXPORT_SYMBOL_NS_GPL(find_cxl_root, "CXL");
1037 
put_cxl_root(struct cxl_root * cxl_root)1038 void put_cxl_root(struct cxl_root *cxl_root)
1039 {
1040 	if (!cxl_root)
1041 		return;
1042 
1043 	put_device(&cxl_root->port.dev);
1044 }
1045 EXPORT_SYMBOL_NS_GPL(put_cxl_root, "CXL");
1046 
find_dport(struct cxl_port * port,int id)1047 static struct cxl_dport *find_dport(struct cxl_port *port, int id)
1048 {
1049 	struct cxl_dport *dport;
1050 	unsigned long index;
1051 
1052 	device_lock_assert(&port->dev);
1053 	xa_for_each(&port->dports, index, dport)
1054 		if (dport->port_id == id)
1055 			return dport;
1056 	return NULL;
1057 }
1058 
add_dport(struct cxl_port * port,struct cxl_dport * dport)1059 static int add_dport(struct cxl_port *port, struct cxl_dport *dport)
1060 {
1061 	struct cxl_dport *dup;
1062 	int rc;
1063 
1064 	device_lock_assert(&port->dev);
1065 	dup = find_dport(port, dport->port_id);
1066 	if (dup) {
1067 		dev_err(&port->dev,
1068 			"unable to add dport%d-%s non-unique port id (%s)\n",
1069 			dport->port_id, dev_name(dport->dport_dev),
1070 			dev_name(dup->dport_dev));
1071 		return -EBUSY;
1072 	}
1073 
1074 	rc = xa_insert(&port->dports, (unsigned long)dport->dport_dev, dport,
1075 		       GFP_KERNEL);
1076 	if (rc)
1077 		return rc;
1078 
1079 	port->nr_dports++;
1080 	return 0;
1081 }
1082 
1083 /*
1084  * Since root-level CXL dports cannot be enumerated by PCI they are not
1085  * enumerated by the common port driver that acquires the port lock over
1086  * dport add/remove. Instead, root dports are manually added by a
1087  * platform driver and cond_cxl_root_lock() is used to take the missing
1088  * port lock in that case.
1089  */
cond_cxl_root_lock(struct cxl_port * port)1090 static void cond_cxl_root_lock(struct cxl_port *port)
1091 {
1092 	if (is_cxl_root(port))
1093 		device_lock(&port->dev);
1094 }
1095 
cond_cxl_root_unlock(struct cxl_port * port)1096 static void cond_cxl_root_unlock(struct cxl_port *port)
1097 {
1098 	if (is_cxl_root(port))
1099 		device_unlock(&port->dev);
1100 }
1101 
cxl_dport_remove(void * data)1102 static void cxl_dport_remove(void *data)
1103 {
1104 	struct cxl_dport *dport = data;
1105 	struct cxl_port *port = dport->port;
1106 
1107 	xa_erase(&port->dports, (unsigned long) dport->dport_dev);
1108 	put_device(dport->dport_dev);
1109 }
1110 
cxl_dport_unlink(void * data)1111 static void cxl_dport_unlink(void *data)
1112 {
1113 	struct cxl_dport *dport = data;
1114 	struct cxl_port *port = dport->port;
1115 	char link_name[CXL_TARGET_STRLEN];
1116 
1117 	sprintf(link_name, "dport%d", dport->port_id);
1118 	sysfs_remove_link(&port->dev.kobj, link_name);
1119 }
1120 
1121 static struct cxl_dport *
__devm_cxl_add_dport(struct cxl_port * port,struct device * dport_dev,int port_id,resource_size_t component_reg_phys,resource_size_t rcrb)1122 __devm_cxl_add_dport(struct cxl_port *port, struct device *dport_dev,
1123 		     int port_id, resource_size_t component_reg_phys,
1124 		     resource_size_t rcrb)
1125 {
1126 	char link_name[CXL_TARGET_STRLEN];
1127 	struct cxl_dport *dport;
1128 	struct device *host;
1129 	int rc;
1130 
1131 	if (is_cxl_root(port))
1132 		host = port->uport_dev;
1133 	else
1134 		host = &port->dev;
1135 
1136 	if (!host->driver) {
1137 		dev_WARN_ONCE(&port->dev, 1, "dport:%s bad devm context\n",
1138 			      dev_name(dport_dev));
1139 		return ERR_PTR(-ENXIO);
1140 	}
1141 
1142 	if (snprintf(link_name, CXL_TARGET_STRLEN, "dport%d", port_id) >=
1143 	    CXL_TARGET_STRLEN)
1144 		return ERR_PTR(-EINVAL);
1145 
1146 	dport = devm_kzalloc(host, sizeof(*dport), GFP_KERNEL);
1147 	if (!dport)
1148 		return ERR_PTR(-ENOMEM);
1149 
1150 	dport->dport_dev = dport_dev;
1151 	dport->port_id = port_id;
1152 	dport->port = port;
1153 
1154 	if (rcrb == CXL_RESOURCE_NONE) {
1155 		rc = cxl_dport_setup_regs(&port->dev, dport,
1156 					  component_reg_phys);
1157 		if (rc)
1158 			return ERR_PTR(rc);
1159 	} else {
1160 		dport->rcrb.base = rcrb;
1161 		component_reg_phys = __rcrb_to_component(dport_dev, &dport->rcrb,
1162 							 CXL_RCRB_DOWNSTREAM);
1163 		if (component_reg_phys == CXL_RESOURCE_NONE) {
1164 			dev_warn(dport_dev, "Invalid Component Registers in RCRB");
1165 			return ERR_PTR(-ENXIO);
1166 		}
1167 
1168 		/*
1169 		 * RCH @dport is not ready to map until associated with its
1170 		 * memdev
1171 		 */
1172 		rc = cxl_dport_setup_regs(NULL, dport, component_reg_phys);
1173 		if (rc)
1174 			return ERR_PTR(rc);
1175 
1176 		dport->rch = true;
1177 	}
1178 
1179 	if (component_reg_phys != CXL_RESOURCE_NONE)
1180 		dev_dbg(dport_dev, "Component Registers found for dport: %pa\n",
1181 			&component_reg_phys);
1182 
1183 	cond_cxl_root_lock(port);
1184 	rc = add_dport(port, dport);
1185 	cond_cxl_root_unlock(port);
1186 	if (rc)
1187 		return ERR_PTR(rc);
1188 
1189 	get_device(dport_dev);
1190 	rc = devm_add_action_or_reset(host, cxl_dport_remove, dport);
1191 	if (rc)
1192 		return ERR_PTR(rc);
1193 
1194 	rc = sysfs_create_link(&port->dev.kobj, &dport_dev->kobj, link_name);
1195 	if (rc)
1196 		return ERR_PTR(rc);
1197 
1198 	rc = devm_add_action_or_reset(host, cxl_dport_unlink, dport);
1199 	if (rc)
1200 		return ERR_PTR(rc);
1201 
1202 	if (dev_is_pci(dport_dev))
1203 		dport->link_latency = cxl_pci_get_latency(to_pci_dev(dport_dev));
1204 
1205 	cxl_debugfs_create_dport_dir(dport);
1206 
1207 	return dport;
1208 }
1209 
1210 /**
1211  * devm_cxl_add_dport - append VH downstream port data to a cxl_port
1212  * @port: the cxl_port that references this dport
1213  * @dport_dev: firmware or PCI device representing the dport
1214  * @port_id: identifier for this dport in a decoder's target list
1215  * @component_reg_phys: optional location of CXL component registers
1216  *
1217  * Note that dports are appended to the devm release action's of the
1218  * either the port's host (for root ports), or the port itself (for
1219  * switch ports)
1220  */
devm_cxl_add_dport(struct cxl_port * port,struct device * dport_dev,int port_id,resource_size_t component_reg_phys)1221 struct cxl_dport *devm_cxl_add_dport(struct cxl_port *port,
1222 				     struct device *dport_dev, int port_id,
1223 				     resource_size_t component_reg_phys)
1224 {
1225 	struct cxl_dport *dport;
1226 
1227 	dport = __devm_cxl_add_dport(port, dport_dev, port_id,
1228 				     component_reg_phys, CXL_RESOURCE_NONE);
1229 	if (IS_ERR(dport)) {
1230 		dev_dbg(dport_dev, "failed to add dport to %s: %ld\n",
1231 			dev_name(&port->dev), PTR_ERR(dport));
1232 	} else {
1233 		dev_dbg(dport_dev, "dport added to %s\n",
1234 			dev_name(&port->dev));
1235 	}
1236 
1237 	return dport;
1238 }
1239 EXPORT_SYMBOL_NS_GPL(devm_cxl_add_dport, "CXL");
1240 
1241 /**
1242  * devm_cxl_add_rch_dport - append RCH downstream port data to a cxl_port
1243  * @port: the cxl_port that references this dport
1244  * @dport_dev: firmware or PCI device representing the dport
1245  * @port_id: identifier for this dport in a decoder's target list
1246  * @rcrb: mandatory location of a Root Complex Register Block
1247  *
1248  * See CXL 3.0 9.11.8 CXL Devices Attached to an RCH
1249  */
devm_cxl_add_rch_dport(struct cxl_port * port,struct device * dport_dev,int port_id,resource_size_t rcrb)1250 struct cxl_dport *devm_cxl_add_rch_dport(struct cxl_port *port,
1251 					 struct device *dport_dev, int port_id,
1252 					 resource_size_t rcrb)
1253 {
1254 	struct cxl_dport *dport;
1255 
1256 	if (rcrb == CXL_RESOURCE_NONE) {
1257 		dev_dbg(&port->dev, "failed to add RCH dport, missing RCRB\n");
1258 		return ERR_PTR(-EINVAL);
1259 	}
1260 
1261 	dport = __devm_cxl_add_dport(port, dport_dev, port_id,
1262 				     CXL_RESOURCE_NONE, rcrb);
1263 	if (IS_ERR(dport)) {
1264 		dev_dbg(dport_dev, "failed to add RCH dport to %s: %ld\n",
1265 			dev_name(&port->dev), PTR_ERR(dport));
1266 	} else {
1267 		dev_dbg(dport_dev, "RCH dport added to %s\n",
1268 			dev_name(&port->dev));
1269 	}
1270 
1271 	return dport;
1272 }
1273 EXPORT_SYMBOL_NS_GPL(devm_cxl_add_rch_dport, "CXL");
1274 
add_ep(struct cxl_ep * new)1275 static int add_ep(struct cxl_ep *new)
1276 {
1277 	struct cxl_port *port = new->dport->port;
1278 
1279 	guard(device)(&port->dev);
1280 	if (port->dead)
1281 		return -ENXIO;
1282 
1283 	return xa_insert(&port->endpoints, (unsigned long)new->ep,
1284 			 new, GFP_KERNEL);
1285 }
1286 
1287 /**
1288  * cxl_add_ep - register an endpoint's interest in a port
1289  * @dport: the dport that routes to @ep_dev
1290  * @ep_dev: device representing the endpoint
1291  *
1292  * Intermediate CXL ports are scanned based on the arrival of endpoints.
1293  * When those endpoints depart the port can be destroyed once all
1294  * endpoints that care about that port have been removed.
1295  */
cxl_add_ep(struct cxl_dport * dport,struct device * ep_dev)1296 static int cxl_add_ep(struct cxl_dport *dport, struct device *ep_dev)
1297 {
1298 	struct cxl_ep *ep;
1299 	int rc;
1300 
1301 	ep = kzalloc(sizeof(*ep), GFP_KERNEL);
1302 	if (!ep)
1303 		return -ENOMEM;
1304 
1305 	ep->ep = get_device(ep_dev);
1306 	ep->dport = dport;
1307 
1308 	rc = add_ep(ep);
1309 	if (rc)
1310 		cxl_ep_release(ep);
1311 	return rc;
1312 }
1313 
1314 struct cxl_find_port_ctx {
1315 	const struct device *dport_dev;
1316 	const struct cxl_port *parent_port;
1317 	struct cxl_dport **dport;
1318 };
1319 
match_port_by_dport(struct device * dev,const void * data)1320 static int match_port_by_dport(struct device *dev, const void *data)
1321 {
1322 	const struct cxl_find_port_ctx *ctx = data;
1323 	struct cxl_dport *dport;
1324 	struct cxl_port *port;
1325 
1326 	if (!is_cxl_port(dev))
1327 		return 0;
1328 	if (ctx->parent_port && dev->parent != &ctx->parent_port->dev)
1329 		return 0;
1330 
1331 	port = to_cxl_port(dev);
1332 	dport = cxl_find_dport_by_dev(port, ctx->dport_dev);
1333 	if (ctx->dport)
1334 		*ctx->dport = dport;
1335 	return dport != NULL;
1336 }
1337 
__find_cxl_port(struct cxl_find_port_ctx * ctx)1338 static struct cxl_port *__find_cxl_port(struct cxl_find_port_ctx *ctx)
1339 {
1340 	struct device *dev;
1341 
1342 	if (!ctx->dport_dev)
1343 		return NULL;
1344 
1345 	dev = bus_find_device(&cxl_bus_type, NULL, ctx, match_port_by_dport);
1346 	if (dev)
1347 		return to_cxl_port(dev);
1348 	return NULL;
1349 }
1350 
find_cxl_port(struct device * dport_dev,struct cxl_dport ** dport)1351 static struct cxl_port *find_cxl_port(struct device *dport_dev,
1352 				      struct cxl_dport **dport)
1353 {
1354 	struct cxl_find_port_ctx ctx = {
1355 		.dport_dev = dport_dev,
1356 		.dport = dport,
1357 	};
1358 	struct cxl_port *port;
1359 
1360 	port = __find_cxl_port(&ctx);
1361 	return port;
1362 }
1363 
find_cxl_port_at(struct cxl_port * parent_port,struct device * dport_dev,struct cxl_dport ** dport)1364 static struct cxl_port *find_cxl_port_at(struct cxl_port *parent_port,
1365 					 struct device *dport_dev,
1366 					 struct cxl_dport **dport)
1367 {
1368 	struct cxl_find_port_ctx ctx = {
1369 		.dport_dev = dport_dev,
1370 		.parent_port = parent_port,
1371 		.dport = dport,
1372 	};
1373 	struct cxl_port *port;
1374 
1375 	port = __find_cxl_port(&ctx);
1376 	return port;
1377 }
1378 
1379 /*
1380  * All users of grandparent() are using it to walk PCIe-like switch port
1381  * hierarchy. A PCIe switch is comprised of a bridge device representing the
1382  * upstream switch port and N bridges representing downstream switch ports. When
1383  * bridges stack the grand-parent of a downstream switch port is another
1384  * downstream switch port in the immediate ancestor switch.
1385  */
grandparent(struct device * dev)1386 static struct device *grandparent(struct device *dev)
1387 {
1388 	if (dev && dev->parent)
1389 		return dev->parent->parent;
1390 	return NULL;
1391 }
1392 
endpoint_host(struct cxl_port * endpoint)1393 static struct device *endpoint_host(struct cxl_port *endpoint)
1394 {
1395 	struct cxl_port *port = to_cxl_port(endpoint->dev.parent);
1396 
1397 	if (is_cxl_root(port))
1398 		return port->uport_dev;
1399 	return &port->dev;
1400 }
1401 
delete_endpoint(void * data)1402 static void delete_endpoint(void *data)
1403 {
1404 	struct cxl_memdev *cxlmd = data;
1405 	struct cxl_port *endpoint = cxlmd->endpoint;
1406 	struct device *host = endpoint_host(endpoint);
1407 
1408 	scoped_guard(device, host) {
1409 		if (host->driver && !endpoint->dead) {
1410 			devm_release_action(host, cxl_unlink_parent_dport, endpoint);
1411 			devm_release_action(host, cxl_unlink_uport, endpoint);
1412 			devm_release_action(host, unregister_port, endpoint);
1413 		}
1414 		cxlmd->endpoint = NULL;
1415 	}
1416 	put_device(&endpoint->dev);
1417 	put_device(host);
1418 }
1419 
cxl_endpoint_autoremove(struct cxl_memdev * cxlmd,struct cxl_port * endpoint)1420 int cxl_endpoint_autoremove(struct cxl_memdev *cxlmd, struct cxl_port *endpoint)
1421 {
1422 	struct device *host = endpoint_host(endpoint);
1423 	struct device *dev = &cxlmd->dev;
1424 
1425 	get_device(host);
1426 	get_device(&endpoint->dev);
1427 	cxlmd->depth = endpoint->depth;
1428 	return devm_add_action_or_reset(dev, delete_endpoint, cxlmd);
1429 }
1430 EXPORT_SYMBOL_NS_GPL(cxl_endpoint_autoremove, "CXL");
1431 
1432 /*
1433  * The natural end of life of a non-root 'cxl_port' is when its parent port goes
1434  * through a ->remove() event ("top-down" unregistration). The unnatural trigger
1435  * for a port to be unregistered is when all memdevs beneath that port have gone
1436  * through ->remove(). This "bottom-up" removal selectively removes individual
1437  * child ports manually. This depends on devm_cxl_add_port() to not change is
1438  * devm action registration order, and for dports to have already been
1439  * destroyed by reap_dports().
1440  */
delete_switch_port(struct cxl_port * port)1441 static void delete_switch_port(struct cxl_port *port)
1442 {
1443 	devm_release_action(port->dev.parent, cxl_unlink_parent_dport, port);
1444 	devm_release_action(port->dev.parent, cxl_unlink_uport, port);
1445 	devm_release_action(port->dev.parent, unregister_port, port);
1446 }
1447 
reap_dports(struct cxl_port * port)1448 static void reap_dports(struct cxl_port *port)
1449 {
1450 	struct cxl_dport *dport;
1451 	unsigned long index;
1452 
1453 	device_lock_assert(&port->dev);
1454 
1455 	xa_for_each(&port->dports, index, dport) {
1456 		devm_release_action(&port->dev, cxl_dport_unlink, dport);
1457 		devm_release_action(&port->dev, cxl_dport_remove, dport);
1458 		devm_kfree(&port->dev, dport);
1459 	}
1460 }
1461 
1462 struct detach_ctx {
1463 	struct cxl_memdev *cxlmd;
1464 	int depth;
1465 };
1466 
port_has_memdev(struct device * dev,const void * data)1467 static int port_has_memdev(struct device *dev, const void *data)
1468 {
1469 	const struct detach_ctx *ctx = data;
1470 	struct cxl_port *port;
1471 
1472 	if (!is_cxl_port(dev))
1473 		return 0;
1474 
1475 	port = to_cxl_port(dev);
1476 	if (port->depth != ctx->depth)
1477 		return 0;
1478 
1479 	return !!cxl_ep_load(port, ctx->cxlmd);
1480 }
1481 
cxl_detach_ep(void * data)1482 static void cxl_detach_ep(void *data)
1483 {
1484 	struct cxl_memdev *cxlmd = data;
1485 
1486 	for (int i = cxlmd->depth - 1; i >= 1; i--) {
1487 		struct cxl_port *port, *parent_port;
1488 		struct detach_ctx ctx = {
1489 			.cxlmd = cxlmd,
1490 			.depth = i,
1491 		};
1492 		struct cxl_ep *ep;
1493 		bool died = false;
1494 
1495 		struct device *dev __free(put_device) =
1496 			bus_find_device(&cxl_bus_type, NULL, &ctx, port_has_memdev);
1497 		if (!dev)
1498 			continue;
1499 		port = to_cxl_port(dev);
1500 
1501 		parent_port = to_cxl_port(port->dev.parent);
1502 		device_lock(&parent_port->dev);
1503 		device_lock(&port->dev);
1504 		ep = cxl_ep_load(port, cxlmd);
1505 		dev_dbg(&cxlmd->dev, "disconnect %s from %s\n",
1506 			ep ? dev_name(ep->ep) : "", dev_name(&port->dev));
1507 		cxl_ep_remove(port, ep);
1508 		if (ep && !port->dead && xa_empty(&port->endpoints) &&
1509 		    !is_cxl_root(parent_port) && parent_port->dev.driver) {
1510 			/*
1511 			 * This was the last ep attached to a dynamically
1512 			 * enumerated port. Block new cxl_add_ep() and garbage
1513 			 * collect the port.
1514 			 */
1515 			died = true;
1516 			port->dead = true;
1517 			reap_dports(port);
1518 		}
1519 		device_unlock(&port->dev);
1520 
1521 		if (died) {
1522 			dev_dbg(&cxlmd->dev, "delete %s\n",
1523 				dev_name(&port->dev));
1524 			delete_switch_port(port);
1525 		}
1526 		device_unlock(&parent_port->dev);
1527 	}
1528 }
1529 
find_component_registers(struct device * dev)1530 static resource_size_t find_component_registers(struct device *dev)
1531 {
1532 	struct cxl_register_map map;
1533 	struct pci_dev *pdev;
1534 
1535 	/*
1536 	 * Theoretically, CXL component registers can be hosted on a
1537 	 * non-PCI device, in practice, only cxl_test hits this case.
1538 	 */
1539 	if (!dev_is_pci(dev))
1540 		return CXL_RESOURCE_NONE;
1541 
1542 	pdev = to_pci_dev(dev);
1543 
1544 	cxl_find_regblock(pdev, CXL_REGLOC_RBI_COMPONENT, &map);
1545 	return map.resource;
1546 }
1547 
add_port_attach_ep(struct cxl_memdev * cxlmd,struct device * uport_dev,struct device * dport_dev)1548 static int add_port_attach_ep(struct cxl_memdev *cxlmd,
1549 			      struct device *uport_dev,
1550 			      struct device *dport_dev)
1551 {
1552 	struct device *dparent = grandparent(dport_dev);
1553 	struct cxl_dport *dport, *parent_dport;
1554 	resource_size_t component_reg_phys;
1555 	int rc;
1556 
1557 	if (!dparent) {
1558 		/*
1559 		 * The iteration reached the topology root without finding the
1560 		 * CXL-root 'cxl_port' on a previous iteration, fail for now to
1561 		 * be re-probed after platform driver attaches.
1562 		 */
1563 		dev_dbg(&cxlmd->dev, "%s is a root dport\n",
1564 			dev_name(dport_dev));
1565 		return -ENXIO;
1566 	}
1567 
1568 	struct cxl_port *parent_port __free(put_cxl_port) =
1569 		find_cxl_port(dparent, &parent_dport);
1570 	if (!parent_port) {
1571 		/* iterate to create this parent_port */
1572 		return -EAGAIN;
1573 	}
1574 
1575 	/*
1576 	 * Definition with __free() here to keep the sequence of
1577 	 * dereferencing the device of the port before the parent_port releasing.
1578 	 */
1579 	struct cxl_port *port __free(put_cxl_port) = NULL;
1580 	scoped_guard(device, &parent_port->dev) {
1581 		if (!parent_port->dev.driver) {
1582 			dev_warn(&cxlmd->dev,
1583 				 "port %s:%s disabled, failed to enumerate CXL.mem\n",
1584 				 dev_name(&parent_port->dev), dev_name(uport_dev));
1585 			return -ENXIO;
1586 		}
1587 
1588 		port = find_cxl_port_at(parent_port, dport_dev, &dport);
1589 		if (!port) {
1590 			component_reg_phys = find_component_registers(uport_dev);
1591 			port = devm_cxl_add_port(&parent_port->dev, uport_dev,
1592 						 component_reg_phys, parent_dport);
1593 			if (IS_ERR(port))
1594 				return PTR_ERR(port);
1595 
1596 			/* retry find to pick up the new dport information */
1597 			port = find_cxl_port_at(parent_port, dport_dev, &dport);
1598 			if (!port)
1599 				return -ENXIO;
1600 		}
1601 	}
1602 
1603 	dev_dbg(&cxlmd->dev, "add to new port %s:%s\n",
1604 		dev_name(&port->dev), dev_name(port->uport_dev));
1605 	rc = cxl_add_ep(dport, &cxlmd->dev);
1606 	if (rc == -EBUSY) {
1607 		/*
1608 		 * "can't" happen, but this error code means
1609 		 * something to the caller, so translate it.
1610 		 */
1611 		rc = -ENXIO;
1612 	}
1613 
1614 	return rc;
1615 }
1616 
devm_cxl_enumerate_ports(struct cxl_memdev * cxlmd)1617 int devm_cxl_enumerate_ports(struct cxl_memdev *cxlmd)
1618 {
1619 	struct device *dev = &cxlmd->dev;
1620 	struct device *iter;
1621 	int rc;
1622 
1623 	/*
1624 	 * Skip intermediate port enumeration in the RCH case, there
1625 	 * are no ports in between a host bridge and an endpoint.
1626 	 */
1627 	if (cxlmd->cxlds->rcd)
1628 		return 0;
1629 
1630 	rc = devm_add_action_or_reset(&cxlmd->dev, cxl_detach_ep, cxlmd);
1631 	if (rc)
1632 		return rc;
1633 
1634 	/*
1635 	 * Scan for and add all cxl_ports in this device's ancestry.
1636 	 * Repeat until no more ports are added. Abort if a port add
1637 	 * attempt fails.
1638 	 */
1639 retry:
1640 	for (iter = dev; iter; iter = grandparent(iter)) {
1641 		struct device *dport_dev = grandparent(iter);
1642 		struct device *uport_dev;
1643 		struct cxl_dport *dport;
1644 
1645 		/*
1646 		 * The terminal "grandparent" in PCI is NULL and @platform_bus
1647 		 * for platform devices
1648 		 */
1649 		if (!dport_dev || dport_dev == &platform_bus)
1650 			return 0;
1651 
1652 		uport_dev = dport_dev->parent;
1653 		if (!uport_dev) {
1654 			dev_warn(dev, "at %s no parent for dport: %s\n",
1655 				 dev_name(iter), dev_name(dport_dev));
1656 			return -ENXIO;
1657 		}
1658 
1659 		dev_dbg(dev, "scan: iter: %s dport_dev: %s parent: %s\n",
1660 			dev_name(iter), dev_name(dport_dev),
1661 			dev_name(uport_dev));
1662 		struct cxl_port *port __free(put_cxl_port) =
1663 			find_cxl_port(dport_dev, &dport);
1664 		if (port) {
1665 			dev_dbg(&cxlmd->dev,
1666 				"found already registered port %s:%s\n",
1667 				dev_name(&port->dev),
1668 				dev_name(port->uport_dev));
1669 			rc = cxl_add_ep(dport, &cxlmd->dev);
1670 
1671 			/*
1672 			 * If the endpoint already exists in the port's list,
1673 			 * that's ok, it was added on a previous pass.
1674 			 * Otherwise, retry in add_port_attach_ep() after taking
1675 			 * the parent_port lock as the current port may be being
1676 			 * reaped.
1677 			 */
1678 			if (rc && rc != -EBUSY)
1679 				return rc;
1680 
1681 			cxl_gpf_port_setup(dport_dev, port);
1682 
1683 			/* Any more ports to add between this one and the root? */
1684 			if (!dev_is_cxl_root_child(&port->dev))
1685 				continue;
1686 
1687 			return 0;
1688 		}
1689 
1690 		rc = add_port_attach_ep(cxlmd, uport_dev, dport_dev);
1691 		/* port missing, try to add parent */
1692 		if (rc == -EAGAIN)
1693 			continue;
1694 		/* failed to add ep or port */
1695 		if (rc)
1696 			return rc;
1697 		/* port added, new descendants possible, start over */
1698 		goto retry;
1699 	}
1700 
1701 	return 0;
1702 }
1703 EXPORT_SYMBOL_NS_GPL(devm_cxl_enumerate_ports, "CXL");
1704 
cxl_pci_find_port(struct pci_dev * pdev,struct cxl_dport ** dport)1705 struct cxl_port *cxl_pci_find_port(struct pci_dev *pdev,
1706 				   struct cxl_dport **dport)
1707 {
1708 	return find_cxl_port(pdev->dev.parent, dport);
1709 }
1710 EXPORT_SYMBOL_NS_GPL(cxl_pci_find_port, "CXL");
1711 
cxl_mem_find_port(struct cxl_memdev * cxlmd,struct cxl_dport ** dport)1712 struct cxl_port *cxl_mem_find_port(struct cxl_memdev *cxlmd,
1713 				   struct cxl_dport **dport)
1714 {
1715 	return find_cxl_port(grandparent(&cxlmd->dev), dport);
1716 }
1717 EXPORT_SYMBOL_NS_GPL(cxl_mem_find_port, "CXL");
1718 
decoder_populate_targets(struct cxl_switch_decoder * cxlsd,struct cxl_port * port,int * target_map)1719 static int decoder_populate_targets(struct cxl_switch_decoder *cxlsd,
1720 				    struct cxl_port *port, int *target_map)
1721 {
1722 	int i;
1723 
1724 	if (!target_map)
1725 		return 0;
1726 
1727 	device_lock_assert(&port->dev);
1728 
1729 	if (xa_empty(&port->dports))
1730 		return -EINVAL;
1731 
1732 	guard(rwsem_write)(&cxl_region_rwsem);
1733 	for (i = 0; i < cxlsd->cxld.interleave_ways; i++) {
1734 		struct cxl_dport *dport = find_dport(port, target_map[i]);
1735 
1736 		if (!dport)
1737 			return -ENXIO;
1738 		cxlsd->target[i] = dport;
1739 	}
1740 
1741 	return 0;
1742 }
1743 
1744 static struct lock_class_key cxl_decoder_key;
1745 
1746 /**
1747  * cxl_decoder_init - Common decoder setup / initialization
1748  * @port: owning port of this decoder
1749  * @cxld: common decoder properties to initialize
1750  *
1751  * A port may contain one or more decoders. Each of those decoders
1752  * enable some address space for CXL.mem utilization. A decoder is
1753  * expected to be configured by the caller before registering via
1754  * cxl_decoder_add()
1755  */
cxl_decoder_init(struct cxl_port * port,struct cxl_decoder * cxld)1756 static int cxl_decoder_init(struct cxl_port *port, struct cxl_decoder *cxld)
1757 {
1758 	struct device *dev;
1759 	int rc;
1760 
1761 	rc = ida_alloc(&port->decoder_ida, GFP_KERNEL);
1762 	if (rc < 0)
1763 		return rc;
1764 
1765 	/* need parent to stick around to release the id */
1766 	get_device(&port->dev);
1767 	cxld->id = rc;
1768 
1769 	dev = &cxld->dev;
1770 	device_initialize(dev);
1771 	lockdep_set_class(&dev->mutex, &cxl_decoder_key);
1772 	device_set_pm_not_required(dev);
1773 	dev->parent = &port->dev;
1774 	dev->bus = &cxl_bus_type;
1775 
1776 	/* Pre initialize an "empty" decoder */
1777 	cxld->interleave_ways = 1;
1778 	cxld->interleave_granularity = PAGE_SIZE;
1779 	cxld->target_type = CXL_DECODER_HOSTONLYMEM;
1780 	cxld->hpa_range = (struct range) {
1781 		.start = 0,
1782 		.end = -1,
1783 	};
1784 
1785 	return 0;
1786 }
1787 
cxl_switch_decoder_init(struct cxl_port * port,struct cxl_switch_decoder * cxlsd,int nr_targets)1788 static int cxl_switch_decoder_init(struct cxl_port *port,
1789 				   struct cxl_switch_decoder *cxlsd,
1790 				   int nr_targets)
1791 {
1792 	if (nr_targets > CXL_DECODER_MAX_INTERLEAVE)
1793 		return -EINVAL;
1794 
1795 	cxlsd->nr_targets = nr_targets;
1796 	return cxl_decoder_init(port, &cxlsd->cxld);
1797 }
1798 
1799 /**
1800  * cxl_root_decoder_alloc - Allocate a root level decoder
1801  * @port: owning CXL root of this decoder
1802  * @nr_targets: static number of downstream targets
1803  *
1804  * Return: A new cxl decoder to be registered by cxl_decoder_add(). A
1805  * 'CXL root' decoder is one that decodes from a top-level / static platform
1806  * firmware description of CXL resources into a CXL standard decode
1807  * topology.
1808  */
cxl_root_decoder_alloc(struct cxl_port * port,unsigned int nr_targets)1809 struct cxl_root_decoder *cxl_root_decoder_alloc(struct cxl_port *port,
1810 						unsigned int nr_targets)
1811 {
1812 	struct cxl_root_decoder *cxlrd;
1813 	struct cxl_switch_decoder *cxlsd;
1814 	struct cxl_decoder *cxld;
1815 	int rc;
1816 
1817 	if (!is_cxl_root(port))
1818 		return ERR_PTR(-EINVAL);
1819 
1820 	cxlrd = kzalloc(struct_size(cxlrd, cxlsd.target, nr_targets),
1821 			GFP_KERNEL);
1822 	if (!cxlrd)
1823 		return ERR_PTR(-ENOMEM);
1824 
1825 	cxlsd = &cxlrd->cxlsd;
1826 	rc = cxl_switch_decoder_init(port, cxlsd, nr_targets);
1827 	if (rc) {
1828 		kfree(cxlrd);
1829 		return ERR_PTR(rc);
1830 	}
1831 
1832 	mutex_init(&cxlrd->range_lock);
1833 
1834 	cxld = &cxlsd->cxld;
1835 	cxld->dev.type = &cxl_decoder_root_type;
1836 	/*
1837 	 * cxl_root_decoder_release() special cases negative ids to
1838 	 * detect memregion_alloc() failures.
1839 	 */
1840 	atomic_set(&cxlrd->region_id, -1);
1841 	rc = memregion_alloc(GFP_KERNEL);
1842 	if (rc < 0) {
1843 		put_device(&cxld->dev);
1844 		return ERR_PTR(rc);
1845 	}
1846 
1847 	atomic_set(&cxlrd->region_id, rc);
1848 	cxlrd->qos_class = CXL_QOS_CLASS_INVALID;
1849 	return cxlrd;
1850 }
1851 EXPORT_SYMBOL_NS_GPL(cxl_root_decoder_alloc, "CXL");
1852 
1853 /**
1854  * cxl_switch_decoder_alloc - Allocate a switch level decoder
1855  * @port: owning CXL switch port of this decoder
1856  * @nr_targets: max number of dynamically addressable downstream targets
1857  *
1858  * Return: A new cxl decoder to be registered by cxl_decoder_add(). A
1859  * 'switch' decoder is any decoder that can be enumerated by PCIe
1860  * topology and the HDM Decoder Capability. This includes the decoders
1861  * that sit between Switch Upstream Ports / Switch Downstream Ports and
1862  * Host Bridges / Root Ports.
1863  */
cxl_switch_decoder_alloc(struct cxl_port * port,unsigned int nr_targets)1864 struct cxl_switch_decoder *cxl_switch_decoder_alloc(struct cxl_port *port,
1865 						    unsigned int nr_targets)
1866 {
1867 	struct cxl_switch_decoder *cxlsd;
1868 	struct cxl_decoder *cxld;
1869 	int rc;
1870 
1871 	if (is_cxl_root(port) || is_cxl_endpoint(port))
1872 		return ERR_PTR(-EINVAL);
1873 
1874 	cxlsd = kzalloc(struct_size(cxlsd, target, nr_targets), GFP_KERNEL);
1875 	if (!cxlsd)
1876 		return ERR_PTR(-ENOMEM);
1877 
1878 	rc = cxl_switch_decoder_init(port, cxlsd, nr_targets);
1879 	if (rc) {
1880 		kfree(cxlsd);
1881 		return ERR_PTR(rc);
1882 	}
1883 
1884 	cxld = &cxlsd->cxld;
1885 	cxld->dev.type = &cxl_decoder_switch_type;
1886 	return cxlsd;
1887 }
1888 EXPORT_SYMBOL_NS_GPL(cxl_switch_decoder_alloc, "CXL");
1889 
1890 /**
1891  * cxl_endpoint_decoder_alloc - Allocate an endpoint decoder
1892  * @port: owning port of this decoder
1893  *
1894  * Return: A new cxl decoder to be registered by cxl_decoder_add()
1895  */
cxl_endpoint_decoder_alloc(struct cxl_port * port)1896 struct cxl_endpoint_decoder *cxl_endpoint_decoder_alloc(struct cxl_port *port)
1897 {
1898 	struct cxl_endpoint_decoder *cxled;
1899 	struct cxl_decoder *cxld;
1900 	int rc;
1901 
1902 	if (!is_cxl_endpoint(port))
1903 		return ERR_PTR(-EINVAL);
1904 
1905 	cxled = kzalloc(sizeof(*cxled), GFP_KERNEL);
1906 	if (!cxled)
1907 		return ERR_PTR(-ENOMEM);
1908 
1909 	cxled->pos = -1;
1910 	cxled->part = -1;
1911 	cxld = &cxled->cxld;
1912 	rc = cxl_decoder_init(port, cxld);
1913 	if (rc)	 {
1914 		kfree(cxled);
1915 		return ERR_PTR(rc);
1916 	}
1917 
1918 	cxld->dev.type = &cxl_decoder_endpoint_type;
1919 	return cxled;
1920 }
1921 EXPORT_SYMBOL_NS_GPL(cxl_endpoint_decoder_alloc, "CXL");
1922 
1923 /**
1924  * cxl_decoder_add_locked - Add a decoder with targets
1925  * @cxld: The cxl decoder allocated by cxl_<type>_decoder_alloc()
1926  * @target_map: A list of downstream ports that this decoder can direct memory
1927  *              traffic to. These numbers should correspond with the port number
1928  *              in the PCIe Link Capabilities structure.
1929  *
1930  * Certain types of decoders may not have any targets. The main example of this
1931  * is an endpoint device. A more awkward example is a hostbridge whose root
1932  * ports get hot added (technically possible, though unlikely).
1933  *
1934  * This is the locked variant of cxl_decoder_add().
1935  *
1936  * Context: Process context. Expects the device lock of the port that owns the
1937  *	    @cxld to be held.
1938  *
1939  * Return: Negative error code if the decoder wasn't properly configured; else
1940  *	   returns 0.
1941  */
cxl_decoder_add_locked(struct cxl_decoder * cxld,int * target_map)1942 int cxl_decoder_add_locked(struct cxl_decoder *cxld, int *target_map)
1943 {
1944 	struct cxl_port *port;
1945 	struct device *dev;
1946 	int rc;
1947 
1948 	if (WARN_ON_ONCE(!cxld))
1949 		return -EINVAL;
1950 
1951 	if (WARN_ON_ONCE(IS_ERR(cxld)))
1952 		return PTR_ERR(cxld);
1953 
1954 	if (cxld->interleave_ways < 1)
1955 		return -EINVAL;
1956 
1957 	dev = &cxld->dev;
1958 
1959 	port = to_cxl_port(cxld->dev.parent);
1960 	if (!is_endpoint_decoder(dev)) {
1961 		struct cxl_switch_decoder *cxlsd = to_cxl_switch_decoder(dev);
1962 
1963 		rc = decoder_populate_targets(cxlsd, port, target_map);
1964 		if (rc && (cxld->flags & CXL_DECODER_F_ENABLE)) {
1965 			dev_err(&port->dev,
1966 				"Failed to populate active decoder targets\n");
1967 			return rc;
1968 		}
1969 	}
1970 
1971 	rc = dev_set_name(dev, "decoder%d.%d", port->id, cxld->id);
1972 	if (rc)
1973 		return rc;
1974 
1975 	return device_add(dev);
1976 }
1977 EXPORT_SYMBOL_NS_GPL(cxl_decoder_add_locked, "CXL");
1978 
1979 /**
1980  * cxl_decoder_add - Add a decoder with targets
1981  * @cxld: The cxl decoder allocated by cxl_<type>_decoder_alloc()
1982  * @target_map: A list of downstream ports that this decoder can direct memory
1983  *              traffic to. These numbers should correspond with the port number
1984  *              in the PCIe Link Capabilities structure.
1985  *
1986  * This is the unlocked variant of cxl_decoder_add_locked().
1987  * See cxl_decoder_add_locked().
1988  *
1989  * Context: Process context. Takes and releases the device lock of the port that
1990  *	    owns the @cxld.
1991  */
cxl_decoder_add(struct cxl_decoder * cxld,int * target_map)1992 int cxl_decoder_add(struct cxl_decoder *cxld, int *target_map)
1993 {
1994 	struct cxl_port *port;
1995 
1996 	if (WARN_ON_ONCE(!cxld))
1997 		return -EINVAL;
1998 
1999 	if (WARN_ON_ONCE(IS_ERR(cxld)))
2000 		return PTR_ERR(cxld);
2001 
2002 	port = to_cxl_port(cxld->dev.parent);
2003 
2004 	guard(device)(&port->dev);
2005 	return cxl_decoder_add_locked(cxld, target_map);
2006 }
2007 EXPORT_SYMBOL_NS_GPL(cxl_decoder_add, "CXL");
2008 
cxld_unregister(void * dev)2009 static void cxld_unregister(void *dev)
2010 {
2011 	struct cxl_endpoint_decoder *cxled;
2012 
2013 	if (is_endpoint_decoder(dev)) {
2014 		cxled = to_cxl_endpoint_decoder(dev);
2015 		cxl_decoder_kill_region(cxled);
2016 	}
2017 
2018 	device_unregister(dev);
2019 }
2020 
cxl_decoder_autoremove(struct device * host,struct cxl_decoder * cxld)2021 int cxl_decoder_autoremove(struct device *host, struct cxl_decoder *cxld)
2022 {
2023 	return devm_add_action_or_reset(host, cxld_unregister, &cxld->dev);
2024 }
2025 EXPORT_SYMBOL_NS_GPL(cxl_decoder_autoremove, "CXL");
2026 
2027 /**
2028  * __cxl_driver_register - register a driver for the cxl bus
2029  * @cxl_drv: cxl driver structure to attach
2030  * @owner: owning module/driver
2031  * @modname: KBUILD_MODNAME for parent driver
2032  */
__cxl_driver_register(struct cxl_driver * cxl_drv,struct module * owner,const char * modname)2033 int __cxl_driver_register(struct cxl_driver *cxl_drv, struct module *owner,
2034 			  const char *modname)
2035 {
2036 	if (!cxl_drv->probe) {
2037 		pr_debug("%s ->probe() must be specified\n", modname);
2038 		return -EINVAL;
2039 	}
2040 
2041 	if (!cxl_drv->name) {
2042 		pr_debug("%s ->name must be specified\n", modname);
2043 		return -EINVAL;
2044 	}
2045 
2046 	if (!cxl_drv->id) {
2047 		pr_debug("%s ->id must be specified\n", modname);
2048 		return -EINVAL;
2049 	}
2050 
2051 	cxl_drv->drv.bus = &cxl_bus_type;
2052 	cxl_drv->drv.owner = owner;
2053 	cxl_drv->drv.mod_name = modname;
2054 	cxl_drv->drv.name = cxl_drv->name;
2055 
2056 	return driver_register(&cxl_drv->drv);
2057 }
2058 EXPORT_SYMBOL_NS_GPL(__cxl_driver_register, "CXL");
2059 
cxl_driver_unregister(struct cxl_driver * cxl_drv)2060 void cxl_driver_unregister(struct cxl_driver *cxl_drv)
2061 {
2062 	driver_unregister(&cxl_drv->drv);
2063 }
2064 EXPORT_SYMBOL_NS_GPL(cxl_driver_unregister, "CXL");
2065 
cxl_bus_uevent(const struct device * dev,struct kobj_uevent_env * env)2066 static int cxl_bus_uevent(const struct device *dev, struct kobj_uevent_env *env)
2067 {
2068 	return add_uevent_var(env, "MODALIAS=" CXL_MODALIAS_FMT,
2069 			      cxl_device_id(dev));
2070 }
2071 
cxl_bus_match(struct device * dev,const struct device_driver * drv)2072 static int cxl_bus_match(struct device *dev, const struct device_driver *drv)
2073 {
2074 	return cxl_device_id(dev) == to_cxl_drv(drv)->id;
2075 }
2076 
cxl_bus_probe(struct device * dev)2077 static int cxl_bus_probe(struct device *dev)
2078 {
2079 	int rc;
2080 
2081 	rc = to_cxl_drv(dev->driver)->probe(dev);
2082 	dev_dbg(dev, "probe: %d\n", rc);
2083 	return rc;
2084 }
2085 
cxl_bus_remove(struct device * dev)2086 static void cxl_bus_remove(struct device *dev)
2087 {
2088 	struct cxl_driver *cxl_drv = to_cxl_drv(dev->driver);
2089 
2090 	if (cxl_drv->remove)
2091 		cxl_drv->remove(dev);
2092 }
2093 
2094 static struct workqueue_struct *cxl_bus_wq;
2095 
cxl_rescan_attach(struct device * dev,void * data)2096 static int cxl_rescan_attach(struct device *dev, void *data)
2097 {
2098 	int rc = device_attach(dev);
2099 
2100 	dev_vdbg(dev, "rescan: %s\n", rc ? "attach" : "detached");
2101 
2102 	return 0;
2103 }
2104 
cxl_bus_rescan_queue(struct work_struct * w)2105 static void cxl_bus_rescan_queue(struct work_struct *w)
2106 {
2107 	bus_for_each_dev(&cxl_bus_type, NULL, NULL, cxl_rescan_attach);
2108 }
2109 
cxl_bus_rescan(void)2110 void cxl_bus_rescan(void)
2111 {
2112 	static DECLARE_WORK(rescan_work, cxl_bus_rescan_queue);
2113 
2114 	queue_work(cxl_bus_wq, &rescan_work);
2115 }
2116 EXPORT_SYMBOL_NS_GPL(cxl_bus_rescan, "CXL");
2117 
cxl_bus_drain(void)2118 void cxl_bus_drain(void)
2119 {
2120 	drain_workqueue(cxl_bus_wq);
2121 }
2122 EXPORT_SYMBOL_NS_GPL(cxl_bus_drain, "CXL");
2123 
schedule_cxl_memdev_detach(struct cxl_memdev * cxlmd)2124 bool schedule_cxl_memdev_detach(struct cxl_memdev *cxlmd)
2125 {
2126 	return queue_work(cxl_bus_wq, &cxlmd->detach_work);
2127 }
2128 EXPORT_SYMBOL_NS_GPL(schedule_cxl_memdev_detach, "CXL");
2129 
add_latency(struct access_coordinate * c,long latency)2130 static void add_latency(struct access_coordinate *c, long latency)
2131 {
2132 	for (int i = 0; i < ACCESS_COORDINATE_MAX; i++) {
2133 		c[i].write_latency += latency;
2134 		c[i].read_latency += latency;
2135 	}
2136 }
2137 
coordinates_valid(struct access_coordinate * c)2138 static bool coordinates_valid(struct access_coordinate *c)
2139 {
2140 	for (int i = 0; i < ACCESS_COORDINATE_MAX; i++) {
2141 		if (c[i].read_bandwidth && c[i].write_bandwidth &&
2142 		    c[i].read_latency && c[i].write_latency)
2143 			continue;
2144 		return false;
2145 	}
2146 
2147 	return true;
2148 }
2149 
set_min_bandwidth(struct access_coordinate * c,unsigned int bw)2150 static void set_min_bandwidth(struct access_coordinate *c, unsigned int bw)
2151 {
2152 	for (int i = 0; i < ACCESS_COORDINATE_MAX; i++) {
2153 		c[i].write_bandwidth = min(c[i].write_bandwidth, bw);
2154 		c[i].read_bandwidth = min(c[i].read_bandwidth, bw);
2155 	}
2156 }
2157 
set_access_coordinates(struct access_coordinate * out,struct access_coordinate * in)2158 static void set_access_coordinates(struct access_coordinate *out,
2159 				   struct access_coordinate *in)
2160 {
2161 	for (int i = 0; i < ACCESS_COORDINATE_MAX; i++)
2162 		out[i] = in[i];
2163 }
2164 
parent_port_is_cxl_root(struct cxl_port * port)2165 static bool parent_port_is_cxl_root(struct cxl_port *port)
2166 {
2167 	return is_cxl_root(to_cxl_port(port->dev.parent));
2168 }
2169 
2170 /**
2171  * cxl_endpoint_get_perf_coordinates - Retrieve performance numbers stored in dports
2172  *				   of CXL path
2173  * @port: endpoint cxl_port
2174  * @coord: output performance data
2175  *
2176  * Return: errno on failure, 0 on success.
2177  */
cxl_endpoint_get_perf_coordinates(struct cxl_port * port,struct access_coordinate * coord)2178 int cxl_endpoint_get_perf_coordinates(struct cxl_port *port,
2179 				      struct access_coordinate *coord)
2180 {
2181 	struct cxl_memdev *cxlmd = to_cxl_memdev(port->uport_dev);
2182 	struct access_coordinate c[] = {
2183 		{
2184 			.read_bandwidth = UINT_MAX,
2185 			.write_bandwidth = UINT_MAX,
2186 		},
2187 		{
2188 			.read_bandwidth = UINT_MAX,
2189 			.write_bandwidth = UINT_MAX,
2190 		},
2191 	};
2192 	struct cxl_port *iter = port;
2193 	struct cxl_dport *dport;
2194 	struct pci_dev *pdev;
2195 	struct device *dev;
2196 	unsigned int bw;
2197 	bool is_cxl_root;
2198 
2199 	if (!is_cxl_endpoint(port))
2200 		return -EINVAL;
2201 
2202 	/*
2203 	 * Skip calculation for RCD. Expectation is HMAT already covers RCD case
2204 	 * since RCH does not support hotplug.
2205 	 */
2206 	if (cxlmd->cxlds->rcd)
2207 		return 0;
2208 
2209 	/*
2210 	 * Exit the loop when the parent port of the current iter port is cxl
2211 	 * root. The iterative loop starts at the endpoint and gathers the
2212 	 * latency of the CXL link from the current device/port to the connected
2213 	 * downstream port each iteration.
2214 	 */
2215 	do {
2216 		dport = iter->parent_dport;
2217 		iter = to_cxl_port(iter->dev.parent);
2218 		is_cxl_root = parent_port_is_cxl_root(iter);
2219 
2220 		/*
2221 		 * There's no valid access_coordinate for a root port since RPs do not
2222 		 * have CDAT and therefore needs to be skipped.
2223 		 */
2224 		if (!is_cxl_root) {
2225 			if (!coordinates_valid(dport->coord))
2226 				return -EINVAL;
2227 			cxl_coordinates_combine(c, c, dport->coord);
2228 		}
2229 		add_latency(c, dport->link_latency);
2230 	} while (!is_cxl_root);
2231 
2232 	dport = iter->parent_dport;
2233 	/* Retrieve HB coords */
2234 	if (!coordinates_valid(dport->coord))
2235 		return -EINVAL;
2236 	cxl_coordinates_combine(c, c, dport->coord);
2237 
2238 	dev = port->uport_dev->parent;
2239 	if (!dev_is_pci(dev))
2240 		return -ENODEV;
2241 
2242 	/* Get the calculated PCI paths bandwidth */
2243 	pdev = to_pci_dev(dev);
2244 	bw = pcie_bandwidth_available(pdev, NULL, NULL, NULL);
2245 	if (bw == 0)
2246 		return -ENXIO;
2247 	bw /= BITS_PER_BYTE;
2248 
2249 	set_min_bandwidth(c, bw);
2250 	set_access_coordinates(coord, c);
2251 
2252 	return 0;
2253 }
2254 EXPORT_SYMBOL_NS_GPL(cxl_endpoint_get_perf_coordinates, "CXL");
2255 
cxl_port_get_switch_dport_bandwidth(struct cxl_port * port,struct access_coordinate * c)2256 int cxl_port_get_switch_dport_bandwidth(struct cxl_port *port,
2257 					struct access_coordinate *c)
2258 {
2259 	struct cxl_dport *dport = port->parent_dport;
2260 
2261 	/* Check this port is connected to a switch DSP and not an RP */
2262 	if (parent_port_is_cxl_root(to_cxl_port(port->dev.parent)))
2263 		return -ENODEV;
2264 
2265 	if (!coordinates_valid(dport->coord))
2266 		return -EINVAL;
2267 
2268 	for (int i = 0; i < ACCESS_COORDINATE_MAX; i++) {
2269 		c[i].read_bandwidth = dport->coord[i].read_bandwidth;
2270 		c[i].write_bandwidth = dport->coord[i].write_bandwidth;
2271 	}
2272 
2273 	return 0;
2274 }
2275 
2276 /* for user tooling to ensure port disable work has completed */
flush_store(const struct bus_type * bus,const char * buf,size_t count)2277 static ssize_t flush_store(const struct bus_type *bus, const char *buf, size_t count)
2278 {
2279 	if (sysfs_streq(buf, "1")) {
2280 		flush_workqueue(cxl_bus_wq);
2281 		return count;
2282 	}
2283 
2284 	return -EINVAL;
2285 }
2286 
2287 static BUS_ATTR_WO(flush);
2288 
2289 static struct attribute *cxl_bus_attributes[] = {
2290 	&bus_attr_flush.attr,
2291 	NULL,
2292 };
2293 
2294 static struct attribute_group cxl_bus_attribute_group = {
2295 	.attrs = cxl_bus_attributes,
2296 };
2297 
2298 static const struct attribute_group *cxl_bus_attribute_groups[] = {
2299 	&cxl_bus_attribute_group,
2300 	NULL,
2301 };
2302 
2303 struct bus_type cxl_bus_type = {
2304 	.name = "cxl",
2305 	.uevent = cxl_bus_uevent,
2306 	.match = cxl_bus_match,
2307 	.probe = cxl_bus_probe,
2308 	.remove = cxl_bus_remove,
2309 	.bus_groups = cxl_bus_attribute_groups,
2310 };
2311 EXPORT_SYMBOL_NS_GPL(cxl_bus_type, "CXL");
2312 
2313 static struct dentry *cxl_debugfs;
2314 
cxl_debugfs_create_dir(const char * dir)2315 struct dentry *cxl_debugfs_create_dir(const char *dir)
2316 {
2317 	return debugfs_create_dir(dir, cxl_debugfs);
2318 }
2319 EXPORT_SYMBOL_NS_GPL(cxl_debugfs_create_dir, "CXL");
2320 
cxl_core_init(void)2321 static __init int cxl_core_init(void)
2322 {
2323 	int rc;
2324 
2325 	cxl_debugfs = debugfs_create_dir("cxl", NULL);
2326 
2327 	if (einj_cxl_is_initialized())
2328 		debugfs_create_file("einj_types", 0400, cxl_debugfs, NULL,
2329 				    &einj_cxl_available_error_type_fops);
2330 
2331 	cxl_mbox_init();
2332 
2333 	rc = cxl_memdev_init();
2334 	if (rc)
2335 		return rc;
2336 
2337 	cxl_bus_wq = alloc_ordered_workqueue("cxl_port", 0);
2338 	if (!cxl_bus_wq) {
2339 		rc = -ENOMEM;
2340 		goto err_wq;
2341 	}
2342 
2343 	rc = bus_register(&cxl_bus_type);
2344 	if (rc)
2345 		goto err_bus;
2346 
2347 	rc = cxl_region_init();
2348 	if (rc)
2349 		goto err_region;
2350 
2351 	rc = cxl_ras_init();
2352 	if (rc)
2353 		goto err_ras;
2354 
2355 	return 0;
2356 
2357 err_ras:
2358 	cxl_region_exit();
2359 err_region:
2360 	bus_unregister(&cxl_bus_type);
2361 err_bus:
2362 	destroy_workqueue(cxl_bus_wq);
2363 err_wq:
2364 	cxl_memdev_exit();
2365 	return rc;
2366 }
2367 
cxl_core_exit(void)2368 static void cxl_core_exit(void)
2369 {
2370 	cxl_ras_exit();
2371 	cxl_region_exit();
2372 	bus_unregister(&cxl_bus_type);
2373 	destroy_workqueue(cxl_bus_wq);
2374 	cxl_memdev_exit();
2375 	debugfs_remove_recursive(cxl_debugfs);
2376 }
2377 
2378 subsys_initcall(cxl_core_init);
2379 module_exit(cxl_core_exit);
2380 MODULE_DESCRIPTION("CXL: Core Compute Express Link support");
2381 MODULE_LICENSE("GPL v2");
2382 MODULE_IMPORT_NS("CXL");
2383