xref: /linux/arch/s390/pci/pci.c (revision 83bce9c2baa51e439480a713119a73d3c8b61083)
1 /*
2  * Copyright IBM Corp. 2012
3  *
4  * Author(s):
5  *   Jan Glauber <jang@linux.vnet.ibm.com>
6  *
7  * The System z PCI code is a rewrite from a prototype by
8  * the following people (Kudoz!):
9  *   Alexander Schmidt
10  *   Christoph Raisch
11  *   Hannes Hering
12  *   Hoang-Nam Nguyen
13  *   Jan-Bernd Themann
14  *   Stefan Roscher
15  *   Thomas Klein
16  */
17 
18 #define KMSG_COMPONENT "zpci"
19 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
20 
21 #include <linux/kernel.h>
22 #include <linux/slab.h>
23 #include <linux/err.h>
24 #include <linux/export.h>
25 #include <linux/delay.h>
26 #include <linux/irq.h>
27 #include <linux/kernel_stat.h>
28 #include <linux/seq_file.h>
29 #include <linux/pci.h>
30 #include <linux/msi.h>
31 
32 #include <asm/isc.h>
33 #include <asm/airq.h>
34 #include <asm/facility.h>
35 #include <asm/pci_insn.h>
36 #include <asm/pci_clp.h>
37 #include <asm/pci_dma.h>
38 
39 #define DEBUG				/* enable pr_debug */
40 
41 #define	SIC_IRQ_MODE_ALL		0
42 #define	SIC_IRQ_MODE_SINGLE		1
43 
44 #define ZPCI_NR_DMA_SPACES		1
45 #define ZPCI_NR_DEVICES			CONFIG_PCI_NR_FUNCTIONS
46 
47 /* list of all detected zpci devices */
48 static LIST_HEAD(zpci_list);
49 static DEFINE_SPINLOCK(zpci_list_lock);
50 
51 static struct irq_chip zpci_irq_chip = {
52 	.name = "zPCI",
53 	.irq_unmask = pci_msi_unmask_irq,
54 	.irq_mask = pci_msi_mask_irq,
55 };
56 
57 static DECLARE_BITMAP(zpci_domain, ZPCI_NR_DEVICES);
58 static DEFINE_SPINLOCK(zpci_domain_lock);
59 
60 static struct airq_iv *zpci_aisb_iv;
61 static struct airq_iv *zpci_aibv[ZPCI_NR_DEVICES];
62 
63 /* Adapter interrupt definitions */
64 static void zpci_irq_handler(struct airq_struct *airq);
65 
66 static struct airq_struct zpci_airq = {
67 	.handler = zpci_irq_handler,
68 	.isc = PCI_ISC,
69 };
70 
71 #define ZPCI_IOMAP_ENTRIES						\
72 	min(((unsigned long) CONFIG_PCI_NR_FUNCTIONS * PCI_BAR_COUNT),	\
73 	    ZPCI_IOMAP_MAX_ENTRIES)
74 
75 static DEFINE_SPINLOCK(zpci_iomap_lock);
76 static unsigned long *zpci_iomap_bitmap;
77 struct zpci_iomap_entry *zpci_iomap_start;
78 EXPORT_SYMBOL_GPL(zpci_iomap_start);
79 
80 static struct kmem_cache *zdev_fmb_cache;
81 
82 struct zpci_dev *get_zdev_by_fid(u32 fid)
83 {
84 	struct zpci_dev *tmp, *zdev = NULL;
85 
86 	spin_lock(&zpci_list_lock);
87 	list_for_each_entry(tmp, &zpci_list, entry) {
88 		if (tmp->fid == fid) {
89 			zdev = tmp;
90 			break;
91 		}
92 	}
93 	spin_unlock(&zpci_list_lock);
94 	return zdev;
95 }
96 
97 static struct zpci_dev *get_zdev_by_bus(struct pci_bus *bus)
98 {
99 	return (bus && bus->sysdata) ? (struct zpci_dev *) bus->sysdata : NULL;
100 }
101 
102 int pci_domain_nr(struct pci_bus *bus)
103 {
104 	return ((struct zpci_dev *) bus->sysdata)->domain;
105 }
106 EXPORT_SYMBOL_GPL(pci_domain_nr);
107 
108 int pci_proc_domain(struct pci_bus *bus)
109 {
110 	return pci_domain_nr(bus);
111 }
112 EXPORT_SYMBOL_GPL(pci_proc_domain);
113 
114 /* Modify PCI: Register adapter interruptions */
115 static int zpci_set_airq(struct zpci_dev *zdev)
116 {
117 	u64 req = ZPCI_CREATE_REQ(zdev->fh, 0, ZPCI_MOD_FC_REG_INT);
118 	struct zpci_fib fib = {0};
119 
120 	fib.isc = PCI_ISC;
121 	fib.sum = 1;		/* enable summary notifications */
122 	fib.noi = airq_iv_end(zdev->aibv);
123 	fib.aibv = (unsigned long) zdev->aibv->vector;
124 	fib.aibvo = 0;		/* each zdev has its own interrupt vector */
125 	fib.aisb = (unsigned long) zpci_aisb_iv->vector + (zdev->aisb/64)*8;
126 	fib.aisbo = zdev->aisb & 63;
127 
128 	return zpci_mod_fc(req, &fib);
129 }
130 
131 struct mod_pci_args {
132 	u64 base;
133 	u64 limit;
134 	u64 iota;
135 	u64 fmb_addr;
136 };
137 
138 static int mod_pci(struct zpci_dev *zdev, int fn, u8 dmaas, struct mod_pci_args *args)
139 {
140 	u64 req = ZPCI_CREATE_REQ(zdev->fh, dmaas, fn);
141 	struct zpci_fib fib = {0};
142 
143 	fib.pba = args->base;
144 	fib.pal = args->limit;
145 	fib.iota = args->iota;
146 	fib.fmb_addr = args->fmb_addr;
147 
148 	return zpci_mod_fc(req, &fib);
149 }
150 
151 /* Modify PCI: Register I/O address translation parameters */
152 int zpci_register_ioat(struct zpci_dev *zdev, u8 dmaas,
153 		       u64 base, u64 limit, u64 iota)
154 {
155 	struct mod_pci_args args = { base, limit, iota, 0 };
156 
157 	WARN_ON_ONCE(iota & 0x3fff);
158 	args.iota |= ZPCI_IOTA_RTTO_FLAG;
159 	return mod_pci(zdev, ZPCI_MOD_FC_REG_IOAT, dmaas, &args);
160 }
161 
162 /* Modify PCI: Unregister I/O address translation parameters */
163 int zpci_unregister_ioat(struct zpci_dev *zdev, u8 dmaas)
164 {
165 	struct mod_pci_args args = { 0, 0, 0, 0 };
166 
167 	return mod_pci(zdev, ZPCI_MOD_FC_DEREG_IOAT, dmaas, &args);
168 }
169 
170 /* Modify PCI: Unregister adapter interruptions */
171 static int zpci_clear_airq(struct zpci_dev *zdev)
172 {
173 	struct mod_pci_args args = { 0, 0, 0, 0 };
174 
175 	return mod_pci(zdev, ZPCI_MOD_FC_DEREG_INT, 0, &args);
176 }
177 
178 /* Modify PCI: Set PCI function measurement parameters */
179 int zpci_fmb_enable_device(struct zpci_dev *zdev)
180 {
181 	struct mod_pci_args args = { 0, 0, 0, 0 };
182 
183 	if (zdev->fmb || sizeof(*zdev->fmb) < zdev->fmb_length)
184 		return -EINVAL;
185 
186 	zdev->fmb = kmem_cache_zalloc(zdev_fmb_cache, GFP_KERNEL);
187 	if (!zdev->fmb)
188 		return -ENOMEM;
189 	WARN_ON((u64) zdev->fmb & 0xf);
190 
191 	/* reset software counters */
192 	atomic64_set(&zdev->allocated_pages, 0);
193 	atomic64_set(&zdev->mapped_pages, 0);
194 	atomic64_set(&zdev->unmapped_pages, 0);
195 
196 	args.fmb_addr = virt_to_phys(zdev->fmb);
197 	return mod_pci(zdev, ZPCI_MOD_FC_SET_MEASURE, 0, &args);
198 }
199 
200 /* Modify PCI: Disable PCI function measurement */
201 int zpci_fmb_disable_device(struct zpci_dev *zdev)
202 {
203 	struct mod_pci_args args = { 0, 0, 0, 0 };
204 	int rc;
205 
206 	if (!zdev->fmb)
207 		return -EINVAL;
208 
209 	/* Function measurement is disabled if fmb address is zero */
210 	rc = mod_pci(zdev, ZPCI_MOD_FC_SET_MEASURE, 0, &args);
211 
212 	kmem_cache_free(zdev_fmb_cache, zdev->fmb);
213 	zdev->fmb = NULL;
214 	return rc;
215 }
216 
217 #define ZPCI_PCIAS_CFGSPC	15
218 
219 static int zpci_cfg_load(struct zpci_dev *zdev, int offset, u32 *val, u8 len)
220 {
221 	u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len);
222 	u64 data;
223 	int rc;
224 
225 	rc = zpci_load(&data, req, offset);
226 	if (!rc) {
227 		data = le64_to_cpu((__force __le64) data);
228 		data >>= (8 - len) * 8;
229 		*val = (u32) data;
230 	} else
231 		*val = 0xffffffff;
232 	return rc;
233 }
234 
235 static int zpci_cfg_store(struct zpci_dev *zdev, int offset, u32 val, u8 len)
236 {
237 	u64 req = ZPCI_CREATE_REQ(zdev->fh, ZPCI_PCIAS_CFGSPC, len);
238 	u64 data = val;
239 	int rc;
240 
241 	data <<= (8 - len) * 8;
242 	data = (__force u64) cpu_to_le64(data);
243 	rc = zpci_store(data, req, offset);
244 	return rc;
245 }
246 
247 void pcibios_fixup_bus(struct pci_bus *bus)
248 {
249 }
250 
251 resource_size_t pcibios_align_resource(void *data, const struct resource *res,
252 				       resource_size_t size,
253 				       resource_size_t align)
254 {
255 	return 0;
256 }
257 
258 /* combine single writes by using store-block insn */
259 void __iowrite64_copy(void __iomem *to, const void *from, size_t count)
260 {
261        zpci_memcpy_toio(to, from, count);
262 }
263 
264 /* Create a virtual mapping cookie for a PCI BAR */
265 void __iomem *pci_iomap_range(struct pci_dev *pdev,
266 			      int bar,
267 			      unsigned long offset,
268 			      unsigned long max)
269 {
270 	struct zpci_dev *zdev =	to_zpci(pdev);
271 	int idx;
272 
273 	if (!pci_resource_len(pdev, bar))
274 		return NULL;
275 
276 	idx = zdev->bars[bar].map_idx;
277 	spin_lock(&zpci_iomap_lock);
278 	/* Detect overrun */
279 	WARN_ON(!++zpci_iomap_start[idx].count);
280 	zpci_iomap_start[idx].fh = zdev->fh;
281 	zpci_iomap_start[idx].bar = bar;
282 	spin_unlock(&zpci_iomap_lock);
283 
284 	return (void __iomem *) ZPCI_ADDR(idx) + offset;
285 }
286 EXPORT_SYMBOL(pci_iomap_range);
287 
288 void __iomem *pci_iomap(struct pci_dev *dev, int bar, unsigned long maxlen)
289 {
290 	return pci_iomap_range(dev, bar, 0, maxlen);
291 }
292 EXPORT_SYMBOL(pci_iomap);
293 
294 void pci_iounmap(struct pci_dev *pdev, void __iomem *addr)
295 {
296 	unsigned int idx = ZPCI_IDX(addr);
297 
298 	spin_lock(&zpci_iomap_lock);
299 	/* Detect underrun */
300 	WARN_ON(!zpci_iomap_start[idx].count);
301 	if (!--zpci_iomap_start[idx].count) {
302 		zpci_iomap_start[idx].fh = 0;
303 		zpci_iomap_start[idx].bar = 0;
304 	}
305 	spin_unlock(&zpci_iomap_lock);
306 }
307 EXPORT_SYMBOL(pci_iounmap);
308 
309 static int pci_read(struct pci_bus *bus, unsigned int devfn, int where,
310 		    int size, u32 *val)
311 {
312 	struct zpci_dev *zdev = get_zdev_by_bus(bus);
313 	int ret;
314 
315 	if (!zdev || devfn != ZPCI_DEVFN)
316 		ret = -ENODEV;
317 	else
318 		ret = zpci_cfg_load(zdev, where, val, size);
319 
320 	return ret;
321 }
322 
323 static int pci_write(struct pci_bus *bus, unsigned int devfn, int where,
324 		     int size, u32 val)
325 {
326 	struct zpci_dev *zdev = get_zdev_by_bus(bus);
327 	int ret;
328 
329 	if (!zdev || devfn != ZPCI_DEVFN)
330 		ret = -ENODEV;
331 	else
332 		ret = zpci_cfg_store(zdev, where, val, size);
333 
334 	return ret;
335 }
336 
337 static struct pci_ops pci_root_ops = {
338 	.read = pci_read,
339 	.write = pci_write,
340 };
341 
342 static void zpci_irq_handler(struct airq_struct *airq)
343 {
344 	unsigned long si, ai;
345 	struct airq_iv *aibv;
346 	int irqs_on = 0;
347 
348 	inc_irq_stat(IRQIO_PCI);
349 	for (si = 0;;) {
350 		/* Scan adapter summary indicator bit vector */
351 		si = airq_iv_scan(zpci_aisb_iv, si, airq_iv_end(zpci_aisb_iv));
352 		if (si == -1UL) {
353 			if (irqs_on++)
354 				/* End of second scan with interrupts on. */
355 				break;
356 			/* First scan complete, reenable interrupts. */
357 			zpci_set_irq_ctrl(SIC_IRQ_MODE_SINGLE, NULL, PCI_ISC);
358 			si = 0;
359 			continue;
360 		}
361 
362 		/* Scan the adapter interrupt vector for this device. */
363 		aibv = zpci_aibv[si];
364 		for (ai = 0;;) {
365 			ai = airq_iv_scan(aibv, ai, airq_iv_end(aibv));
366 			if (ai == -1UL)
367 				break;
368 			inc_irq_stat(IRQIO_MSI);
369 			airq_iv_lock(aibv, ai);
370 			generic_handle_irq(airq_iv_get_data(aibv, ai));
371 			airq_iv_unlock(aibv, ai);
372 		}
373 	}
374 }
375 
376 int arch_setup_msi_irqs(struct pci_dev *pdev, int nvec, int type)
377 {
378 	struct zpci_dev *zdev = to_zpci(pdev);
379 	unsigned int hwirq, msi_vecs;
380 	unsigned long aisb;
381 	struct msi_desc *msi;
382 	struct msi_msg msg;
383 	int rc, irq;
384 
385 	if (type == PCI_CAP_ID_MSI && nvec > 1)
386 		return 1;
387 	msi_vecs = min_t(unsigned int, nvec, zdev->max_msi);
388 
389 	/* Allocate adapter summary indicator bit */
390 	rc = -EIO;
391 	aisb = airq_iv_alloc_bit(zpci_aisb_iv);
392 	if (aisb == -1UL)
393 		goto out;
394 	zdev->aisb = aisb;
395 
396 	/* Create adapter interrupt vector */
397 	rc = -ENOMEM;
398 	zdev->aibv = airq_iv_create(msi_vecs, AIRQ_IV_DATA | AIRQ_IV_BITLOCK);
399 	if (!zdev->aibv)
400 		goto out_si;
401 
402 	/* Wire up shortcut pointer */
403 	zpci_aibv[aisb] = zdev->aibv;
404 
405 	/* Request MSI interrupts */
406 	hwirq = 0;
407 	for_each_pci_msi_entry(msi, pdev) {
408 		rc = -EIO;
409 		irq = irq_alloc_desc(0);	/* Alloc irq on node 0 */
410 		if (irq < 0)
411 			goto out_msi;
412 		rc = irq_set_msi_desc(irq, msi);
413 		if (rc)
414 			goto out_msi;
415 		irq_set_chip_and_handler(irq, &zpci_irq_chip,
416 					 handle_simple_irq);
417 		msg.data = hwirq;
418 		msg.address_lo = zdev->msi_addr & 0xffffffff;
419 		msg.address_hi = zdev->msi_addr >> 32;
420 		pci_write_msi_msg(irq, &msg);
421 		airq_iv_set_data(zdev->aibv, hwirq, irq);
422 		hwirq++;
423 	}
424 
425 	/* Enable adapter interrupts */
426 	rc = zpci_set_airq(zdev);
427 	if (rc)
428 		goto out_msi;
429 
430 	return (msi_vecs == nvec) ? 0 : msi_vecs;
431 
432 out_msi:
433 	for_each_pci_msi_entry(msi, pdev) {
434 		if (hwirq-- == 0)
435 			break;
436 		irq_set_msi_desc(msi->irq, NULL);
437 		irq_free_desc(msi->irq);
438 		msi->msg.address_lo = 0;
439 		msi->msg.address_hi = 0;
440 		msi->msg.data = 0;
441 		msi->irq = 0;
442 	}
443 	zpci_aibv[aisb] = NULL;
444 	airq_iv_release(zdev->aibv);
445 out_si:
446 	airq_iv_free_bit(zpci_aisb_iv, aisb);
447 out:
448 	return rc;
449 }
450 
451 void arch_teardown_msi_irqs(struct pci_dev *pdev)
452 {
453 	struct zpci_dev *zdev = to_zpci(pdev);
454 	struct msi_desc *msi;
455 	int rc;
456 
457 	/* Disable adapter interrupts */
458 	rc = zpci_clear_airq(zdev);
459 	if (rc)
460 		return;
461 
462 	/* Release MSI interrupts */
463 	for_each_pci_msi_entry(msi, pdev) {
464 		if (msi->msi_attrib.is_msix)
465 			__pci_msix_desc_mask_irq(msi, 1);
466 		else
467 			__pci_msi_desc_mask_irq(msi, 1, 1);
468 		irq_set_msi_desc(msi->irq, NULL);
469 		irq_free_desc(msi->irq);
470 		msi->msg.address_lo = 0;
471 		msi->msg.address_hi = 0;
472 		msi->msg.data = 0;
473 		msi->irq = 0;
474 	}
475 
476 	zpci_aibv[zdev->aisb] = NULL;
477 	airq_iv_release(zdev->aibv);
478 	airq_iv_free_bit(zpci_aisb_iv, zdev->aisb);
479 }
480 
481 static void zpci_map_resources(struct pci_dev *pdev)
482 {
483 	resource_size_t len;
484 	int i;
485 
486 	for (i = 0; i < PCI_BAR_COUNT; i++) {
487 		len = pci_resource_len(pdev, i);
488 		if (!len)
489 			continue;
490 		pdev->resource[i].start =
491 			(resource_size_t __force) pci_iomap(pdev, i, 0);
492 		pdev->resource[i].end = pdev->resource[i].start + len - 1;
493 	}
494 }
495 
496 static void zpci_unmap_resources(struct pci_dev *pdev)
497 {
498 	resource_size_t len;
499 	int i;
500 
501 	for (i = 0; i < PCI_BAR_COUNT; i++) {
502 		len = pci_resource_len(pdev, i);
503 		if (!len)
504 			continue;
505 		pci_iounmap(pdev, (void __iomem __force *)
506 			    pdev->resource[i].start);
507 	}
508 }
509 
510 static int __init zpci_irq_init(void)
511 {
512 	int rc;
513 
514 	rc = register_adapter_interrupt(&zpci_airq);
515 	if (rc)
516 		goto out;
517 	/* Set summary to 1 to be called every time for the ISC. */
518 	*zpci_airq.lsi_ptr = 1;
519 
520 	rc = -ENOMEM;
521 	zpci_aisb_iv = airq_iv_create(ZPCI_NR_DEVICES, AIRQ_IV_ALLOC);
522 	if (!zpci_aisb_iv)
523 		goto out_airq;
524 
525 	zpci_set_irq_ctrl(SIC_IRQ_MODE_SINGLE, NULL, PCI_ISC);
526 	return 0;
527 
528 out_airq:
529 	unregister_adapter_interrupt(&zpci_airq);
530 out:
531 	return rc;
532 }
533 
534 static void zpci_irq_exit(void)
535 {
536 	airq_iv_release(zpci_aisb_iv);
537 	unregister_adapter_interrupt(&zpci_airq);
538 }
539 
540 static int zpci_alloc_iomap(struct zpci_dev *zdev)
541 {
542 	unsigned long entry;
543 
544 	spin_lock(&zpci_iomap_lock);
545 	entry = find_first_zero_bit(zpci_iomap_bitmap, ZPCI_IOMAP_ENTRIES);
546 	if (entry == ZPCI_IOMAP_ENTRIES) {
547 		spin_unlock(&zpci_iomap_lock);
548 		return -ENOSPC;
549 	}
550 	set_bit(entry, zpci_iomap_bitmap);
551 	spin_unlock(&zpci_iomap_lock);
552 	return entry;
553 }
554 
555 static void zpci_free_iomap(struct zpci_dev *zdev, int entry)
556 {
557 	spin_lock(&zpci_iomap_lock);
558 	memset(&zpci_iomap_start[entry], 0, sizeof(struct zpci_iomap_entry));
559 	clear_bit(entry, zpci_iomap_bitmap);
560 	spin_unlock(&zpci_iomap_lock);
561 }
562 
563 static struct resource *__alloc_res(struct zpci_dev *zdev, unsigned long start,
564 				    unsigned long size, unsigned long flags)
565 {
566 	struct resource *r;
567 
568 	r = kzalloc(sizeof(*r), GFP_KERNEL);
569 	if (!r)
570 		return NULL;
571 
572 	r->start = start;
573 	r->end = r->start + size - 1;
574 	r->flags = flags;
575 	r->name = zdev->res_name;
576 
577 	if (request_resource(&iomem_resource, r)) {
578 		kfree(r);
579 		return NULL;
580 	}
581 	return r;
582 }
583 
584 static int zpci_setup_bus_resources(struct zpci_dev *zdev,
585 				    struct list_head *resources)
586 {
587 	unsigned long addr, size, flags;
588 	struct resource *res;
589 	int i, entry;
590 
591 	snprintf(zdev->res_name, sizeof(zdev->res_name),
592 		 "PCI Bus %04x:%02x", zdev->domain, ZPCI_BUS_NR);
593 
594 	for (i = 0; i < PCI_BAR_COUNT; i++) {
595 		if (!zdev->bars[i].size)
596 			continue;
597 		entry = zpci_alloc_iomap(zdev);
598 		if (entry < 0)
599 			return entry;
600 		zdev->bars[i].map_idx = entry;
601 
602 		/* only MMIO is supported */
603 		flags = IORESOURCE_MEM;
604 		if (zdev->bars[i].val & 8)
605 			flags |= IORESOURCE_PREFETCH;
606 		if (zdev->bars[i].val & 4)
607 			flags |= IORESOURCE_MEM_64;
608 
609 		addr = ZPCI_ADDR(entry);
610 		size = 1UL << zdev->bars[i].size;
611 
612 		res = __alloc_res(zdev, addr, size, flags);
613 		if (!res) {
614 			zpci_free_iomap(zdev, entry);
615 			return -ENOMEM;
616 		}
617 		zdev->bars[i].res = res;
618 		pci_add_resource(resources, res);
619 	}
620 
621 	return 0;
622 }
623 
624 static void zpci_cleanup_bus_resources(struct zpci_dev *zdev)
625 {
626 	int i;
627 
628 	for (i = 0; i < PCI_BAR_COUNT; i++) {
629 		if (!zdev->bars[i].size || !zdev->bars[i].res)
630 			continue;
631 
632 		zpci_free_iomap(zdev, zdev->bars[i].map_idx);
633 		release_resource(zdev->bars[i].res);
634 		kfree(zdev->bars[i].res);
635 	}
636 }
637 
638 int pcibios_add_device(struct pci_dev *pdev)
639 {
640 	struct resource *res;
641 	int i;
642 
643 	pdev->dev.groups = zpci_attr_groups;
644 	pdev->dev.dma_ops = &s390_pci_dma_ops;
645 	zpci_map_resources(pdev);
646 
647 	for (i = 0; i < PCI_BAR_COUNT; i++) {
648 		res = &pdev->resource[i];
649 		if (res->parent || !res->flags)
650 			continue;
651 		pci_claim_resource(pdev, i);
652 	}
653 
654 	return 0;
655 }
656 
657 void pcibios_release_device(struct pci_dev *pdev)
658 {
659 	zpci_unmap_resources(pdev);
660 }
661 
662 int pcibios_enable_device(struct pci_dev *pdev, int mask)
663 {
664 	struct zpci_dev *zdev = to_zpci(pdev);
665 
666 	zpci_debug_init_device(zdev, dev_name(&pdev->dev));
667 	zpci_fmb_enable_device(zdev);
668 
669 	return pci_enable_resources(pdev, mask);
670 }
671 
672 void pcibios_disable_device(struct pci_dev *pdev)
673 {
674 	struct zpci_dev *zdev = to_zpci(pdev);
675 
676 	zpci_fmb_disable_device(zdev);
677 	zpci_debug_exit_device(zdev);
678 }
679 
680 #ifdef CONFIG_HIBERNATE_CALLBACKS
681 static int zpci_restore(struct device *dev)
682 {
683 	struct pci_dev *pdev = to_pci_dev(dev);
684 	struct zpci_dev *zdev = to_zpci(pdev);
685 	int ret = 0;
686 
687 	if (zdev->state != ZPCI_FN_STATE_ONLINE)
688 		goto out;
689 
690 	ret = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES);
691 	if (ret)
692 		goto out;
693 
694 	zpci_map_resources(pdev);
695 	zpci_register_ioat(zdev, 0, zdev->start_dma, zdev->end_dma,
696 			   (u64) zdev->dma_table);
697 
698 out:
699 	return ret;
700 }
701 
702 static int zpci_freeze(struct device *dev)
703 {
704 	struct pci_dev *pdev = to_pci_dev(dev);
705 	struct zpci_dev *zdev = to_zpci(pdev);
706 
707 	if (zdev->state != ZPCI_FN_STATE_ONLINE)
708 		return 0;
709 
710 	zpci_unregister_ioat(zdev, 0);
711 	zpci_unmap_resources(pdev);
712 	return clp_disable_fh(zdev);
713 }
714 
715 struct dev_pm_ops pcibios_pm_ops = {
716 	.thaw_noirq = zpci_restore,
717 	.freeze_noirq = zpci_freeze,
718 	.restore_noirq = zpci_restore,
719 	.poweroff_noirq = zpci_freeze,
720 };
721 #endif /* CONFIG_HIBERNATE_CALLBACKS */
722 
723 static int zpci_alloc_domain(struct zpci_dev *zdev)
724 {
725 	if (zpci_unique_uid) {
726 		zdev->domain = (u16) zdev->uid;
727 		return 0;
728 	}
729 
730 	spin_lock(&zpci_domain_lock);
731 	zdev->domain = find_first_zero_bit(zpci_domain, ZPCI_NR_DEVICES);
732 	if (zdev->domain == ZPCI_NR_DEVICES) {
733 		spin_unlock(&zpci_domain_lock);
734 		return -ENOSPC;
735 	}
736 	set_bit(zdev->domain, zpci_domain);
737 	spin_unlock(&zpci_domain_lock);
738 	return 0;
739 }
740 
741 static void zpci_free_domain(struct zpci_dev *zdev)
742 {
743 	if (zpci_unique_uid)
744 		return;
745 
746 	spin_lock(&zpci_domain_lock);
747 	clear_bit(zdev->domain, zpci_domain);
748 	spin_unlock(&zpci_domain_lock);
749 }
750 
751 void pcibios_remove_bus(struct pci_bus *bus)
752 {
753 	struct zpci_dev *zdev = get_zdev_by_bus(bus);
754 
755 	zpci_exit_slot(zdev);
756 	zpci_cleanup_bus_resources(zdev);
757 	zpci_free_domain(zdev);
758 
759 	spin_lock(&zpci_list_lock);
760 	list_del(&zdev->entry);
761 	spin_unlock(&zpci_list_lock);
762 
763 	kfree(zdev);
764 }
765 
766 static int zpci_scan_bus(struct zpci_dev *zdev)
767 {
768 	LIST_HEAD(resources);
769 	int ret;
770 
771 	ret = zpci_setup_bus_resources(zdev, &resources);
772 	if (ret)
773 		goto error;
774 
775 	zdev->bus = pci_scan_root_bus(NULL, ZPCI_BUS_NR, &pci_root_ops,
776 				      zdev, &resources);
777 	if (!zdev->bus) {
778 		ret = -EIO;
779 		goto error;
780 	}
781 	zdev->bus->max_bus_speed = zdev->max_bus_speed;
782 	pci_bus_add_devices(zdev->bus);
783 	return 0;
784 
785 error:
786 	zpci_cleanup_bus_resources(zdev);
787 	pci_free_resource_list(&resources);
788 	return ret;
789 }
790 
791 int zpci_enable_device(struct zpci_dev *zdev)
792 {
793 	int rc;
794 
795 	rc = clp_enable_fh(zdev, ZPCI_NR_DMA_SPACES);
796 	if (rc)
797 		goto out;
798 
799 	rc = zpci_dma_init_device(zdev);
800 	if (rc)
801 		goto out_dma;
802 
803 	zdev->state = ZPCI_FN_STATE_ONLINE;
804 	return 0;
805 
806 out_dma:
807 	clp_disable_fh(zdev);
808 out:
809 	return rc;
810 }
811 EXPORT_SYMBOL_GPL(zpci_enable_device);
812 
813 int zpci_disable_device(struct zpci_dev *zdev)
814 {
815 	zpci_dma_exit_device(zdev);
816 	return clp_disable_fh(zdev);
817 }
818 EXPORT_SYMBOL_GPL(zpci_disable_device);
819 
820 int zpci_create_device(struct zpci_dev *zdev)
821 {
822 	int rc;
823 
824 	rc = zpci_alloc_domain(zdev);
825 	if (rc)
826 		goto out;
827 
828 	mutex_init(&zdev->lock);
829 	if (zdev->state == ZPCI_FN_STATE_CONFIGURED) {
830 		rc = zpci_enable_device(zdev);
831 		if (rc)
832 			goto out_free;
833 	}
834 	rc = zpci_scan_bus(zdev);
835 	if (rc)
836 		goto out_disable;
837 
838 	spin_lock(&zpci_list_lock);
839 	list_add_tail(&zdev->entry, &zpci_list);
840 	spin_unlock(&zpci_list_lock);
841 
842 	zpci_init_slot(zdev);
843 
844 	return 0;
845 
846 out_disable:
847 	if (zdev->state == ZPCI_FN_STATE_ONLINE)
848 		zpci_disable_device(zdev);
849 out_free:
850 	zpci_free_domain(zdev);
851 out:
852 	return rc;
853 }
854 
855 void zpci_stop_device(struct zpci_dev *zdev)
856 {
857 	zpci_dma_exit_device(zdev);
858 	/*
859 	 * Note: SCLP disables fh via set-pci-fn so don't
860 	 * do that here.
861 	 */
862 }
863 EXPORT_SYMBOL_GPL(zpci_stop_device);
864 
865 int zpci_report_error(struct pci_dev *pdev,
866 		      struct zpci_report_error_header *report)
867 {
868 	struct zpci_dev *zdev = to_zpci(pdev);
869 
870 	return sclp_pci_report(report, zdev->fh, zdev->fid);
871 }
872 EXPORT_SYMBOL(zpci_report_error);
873 
874 static inline int barsize(u8 size)
875 {
876 	return (size) ? (1 << size) >> 10 : 0;
877 }
878 
879 static int zpci_mem_init(void)
880 {
881 	BUILD_BUG_ON(!is_power_of_2(__alignof__(struct zpci_fmb)) ||
882 		     __alignof__(struct zpci_fmb) < sizeof(struct zpci_fmb));
883 
884 	zdev_fmb_cache = kmem_cache_create("PCI_FMB_cache", sizeof(struct zpci_fmb),
885 					   __alignof__(struct zpci_fmb), 0, NULL);
886 	if (!zdev_fmb_cache)
887 		goto error_fmb;
888 
889 	zpci_iomap_start = kcalloc(ZPCI_IOMAP_ENTRIES,
890 				   sizeof(*zpci_iomap_start), GFP_KERNEL);
891 	if (!zpci_iomap_start)
892 		goto error_iomap;
893 
894 	zpci_iomap_bitmap = kcalloc(BITS_TO_LONGS(ZPCI_IOMAP_ENTRIES),
895 				    sizeof(*zpci_iomap_bitmap), GFP_KERNEL);
896 	if (!zpci_iomap_bitmap)
897 		goto error_iomap_bitmap;
898 
899 	return 0;
900 error_iomap_bitmap:
901 	kfree(zpci_iomap_start);
902 error_iomap:
903 	kmem_cache_destroy(zdev_fmb_cache);
904 error_fmb:
905 	return -ENOMEM;
906 }
907 
908 static void zpci_mem_exit(void)
909 {
910 	kfree(zpci_iomap_bitmap);
911 	kfree(zpci_iomap_start);
912 	kmem_cache_destroy(zdev_fmb_cache);
913 }
914 
915 static unsigned int s390_pci_probe = 1;
916 static unsigned int s390_pci_initialized;
917 
918 char * __init pcibios_setup(char *str)
919 {
920 	if (!strcmp(str, "off")) {
921 		s390_pci_probe = 0;
922 		return NULL;
923 	}
924 	return str;
925 }
926 
927 bool zpci_is_enabled(void)
928 {
929 	return s390_pci_initialized;
930 }
931 
932 static int __init pci_base_init(void)
933 {
934 	int rc;
935 
936 	if (!s390_pci_probe)
937 		return 0;
938 
939 	if (!test_facility(69) || !test_facility(71) || !test_facility(72))
940 		return 0;
941 
942 	rc = zpci_debug_init();
943 	if (rc)
944 		goto out;
945 
946 	rc = zpci_mem_init();
947 	if (rc)
948 		goto out_mem;
949 
950 	rc = zpci_irq_init();
951 	if (rc)
952 		goto out_irq;
953 
954 	rc = zpci_dma_init();
955 	if (rc)
956 		goto out_dma;
957 
958 	rc = clp_scan_pci_devices();
959 	if (rc)
960 		goto out_find;
961 
962 	s390_pci_initialized = 1;
963 	return 0;
964 
965 out_find:
966 	zpci_dma_exit();
967 out_dma:
968 	zpci_irq_exit();
969 out_irq:
970 	zpci_mem_exit();
971 out_mem:
972 	zpci_debug_exit();
973 out:
974 	return rc;
975 }
976 subsys_initcall_sync(pci_base_init);
977 
978 void zpci_rescan(void)
979 {
980 	if (zpci_is_enabled())
981 		clp_rescan_pci_devices_simple();
982 }
983