xref: /linux/arch/powerpc/kernel/eeh_pe.c (revision 9d535e200f09ce347afc38c81ec7f2901187e5f0)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * The file intends to implement PE based on the information from
4  * platforms. Basically, there have 3 types of PEs: PHB/Bus/Device.
5  * All the PEs should be organized as hierarchy tree. The first level
6  * of the tree will be associated to existing PHBs since the particular
7  * PE is only meaningful in one PHB domain.
8  *
9  * Copyright Benjamin Herrenschmidt & Gavin Shan, IBM Corporation 2012.
10  */
11 
12 #include <linux/delay.h>
13 #include <linux/export.h>
14 #include <linux/gfp.h>
15 #include <linux/kernel.h>
16 #include <linux/pci.h>
17 #include <linux/string.h>
18 
19 #include <asm/pci-bridge.h>
20 #include <asm/ppc-pci.h>
21 
22 static int eeh_pe_aux_size = 0;
23 static LIST_HEAD(eeh_phb_pe);
24 
25 /**
26  * eeh_set_pe_aux_size - Set PE auxillary data size
27  * @size: PE auxillary data size
28  *
29  * Set PE auxillary data size
30  */
31 void eeh_set_pe_aux_size(int size)
32 {
33 	if (size < 0)
34 		return;
35 
36 	eeh_pe_aux_size = size;
37 }
38 
39 /**
40  * eeh_pe_alloc - Allocate PE
41  * @phb: PCI controller
42  * @type: PE type
43  *
44  * Allocate PE instance dynamically.
45  */
46 static struct eeh_pe *eeh_pe_alloc(struct pci_controller *phb, int type)
47 {
48 	struct eeh_pe *pe;
49 	size_t alloc_size;
50 
51 	alloc_size = sizeof(struct eeh_pe);
52 	if (eeh_pe_aux_size) {
53 		alloc_size = ALIGN(alloc_size, cache_line_size());
54 		alloc_size += eeh_pe_aux_size;
55 	}
56 
57 	/* Allocate PHB PE */
58 	pe = kzalloc(alloc_size, GFP_KERNEL);
59 	if (!pe) return NULL;
60 
61 	/* Initialize PHB PE */
62 	pe->type = type;
63 	pe->phb = phb;
64 	INIT_LIST_HEAD(&pe->child_list);
65 	INIT_LIST_HEAD(&pe->edevs);
66 
67 	pe->data = (void *)pe + ALIGN(sizeof(struct eeh_pe),
68 				      cache_line_size());
69 	return pe;
70 }
71 
72 /**
73  * eeh_phb_pe_create - Create PHB PE
74  * @phb: PCI controller
75  *
76  * The function should be called while the PHB is detected during
77  * system boot or PCI hotplug in order to create PHB PE.
78  */
79 int eeh_phb_pe_create(struct pci_controller *phb)
80 {
81 	struct eeh_pe *pe;
82 
83 	/* Allocate PHB PE */
84 	pe = eeh_pe_alloc(phb, EEH_PE_PHB);
85 	if (!pe) {
86 		pr_err("%s: out of memory!\n", __func__);
87 		return -ENOMEM;
88 	}
89 
90 	/* Put it into the list */
91 	list_add_tail(&pe->child, &eeh_phb_pe);
92 
93 	pr_debug("EEH: Add PE for PHB#%x\n", phb->global_number);
94 
95 	return 0;
96 }
97 
98 /**
99  * eeh_wait_state - Wait for PE state
100  * @pe: EEH PE
101  * @max_wait: maximal period in millisecond
102  *
103  * Wait for the state of associated PE. It might take some time
104  * to retrieve the PE's state.
105  */
106 int eeh_wait_state(struct eeh_pe *pe, int max_wait)
107 {
108 	int ret;
109 	int mwait;
110 
111 	/*
112 	 * According to PAPR, the state of PE might be temporarily
113 	 * unavailable. Under the circumstance, we have to wait
114 	 * for indicated time determined by firmware. The maximal
115 	 * wait time is 5 minutes, which is acquired from the original
116 	 * EEH implementation. Also, the original implementation
117 	 * also defined the minimal wait time as 1 second.
118 	 */
119 #define EEH_STATE_MIN_WAIT_TIME	(1000)
120 #define EEH_STATE_MAX_WAIT_TIME	(300 * 1000)
121 
122 	while (1) {
123 		ret = eeh_ops->get_state(pe, &mwait);
124 
125 		if (ret != EEH_STATE_UNAVAILABLE)
126 			return ret;
127 
128 		if (max_wait <= 0) {
129 			pr_warn("%s: Timeout when getting PE's state (%d)\n",
130 				__func__, max_wait);
131 			return EEH_STATE_NOT_SUPPORT;
132 		}
133 
134 		if (mwait < EEH_STATE_MIN_WAIT_TIME) {
135 			pr_warn("%s: Firmware returned bad wait value %d\n",
136 				__func__, mwait);
137 			mwait = EEH_STATE_MIN_WAIT_TIME;
138 		} else if (mwait > EEH_STATE_MAX_WAIT_TIME) {
139 			pr_warn("%s: Firmware returned too long wait value %d\n",
140 				__func__, mwait);
141 			mwait = EEH_STATE_MAX_WAIT_TIME;
142 		}
143 
144 		msleep(min(mwait, max_wait));
145 		max_wait -= mwait;
146 	}
147 }
148 
149 /**
150  * eeh_phb_pe_get - Retrieve PHB PE based on the given PHB
151  * @phb: PCI controller
152  *
153  * The overall PEs form hierarchy tree. The first layer of the
154  * hierarchy tree is composed of PHB PEs. The function is used
155  * to retrieve the corresponding PHB PE according to the given PHB.
156  */
157 struct eeh_pe *eeh_phb_pe_get(struct pci_controller *phb)
158 {
159 	struct eeh_pe *pe;
160 
161 	list_for_each_entry(pe, &eeh_phb_pe, child) {
162 		/*
163 		 * Actually, we needn't check the type since
164 		 * the PE for PHB has been determined when that
165 		 * was created.
166 		 */
167 		if ((pe->type & EEH_PE_PHB) && pe->phb == phb)
168 			return pe;
169 	}
170 
171 	return NULL;
172 }
173 
174 /**
175  * eeh_pe_next - Retrieve the next PE in the tree
176  * @pe: current PE
177  * @root: root PE
178  *
179  * The function is used to retrieve the next PE in the
180  * hierarchy PE tree.
181  */
182 struct eeh_pe *eeh_pe_next(struct eeh_pe *pe, struct eeh_pe *root)
183 {
184 	struct list_head *next = pe->child_list.next;
185 
186 	if (next == &pe->child_list) {
187 		while (1) {
188 			if (pe == root)
189 				return NULL;
190 			next = pe->child.next;
191 			if (next != &pe->parent->child_list)
192 				break;
193 			pe = pe->parent;
194 		}
195 	}
196 
197 	return list_entry(next, struct eeh_pe, child);
198 }
199 
200 /**
201  * eeh_pe_traverse - Traverse PEs in the specified PHB
202  * @root: root PE
203  * @fn: callback
204  * @flag: extra parameter to callback
205  *
206  * The function is used to traverse the specified PE and its
207  * child PEs. The traversing is to be terminated once the
208  * callback returns something other than NULL, or no more PEs
209  * to be traversed.
210  */
211 void *eeh_pe_traverse(struct eeh_pe *root,
212 		      eeh_pe_traverse_func fn, void *flag)
213 {
214 	struct eeh_pe *pe;
215 	void *ret;
216 
217 	eeh_for_each_pe(root, pe) {
218 		ret = fn(pe, flag);
219 		if (ret) return ret;
220 	}
221 
222 	return NULL;
223 }
224 
225 /**
226  * eeh_pe_dev_traverse - Traverse the devices from the PE
227  * @root: EEH PE
228  * @fn: function callback
229  * @flag: extra parameter to callback
230  *
231  * The function is used to traverse the devices of the specified
232  * PE and its child PEs.
233  */
234 void eeh_pe_dev_traverse(struct eeh_pe *root,
235 			  eeh_edev_traverse_func fn, void *flag)
236 {
237 	struct eeh_pe *pe;
238 	struct eeh_dev *edev, *tmp;
239 
240 	if (!root) {
241 		pr_warn("%s: Invalid PE %p\n",
242 			__func__, root);
243 		return;
244 	}
245 
246 	/* Traverse root PE */
247 	eeh_for_each_pe(root, pe)
248 		eeh_pe_for_each_dev(pe, edev, tmp)
249 			fn(edev, flag);
250 }
251 
252 /**
253  * __eeh_pe_get - Check the PE address
254  * @data: EEH PE
255  * @flag: EEH device
256  *
257  * For one particular PE, it can be identified by PE address
258  * or tranditional BDF address. BDF address is composed of
259  * Bus/Device/Function number. The extra data referred by flag
260  * indicates which type of address should be used.
261  */
262 struct eeh_pe_get_flag {
263 	int pe_no;
264 	int config_addr;
265 };
266 
267 static void *__eeh_pe_get(struct eeh_pe *pe, void *flag)
268 {
269 	struct eeh_pe_get_flag *tmp = (struct eeh_pe_get_flag *) flag;
270 
271 	/* Unexpected PHB PE */
272 	if (pe->type & EEH_PE_PHB)
273 		return NULL;
274 
275 	/*
276 	 * We prefer PE address. For most cases, we should
277 	 * have non-zero PE address
278 	 */
279 	if (eeh_has_flag(EEH_VALID_PE_ZERO)) {
280 		if (tmp->pe_no == pe->addr)
281 			return pe;
282 	} else {
283 		if (tmp->pe_no &&
284 		    (tmp->pe_no == pe->addr))
285 			return pe;
286 	}
287 
288 	/* Try BDF address */
289 	if (tmp->config_addr &&
290 	   (tmp->config_addr == pe->config_addr))
291 		return pe;
292 
293 	return NULL;
294 }
295 
296 /**
297  * eeh_pe_get - Search PE based on the given address
298  * @phb: PCI controller
299  * @pe_no: PE number
300  * @config_addr: Config address
301  *
302  * Search the corresponding PE based on the specified address which
303  * is included in the eeh device. The function is used to check if
304  * the associated PE has been created against the PE address. It's
305  * notable that the PE address has 2 format: traditional PE address
306  * which is composed of PCI bus/device/function number, or unified
307  * PE address.
308  */
309 struct eeh_pe *eeh_pe_get(struct pci_controller *phb,
310 		int pe_no, int config_addr)
311 {
312 	struct eeh_pe *root = eeh_phb_pe_get(phb);
313 	struct eeh_pe_get_flag tmp = { pe_no, config_addr };
314 	struct eeh_pe *pe;
315 
316 	pe = eeh_pe_traverse(root, __eeh_pe_get, &tmp);
317 
318 	return pe;
319 }
320 
321 /**
322  * eeh_pe_get_parent - Retrieve the parent PE
323  * @edev: EEH device
324  *
325  * The whole PEs existing in the system are organized as hierarchy
326  * tree. The function is used to retrieve the parent PE according
327  * to the parent EEH device.
328  */
329 static struct eeh_pe *eeh_pe_get_parent(struct eeh_dev *edev)
330 {
331 	struct eeh_dev *parent;
332 	struct pci_dn *pdn = eeh_dev_to_pdn(edev);
333 
334 	/*
335 	 * It might have the case for the indirect parent
336 	 * EEH device already having associated PE, but
337 	 * the direct parent EEH device doesn't have yet.
338 	 */
339 	if (edev->physfn)
340 		pdn = pci_get_pdn(edev->physfn);
341 	else
342 		pdn = pdn ? pdn->parent : NULL;
343 	while (pdn) {
344 		/* We're poking out of PCI territory */
345 		parent = pdn_to_eeh_dev(pdn);
346 		if (!parent)
347 			return NULL;
348 
349 		if (parent->pe)
350 			return parent->pe;
351 
352 		pdn = pdn->parent;
353 	}
354 
355 	return NULL;
356 }
357 
358 /**
359  * eeh_add_to_parent_pe - Add EEH device to parent PE
360  * @edev: EEH device
361  *
362  * Add EEH device to the parent PE. If the parent PE already
363  * exists, the PE type will be changed to EEH_PE_BUS. Otherwise,
364  * we have to create new PE to hold the EEH device and the new
365  * PE will be linked to its parent PE as well.
366  */
367 int eeh_add_to_parent_pe(struct eeh_dev *edev)
368 {
369 	struct eeh_pe *pe, *parent;
370 	struct pci_dn *pdn = eeh_dev_to_pdn(edev);
371 	int config_addr = (pdn->busno << 8) | (pdn->devfn);
372 
373 	/* Check if the PE number is valid */
374 	if (!eeh_has_flag(EEH_VALID_PE_ZERO) && !edev->pe_config_addr) {
375 		eeh_edev_err(edev, "PE#0 is invalid for this PHB!\n");
376 		return -EINVAL;
377 	}
378 
379 	/*
380 	 * Search the PE has been existing or not according
381 	 * to the PE address. If that has been existing, the
382 	 * PE should be composed of PCI bus and its subordinate
383 	 * components.
384 	 */
385 	pe = eeh_pe_get(pdn->phb, edev->pe_config_addr, config_addr);
386 	if (pe) {
387 		if (pe->type & EEH_PE_INVALID) {
388 			list_add_tail(&edev->entry, &pe->edevs);
389 			edev->pe = pe;
390 			/*
391 			 * We're running to here because of PCI hotplug caused by
392 			 * EEH recovery. We need clear EEH_PE_INVALID until the top.
393 			 */
394 			parent = pe;
395 			while (parent) {
396 				if (!(parent->type & EEH_PE_INVALID))
397 					break;
398 				parent->type &= ~EEH_PE_INVALID;
399 				parent = parent->parent;
400 			}
401 
402 			eeh_edev_dbg(edev,
403 				     "Added to device PE (parent: PE#%x)\n",
404 				     pe->parent->addr);
405 		} else {
406 			/* Mark the PE as type of PCI bus */
407 			pe->type = EEH_PE_BUS;
408 			edev->pe = pe;
409 
410 			/* Put the edev to PE */
411 			list_add_tail(&edev->entry, &pe->edevs);
412 			eeh_edev_dbg(edev, "Added to bus PE\n");
413 		}
414 		return 0;
415 	}
416 
417 	/* Create a new EEH PE */
418 	if (edev->physfn)
419 		pe = eeh_pe_alloc(pdn->phb, EEH_PE_VF);
420 	else
421 		pe = eeh_pe_alloc(pdn->phb, EEH_PE_DEVICE);
422 	if (!pe) {
423 		pr_err("%s: out of memory!\n", __func__);
424 		return -ENOMEM;
425 	}
426 	pe->addr	= edev->pe_config_addr;
427 	pe->config_addr	= config_addr;
428 
429 	/*
430 	 * Put the new EEH PE into hierarchy tree. If the parent
431 	 * can't be found, the newly created PE will be attached
432 	 * to PHB directly. Otherwise, we have to associate the
433 	 * PE with its parent.
434 	 */
435 	parent = eeh_pe_get_parent(edev);
436 	if (!parent) {
437 		parent = eeh_phb_pe_get(pdn->phb);
438 		if (!parent) {
439 			pr_err("%s: No PHB PE is found (PHB Domain=%d)\n",
440 				__func__, pdn->phb->global_number);
441 			edev->pe = NULL;
442 			kfree(pe);
443 			return -EEXIST;
444 		}
445 	}
446 	pe->parent = parent;
447 
448 	/*
449 	 * Put the newly created PE into the child list and
450 	 * link the EEH device accordingly.
451 	 */
452 	list_add_tail(&pe->child, &parent->child_list);
453 	list_add_tail(&edev->entry, &pe->edevs);
454 	edev->pe = pe;
455 	eeh_edev_dbg(edev, "Added to device PE (parent: PE#%x)\n",
456 		     pe->parent->addr);
457 
458 	return 0;
459 }
460 
461 /**
462  * eeh_rmv_from_parent_pe - Remove one EEH device from the associated PE
463  * @edev: EEH device
464  *
465  * The PE hierarchy tree might be changed when doing PCI hotplug.
466  * Also, the PCI devices or buses could be removed from the system
467  * during EEH recovery. So we have to call the function remove the
468  * corresponding PE accordingly if necessary.
469  */
470 int eeh_rmv_from_parent_pe(struct eeh_dev *edev)
471 {
472 	struct eeh_pe *pe, *parent, *child;
473 	int cnt;
474 
475 	pe = eeh_dev_to_pe(edev);
476 	if (!pe) {
477 		eeh_edev_dbg(edev, "No PE found for device.\n");
478 		return -EEXIST;
479 	}
480 
481 	/* Remove the EEH device */
482 	edev->pe = NULL;
483 	list_del(&edev->entry);
484 
485 	/*
486 	 * Check if the parent PE includes any EEH devices.
487 	 * If not, we should delete that. Also, we should
488 	 * delete the parent PE if it doesn't have associated
489 	 * child PEs and EEH devices.
490 	 */
491 	while (1) {
492 		parent = pe->parent;
493 		if (pe->type & EEH_PE_PHB)
494 			break;
495 
496 		if (!(pe->state & EEH_PE_KEEP)) {
497 			if (list_empty(&pe->edevs) &&
498 			    list_empty(&pe->child_list)) {
499 				list_del(&pe->child);
500 				kfree(pe);
501 			} else {
502 				break;
503 			}
504 		} else {
505 			if (list_empty(&pe->edevs)) {
506 				cnt = 0;
507 				list_for_each_entry(child, &pe->child_list, child) {
508 					if (!(child->type & EEH_PE_INVALID)) {
509 						cnt++;
510 						break;
511 					}
512 				}
513 
514 				if (!cnt)
515 					pe->type |= EEH_PE_INVALID;
516 				else
517 					break;
518 			}
519 		}
520 
521 		pe = parent;
522 	}
523 
524 	return 0;
525 }
526 
527 /**
528  * eeh_pe_update_time_stamp - Update PE's frozen time stamp
529  * @pe: EEH PE
530  *
531  * We have time stamp for each PE to trace its time of getting
532  * frozen in last hour. The function should be called to update
533  * the time stamp on first error of the specific PE. On the other
534  * handle, we needn't account for errors happened in last hour.
535  */
536 void eeh_pe_update_time_stamp(struct eeh_pe *pe)
537 {
538 	time64_t tstamp;
539 
540 	if (!pe) return;
541 
542 	if (pe->freeze_count <= 0) {
543 		pe->freeze_count = 0;
544 		pe->tstamp = ktime_get_seconds();
545 	} else {
546 		tstamp = ktime_get_seconds();
547 		if (tstamp - pe->tstamp > 3600) {
548 			pe->tstamp = tstamp;
549 			pe->freeze_count = 0;
550 		}
551 	}
552 }
553 
554 /**
555  * eeh_pe_state_mark - Mark specified state for PE and its associated device
556  * @pe: EEH PE
557  *
558  * EEH error affects the current PE and its child PEs. The function
559  * is used to mark appropriate state for the affected PEs and the
560  * associated devices.
561  */
562 void eeh_pe_state_mark(struct eeh_pe *root, int state)
563 {
564 	struct eeh_pe *pe;
565 
566 	eeh_for_each_pe(root, pe)
567 		if (!(pe->state & EEH_PE_REMOVED))
568 			pe->state |= state;
569 }
570 EXPORT_SYMBOL_GPL(eeh_pe_state_mark);
571 
572 /**
573  * eeh_pe_mark_isolated
574  * @pe: EEH PE
575  *
576  * Record that a PE has been isolated by marking the PE and it's children as
577  * EEH_PE_ISOLATED (and EEH_PE_CFG_BLOCKED, if required) and their PCI devices
578  * as pci_channel_io_frozen.
579  */
580 void eeh_pe_mark_isolated(struct eeh_pe *root)
581 {
582 	struct eeh_pe *pe;
583 	struct eeh_dev *edev;
584 	struct pci_dev *pdev;
585 
586 	eeh_pe_state_mark(root, EEH_PE_ISOLATED);
587 	eeh_for_each_pe(root, pe) {
588 		list_for_each_entry(edev, &pe->edevs, entry) {
589 			pdev = eeh_dev_to_pci_dev(edev);
590 			if (pdev)
591 				pdev->error_state = pci_channel_io_frozen;
592 		}
593 		/* Block PCI config access if required */
594 		if (pe->state & EEH_PE_CFG_RESTRICTED)
595 			pe->state |= EEH_PE_CFG_BLOCKED;
596 	}
597 }
598 EXPORT_SYMBOL_GPL(eeh_pe_mark_isolated);
599 
600 static void __eeh_pe_dev_mode_mark(struct eeh_dev *edev, void *flag)
601 {
602 	int mode = *((int *)flag);
603 
604 	edev->mode |= mode;
605 }
606 
607 /**
608  * eeh_pe_dev_state_mark - Mark state for all device under the PE
609  * @pe: EEH PE
610  *
611  * Mark specific state for all child devices of the PE.
612  */
613 void eeh_pe_dev_mode_mark(struct eeh_pe *pe, int mode)
614 {
615 	eeh_pe_dev_traverse(pe, __eeh_pe_dev_mode_mark, &mode);
616 }
617 
618 /**
619  * eeh_pe_state_clear - Clear state for the PE
620  * @data: EEH PE
621  * @state: state
622  * @include_passed: include passed-through devices?
623  *
624  * The function is used to clear the indicated state from the
625  * given PE. Besides, we also clear the check count of the PE
626  * as well.
627  */
628 void eeh_pe_state_clear(struct eeh_pe *root, int state, bool include_passed)
629 {
630 	struct eeh_pe *pe;
631 	struct eeh_dev *edev, *tmp;
632 	struct pci_dev *pdev;
633 
634 	eeh_for_each_pe(root, pe) {
635 		/* Keep the state of permanently removed PE intact */
636 		if (pe->state & EEH_PE_REMOVED)
637 			continue;
638 
639 		if (!include_passed && eeh_pe_passed(pe))
640 			continue;
641 
642 		pe->state &= ~state;
643 
644 		/*
645 		 * Special treatment on clearing isolated state. Clear
646 		 * check count since last isolation and put all affected
647 		 * devices to normal state.
648 		 */
649 		if (!(state & EEH_PE_ISOLATED))
650 			continue;
651 
652 		pe->check_count = 0;
653 		eeh_pe_for_each_dev(pe, edev, tmp) {
654 			pdev = eeh_dev_to_pci_dev(edev);
655 			if (!pdev)
656 				continue;
657 
658 			pdev->error_state = pci_channel_io_normal;
659 		}
660 
661 		/* Unblock PCI config access if required */
662 		if (pe->state & EEH_PE_CFG_RESTRICTED)
663 			pe->state &= ~EEH_PE_CFG_BLOCKED;
664 	}
665 }
666 
667 /*
668  * Some PCI bridges (e.g. PLX bridges) have primary/secondary
669  * buses assigned explicitly by firmware, and we probably have
670  * lost that after reset. So we have to delay the check until
671  * the PCI-CFG registers have been restored for the parent
672  * bridge.
673  *
674  * Don't use normal PCI-CFG accessors, which probably has been
675  * blocked on normal path during the stage. So we need utilize
676  * eeh operations, which is always permitted.
677  */
678 static void eeh_bridge_check_link(struct eeh_dev *edev)
679 {
680 	struct pci_dn *pdn = eeh_dev_to_pdn(edev);
681 	int cap;
682 	uint32_t val;
683 	int timeout = 0;
684 
685 	/*
686 	 * We only check root port and downstream ports of
687 	 * PCIe switches
688 	 */
689 	if (!(edev->mode & (EEH_DEV_ROOT_PORT | EEH_DEV_DS_PORT)))
690 		return;
691 
692 	eeh_edev_dbg(edev, "Checking PCIe link...\n");
693 
694 	/* Check slot status */
695 	cap = edev->pcie_cap;
696 	eeh_ops->read_config(pdn, cap + PCI_EXP_SLTSTA, 2, &val);
697 	if (!(val & PCI_EXP_SLTSTA_PDS)) {
698 		eeh_edev_dbg(edev, "No card in the slot (0x%04x) !\n", val);
699 		return;
700 	}
701 
702 	/* Check power status if we have the capability */
703 	eeh_ops->read_config(pdn, cap + PCI_EXP_SLTCAP, 2, &val);
704 	if (val & PCI_EXP_SLTCAP_PCP) {
705 		eeh_ops->read_config(pdn, cap + PCI_EXP_SLTCTL, 2, &val);
706 		if (val & PCI_EXP_SLTCTL_PCC) {
707 			eeh_edev_dbg(edev, "In power-off state, power it on ...\n");
708 			val &= ~(PCI_EXP_SLTCTL_PCC | PCI_EXP_SLTCTL_PIC);
709 			val |= (0x0100 & PCI_EXP_SLTCTL_PIC);
710 			eeh_ops->write_config(pdn, cap + PCI_EXP_SLTCTL, 2, val);
711 			msleep(2 * 1000);
712 		}
713 	}
714 
715 	/* Enable link */
716 	eeh_ops->read_config(pdn, cap + PCI_EXP_LNKCTL, 2, &val);
717 	val &= ~PCI_EXP_LNKCTL_LD;
718 	eeh_ops->write_config(pdn, cap + PCI_EXP_LNKCTL, 2, val);
719 
720 	/* Check link */
721 	eeh_ops->read_config(pdn, cap + PCI_EXP_LNKCAP, 4, &val);
722 	if (!(val & PCI_EXP_LNKCAP_DLLLARC)) {
723 		eeh_edev_dbg(edev, "No link reporting capability (0x%08x) \n", val);
724 		msleep(1000);
725 		return;
726 	}
727 
728 	/* Wait the link is up until timeout (5s) */
729 	timeout = 0;
730 	while (timeout < 5000) {
731 		msleep(20);
732 		timeout += 20;
733 
734 		eeh_ops->read_config(pdn, cap + PCI_EXP_LNKSTA, 2, &val);
735 		if (val & PCI_EXP_LNKSTA_DLLLA)
736 			break;
737 	}
738 
739 	if (val & PCI_EXP_LNKSTA_DLLLA)
740 		eeh_edev_dbg(edev, "Link up (%s)\n",
741 			 (val & PCI_EXP_LNKSTA_CLS_2_5GB) ? "2.5GB" : "5GB");
742 	else
743 		eeh_edev_dbg(edev, "Link not ready (0x%04x)\n", val);
744 }
745 
746 #define BYTE_SWAP(OFF)	(8*((OFF)/4)+3-(OFF))
747 #define SAVED_BYTE(OFF)	(((u8 *)(edev->config_space))[BYTE_SWAP(OFF)])
748 
749 static void eeh_restore_bridge_bars(struct eeh_dev *edev)
750 {
751 	struct pci_dn *pdn = eeh_dev_to_pdn(edev);
752 	int i;
753 
754 	/*
755 	 * Device BARs: 0x10 - 0x18
756 	 * Bus numbers and windows: 0x18 - 0x30
757 	 */
758 	for (i = 4; i < 13; i++)
759 		eeh_ops->write_config(pdn, i*4, 4, edev->config_space[i]);
760 	/* Rom: 0x38 */
761 	eeh_ops->write_config(pdn, 14*4, 4, edev->config_space[14]);
762 
763 	/* Cache line & Latency timer: 0xC 0xD */
764 	eeh_ops->write_config(pdn, PCI_CACHE_LINE_SIZE, 1,
765                 SAVED_BYTE(PCI_CACHE_LINE_SIZE));
766         eeh_ops->write_config(pdn, PCI_LATENCY_TIMER, 1,
767                 SAVED_BYTE(PCI_LATENCY_TIMER));
768 	/* Max latency, min grant, interrupt ping and line: 0x3C */
769 	eeh_ops->write_config(pdn, 15*4, 4, edev->config_space[15]);
770 
771 	/* PCI Command: 0x4 */
772 	eeh_ops->write_config(pdn, PCI_COMMAND, 4, edev->config_space[1] |
773 			      PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER);
774 
775 	/* Check the PCIe link is ready */
776 	eeh_bridge_check_link(edev);
777 }
778 
779 static void eeh_restore_device_bars(struct eeh_dev *edev)
780 {
781 	struct pci_dn *pdn = eeh_dev_to_pdn(edev);
782 	int i;
783 	u32 cmd;
784 
785 	for (i = 4; i < 10; i++)
786 		eeh_ops->write_config(pdn, i*4, 4, edev->config_space[i]);
787 	/* 12 == Expansion ROM Address */
788 	eeh_ops->write_config(pdn, 12*4, 4, edev->config_space[12]);
789 
790 	eeh_ops->write_config(pdn, PCI_CACHE_LINE_SIZE, 1,
791 		SAVED_BYTE(PCI_CACHE_LINE_SIZE));
792 	eeh_ops->write_config(pdn, PCI_LATENCY_TIMER, 1,
793 		SAVED_BYTE(PCI_LATENCY_TIMER));
794 
795 	/* max latency, min grant, interrupt pin and line */
796 	eeh_ops->write_config(pdn, 15*4, 4, edev->config_space[15]);
797 
798 	/*
799 	 * Restore PERR & SERR bits, some devices require it,
800 	 * don't touch the other command bits
801 	 */
802 	eeh_ops->read_config(pdn, PCI_COMMAND, 4, &cmd);
803 	if (edev->config_space[1] & PCI_COMMAND_PARITY)
804 		cmd |= PCI_COMMAND_PARITY;
805 	else
806 		cmd &= ~PCI_COMMAND_PARITY;
807 	if (edev->config_space[1] & PCI_COMMAND_SERR)
808 		cmd |= PCI_COMMAND_SERR;
809 	else
810 		cmd &= ~PCI_COMMAND_SERR;
811 	eeh_ops->write_config(pdn, PCI_COMMAND, 4, cmd);
812 }
813 
814 /**
815  * eeh_restore_one_device_bars - Restore the Base Address Registers for one device
816  * @data: EEH device
817  * @flag: Unused
818  *
819  * Loads the PCI configuration space base address registers,
820  * the expansion ROM base address, the latency timer, and etc.
821  * from the saved values in the device node.
822  */
823 static void eeh_restore_one_device_bars(struct eeh_dev *edev, void *flag)
824 {
825 	struct pci_dn *pdn = eeh_dev_to_pdn(edev);
826 
827 	/* Do special restore for bridges */
828 	if (edev->mode & EEH_DEV_BRIDGE)
829 		eeh_restore_bridge_bars(edev);
830 	else
831 		eeh_restore_device_bars(edev);
832 
833 	if (eeh_ops->restore_config && pdn)
834 		eeh_ops->restore_config(pdn);
835 }
836 
837 /**
838  * eeh_pe_restore_bars - Restore the PCI config space info
839  * @pe: EEH PE
840  *
841  * This routine performs a recursive walk to the children
842  * of this device as well.
843  */
844 void eeh_pe_restore_bars(struct eeh_pe *pe)
845 {
846 	/*
847 	 * We needn't take the EEH lock since eeh_pe_dev_traverse()
848 	 * will take that.
849 	 */
850 	eeh_pe_dev_traverse(pe, eeh_restore_one_device_bars, NULL);
851 }
852 
853 /**
854  * eeh_pe_loc_get - Retrieve location code binding to the given PE
855  * @pe: EEH PE
856  *
857  * Retrieve the location code of the given PE. If the primary PE bus
858  * is root bus, we will grab location code from PHB device tree node
859  * or root port. Otherwise, the upstream bridge's device tree node
860  * of the primary PE bus will be checked for the location code.
861  */
862 const char *eeh_pe_loc_get(struct eeh_pe *pe)
863 {
864 	struct pci_bus *bus = eeh_pe_bus_get(pe);
865 	struct device_node *dn;
866 	const char *loc = NULL;
867 
868 	while (bus) {
869 		dn = pci_bus_to_OF_node(bus);
870 		if (!dn) {
871 			bus = bus->parent;
872 			continue;
873 		}
874 
875 		if (pci_is_root_bus(bus))
876 			loc = of_get_property(dn, "ibm,io-base-loc-code", NULL);
877 		else
878 			loc = of_get_property(dn, "ibm,slot-location-code",
879 					      NULL);
880 
881 		if (loc)
882 			return loc;
883 
884 		bus = bus->parent;
885 	}
886 
887 	return "N/A";
888 }
889 
890 /**
891  * eeh_pe_bus_get - Retrieve PCI bus according to the given PE
892  * @pe: EEH PE
893  *
894  * Retrieve the PCI bus according to the given PE. Basically,
895  * there're 3 types of PEs: PHB/Bus/Device. For PHB PE, the
896  * primary PCI bus will be retrieved. The parent bus will be
897  * returned for BUS PE. However, we don't have associated PCI
898  * bus for DEVICE PE.
899  */
900 struct pci_bus *eeh_pe_bus_get(struct eeh_pe *pe)
901 {
902 	struct eeh_dev *edev;
903 	struct pci_dev *pdev;
904 
905 	if (pe->type & EEH_PE_PHB)
906 		return pe->phb->bus;
907 
908 	/* The primary bus might be cached during probe time */
909 	if (pe->state & EEH_PE_PRI_BUS)
910 		return pe->bus;
911 
912 	/* Retrieve the parent PCI bus of first (top) PCI device */
913 	edev = list_first_entry_or_null(&pe->edevs, struct eeh_dev, entry);
914 	pdev = eeh_dev_to_pci_dev(edev);
915 	if (pdev)
916 		return pdev->bus;
917 
918 	return NULL;
919 }
920