xref: /linux/arch/x86/pci/acpi.c (revision d39d0ed196aa1685bb24771e92f78633c66ac9cb)
1 #include <linux/pci.h>
2 #include <linux/acpi.h>
3 #include <linux/init.h>
4 #include <linux/irq.h>
5 #include <linux/dmi.h>
6 #include <linux/slab.h>
7 #include <asm/numa.h>
8 #include <asm/pci_x86.h>
9 
10 struct pci_root_info {
11 	struct acpi_device *bridge;
12 	char *name;
13 	unsigned int res_num;
14 	struct resource *res;
15 	struct pci_bus *bus;
16 	int busnum;
17 };
18 
19 static bool pci_use_crs = true;
20 
21 static int __init set_use_crs(const struct dmi_system_id *id)
22 {
23 	pci_use_crs = true;
24 	return 0;
25 }
26 
27 static const struct dmi_system_id pci_use_crs_table[] __initconst = {
28 	/* http://bugzilla.kernel.org/show_bug.cgi?id=14183 */
29 	{
30 		.callback = set_use_crs,
31 		.ident = "IBM System x3800",
32 		.matches = {
33 			DMI_MATCH(DMI_SYS_VENDOR, "IBM"),
34 			DMI_MATCH(DMI_PRODUCT_NAME, "x3800"),
35 		},
36 	},
37 	/* https://bugzilla.kernel.org/show_bug.cgi?id=16007 */
38 	/* 2006 AMD HT/VIA system with two host bridges */
39         {
40 		.callback = set_use_crs,
41 		.ident = "ASRock ALiveSATA2-GLAN",
42 		.matches = {
43 			DMI_MATCH(DMI_PRODUCT_NAME, "ALiveSATA2-GLAN"),
44                 },
45         },
46 	{}
47 };
48 
49 void __init pci_acpi_crs_quirks(void)
50 {
51 	int year;
52 
53 	if (dmi_get_date(DMI_BIOS_DATE, &year, NULL, NULL) && year < 2008)
54 		pci_use_crs = false;
55 
56 	dmi_check_system(pci_use_crs_table);
57 
58 	/*
59 	 * If the user specifies "pci=use_crs" or "pci=nocrs" explicitly, that
60 	 * takes precedence over anything we figured out above.
61 	 */
62 	if (pci_probe & PCI_ROOT_NO_CRS)
63 		pci_use_crs = false;
64 	else if (pci_probe & PCI_USE__CRS)
65 		pci_use_crs = true;
66 
67 	printk(KERN_INFO "PCI: %s host bridge windows from ACPI; "
68 	       "if necessary, use \"pci=%s\" and report a bug\n",
69 	       pci_use_crs ? "Using" : "Ignoring",
70 	       pci_use_crs ? "nocrs" : "use_crs");
71 }
72 
73 static acpi_status
74 resource_to_addr(struct acpi_resource *resource,
75 			struct acpi_resource_address64 *addr)
76 {
77 	acpi_status status;
78 	struct acpi_resource_memory24 *memory24;
79 	struct acpi_resource_memory32 *memory32;
80 	struct acpi_resource_fixed_memory32 *fixed_memory32;
81 
82 	memset(addr, 0, sizeof(*addr));
83 	switch (resource->type) {
84 	case ACPI_RESOURCE_TYPE_MEMORY24:
85 		memory24 = &resource->data.memory24;
86 		addr->resource_type = ACPI_MEMORY_RANGE;
87 		addr->minimum = memory24->minimum;
88 		addr->address_length = memory24->address_length;
89 		addr->maximum = addr->minimum + addr->address_length - 1;
90 		return AE_OK;
91 	case ACPI_RESOURCE_TYPE_MEMORY32:
92 		memory32 = &resource->data.memory32;
93 		addr->resource_type = ACPI_MEMORY_RANGE;
94 		addr->minimum = memory32->minimum;
95 		addr->address_length = memory32->address_length;
96 		addr->maximum = addr->minimum + addr->address_length - 1;
97 		return AE_OK;
98 	case ACPI_RESOURCE_TYPE_FIXED_MEMORY32:
99 		fixed_memory32 = &resource->data.fixed_memory32;
100 		addr->resource_type = ACPI_MEMORY_RANGE;
101 		addr->minimum = fixed_memory32->address;
102 		addr->address_length = fixed_memory32->address_length;
103 		addr->maximum = addr->minimum + addr->address_length - 1;
104 		return AE_OK;
105 	case ACPI_RESOURCE_TYPE_ADDRESS16:
106 	case ACPI_RESOURCE_TYPE_ADDRESS32:
107 	case ACPI_RESOURCE_TYPE_ADDRESS64:
108 		status = acpi_resource_to_address64(resource, addr);
109 		if (ACPI_SUCCESS(status) &&
110 		    (addr->resource_type == ACPI_MEMORY_RANGE ||
111 		    addr->resource_type == ACPI_IO_RANGE) &&
112 		    addr->address_length > 0) {
113 			return AE_OK;
114 		}
115 		break;
116 	}
117 	return AE_ERROR;
118 }
119 
120 static acpi_status
121 count_resource(struct acpi_resource *acpi_res, void *data)
122 {
123 	struct pci_root_info *info = data;
124 	struct acpi_resource_address64 addr;
125 	acpi_status status;
126 
127 	status = resource_to_addr(acpi_res, &addr);
128 	if (ACPI_SUCCESS(status))
129 		info->res_num++;
130 	return AE_OK;
131 }
132 
133 static acpi_status
134 setup_resource(struct acpi_resource *acpi_res, void *data)
135 {
136 	struct pci_root_info *info = data;
137 	struct resource *res;
138 	struct acpi_resource_address64 addr;
139 	acpi_status status;
140 	unsigned long flags;
141 	struct resource *root, *conflict;
142 	u64 start, end;
143 
144 	status = resource_to_addr(acpi_res, &addr);
145 	if (!ACPI_SUCCESS(status))
146 		return AE_OK;
147 
148 	if (addr.resource_type == ACPI_MEMORY_RANGE) {
149 		root = &iomem_resource;
150 		flags = IORESOURCE_MEM;
151 		if (addr.info.mem.caching == ACPI_PREFETCHABLE_MEMORY)
152 			flags |= IORESOURCE_PREFETCH;
153 	} else if (addr.resource_type == ACPI_IO_RANGE) {
154 		root = &ioport_resource;
155 		flags = IORESOURCE_IO;
156 	} else
157 		return AE_OK;
158 
159 	start = addr.minimum + addr.translation_offset;
160 	end = addr.maximum + addr.translation_offset;
161 
162 	res = &info->res[info->res_num];
163 	res->name = info->name;
164 	res->flags = flags;
165 	res->start = start;
166 	res->end = end;
167 	res->child = NULL;
168 
169 	if (!pci_use_crs) {
170 		dev_printk(KERN_DEBUG, &info->bridge->dev,
171 			   "host bridge window %pR (ignored)\n", res);
172 		return AE_OK;
173 	}
174 
175 	conflict = insert_resource_conflict(root, res);
176 	if (conflict) {
177 		dev_err(&info->bridge->dev,
178 			"address space collision: host bridge window %pR "
179 			"conflicts with %s %pR\n",
180 			res, conflict->name, conflict);
181 	} else {
182 		pci_bus_add_resource(info->bus, res, 0);
183 		info->res_num++;
184 		if (addr.translation_offset)
185 			dev_info(&info->bridge->dev, "host bridge window %pR "
186 				 "(PCI address [%#llx-%#llx])\n",
187 				 res, res->start - addr.translation_offset,
188 				 res->end - addr.translation_offset);
189 		else
190 			dev_info(&info->bridge->dev,
191 				 "host bridge window %pR\n", res);
192 	}
193 	return AE_OK;
194 }
195 
196 static void
197 get_current_resources(struct acpi_device *device, int busnum,
198 			int domain, struct pci_bus *bus)
199 {
200 	struct pci_root_info info;
201 	size_t size;
202 
203 	if (pci_use_crs)
204 		pci_bus_remove_resources(bus);
205 
206 	info.bridge = device;
207 	info.bus = bus;
208 	info.res_num = 0;
209 	acpi_walk_resources(device->handle, METHOD_NAME__CRS, count_resource,
210 				&info);
211 	if (!info.res_num)
212 		return;
213 
214 	size = sizeof(*info.res) * info.res_num;
215 	info.res = kmalloc(size, GFP_KERNEL);
216 	if (!info.res)
217 		goto res_alloc_fail;
218 
219 	info.name = kasprintf(GFP_KERNEL, "PCI Bus %04x:%02x", domain, busnum);
220 	if (!info.name)
221 		goto name_alloc_fail;
222 
223 	info.res_num = 0;
224 	acpi_walk_resources(device->handle, METHOD_NAME__CRS, setup_resource,
225 				&info);
226 
227 	return;
228 
229 name_alloc_fail:
230 	kfree(info.res);
231 res_alloc_fail:
232 	return;
233 }
234 
235 struct pci_bus * __devinit pci_acpi_scan_root(struct acpi_pci_root *root)
236 {
237 	struct acpi_device *device = root->device;
238 	int domain = root->segment;
239 	int busnum = root->secondary.start;
240 	struct pci_bus *bus;
241 	struct pci_sysdata *sd;
242 	int node;
243 #ifdef CONFIG_ACPI_NUMA
244 	int pxm;
245 #endif
246 
247 	if (domain && !pci_domains_supported) {
248 		printk(KERN_WARNING "pci_bus %04x:%02x: "
249 		       "ignored (multiple domains not supported)\n",
250 		       domain, busnum);
251 		return NULL;
252 	}
253 
254 	node = -1;
255 #ifdef CONFIG_ACPI_NUMA
256 	pxm = acpi_get_pxm(device->handle);
257 	if (pxm >= 0)
258 		node = pxm_to_node(pxm);
259 	if (node != -1)
260 		set_mp_bus_to_node(busnum, node);
261 	else
262 #endif
263 		node = get_mp_bus_to_node(busnum);
264 
265 	if (node != -1 && !node_online(node))
266 		node = -1;
267 
268 	/* Allocate per-root-bus (not per bus) arch-specific data.
269 	 * TODO: leak; this memory is never freed.
270 	 * It's arguable whether it's worth the trouble to care.
271 	 */
272 	sd = kzalloc(sizeof(*sd), GFP_KERNEL);
273 	if (!sd) {
274 		printk(KERN_WARNING "pci_bus %04x:%02x: "
275 		       "ignored (out of memory)\n", domain, busnum);
276 		return NULL;
277 	}
278 
279 	sd->domain = domain;
280 	sd->node = node;
281 	/*
282 	 * Maybe the desired pci bus has been already scanned. In such case
283 	 * it is unnecessary to scan the pci bus with the given domain,busnum.
284 	 */
285 	bus = pci_find_bus(domain, busnum);
286 	if (bus) {
287 		/*
288 		 * If the desired bus exits, the content of bus->sysdata will
289 		 * be replaced by sd.
290 		 */
291 		memcpy(bus->sysdata, sd, sizeof(*sd));
292 		kfree(sd);
293 	} else {
294 		bus = pci_create_bus(NULL, busnum, &pci_root_ops, sd);
295 		if (bus) {
296 			get_current_resources(device, busnum, domain, bus);
297 			bus->subordinate = pci_scan_child_bus(bus);
298 		}
299 	}
300 
301 	if (!bus)
302 		kfree(sd);
303 
304 	if (bus && node != -1) {
305 #ifdef CONFIG_ACPI_NUMA
306 		if (pxm >= 0)
307 			dev_printk(KERN_DEBUG, &bus->dev,
308 				   "on NUMA node %d (pxm %d)\n", node, pxm);
309 #else
310 		dev_printk(KERN_DEBUG, &bus->dev, "on NUMA node %d\n", node);
311 #endif
312 	}
313 
314 	return bus;
315 }
316 
317 int __init pci_acpi_init(void)
318 {
319 	struct pci_dev *dev = NULL;
320 
321 	if (acpi_noirq)
322 		return -ENODEV;
323 
324 	printk(KERN_INFO "PCI: Using ACPI for IRQ routing\n");
325 	acpi_irq_penalty_init();
326 	pcibios_enable_irq = acpi_pci_irq_enable;
327 	pcibios_disable_irq = acpi_pci_irq_disable;
328 	x86_init.pci.init_irq = x86_init_noop;
329 
330 	if (pci_routeirq) {
331 		/*
332 		 * PCI IRQ routing is set up by pci_enable_device(), but we
333 		 * also do it here in case there are still broken drivers that
334 		 * don't use pci_enable_device().
335 		 */
336 		printk(KERN_INFO "PCI: Routing PCI interrupts for all devices because \"pci=routeirq\" specified\n");
337 		for_each_pci_dev(dev)
338 			acpi_pci_irq_enable(dev);
339 	}
340 
341 	return 0;
342 }
343