1 /* 2 * arch/xtensa/kernel/pci.c 3 * 4 * PCI bios-type initialisation for PCI machines 5 * 6 * This program is free software; you can redistribute it and/or modify it 7 * under the terms of the GNU General Public License as published by the 8 * Free Software Foundation; either version 2 of the License, or (at your 9 * option) any later version. 10 * 11 * Copyright (C) 2001-2005 Tensilica Inc. 12 * 13 * Based largely on work from Cort (ppc/kernel/pci.c) 14 * IO functions copied from sparc. 15 * 16 * Chris Zankel <chris@zankel.net> 17 * 18 */ 19 20 #include <linux/kernel.h> 21 #include <linux/pci.h> 22 #include <linux/delay.h> 23 #include <linux/string.h> 24 #include <linux/init.h> 25 #include <linux/sched.h> 26 #include <linux/errno.h> 27 #include <linux/bootmem.h> 28 29 #include <asm/pci-bridge.h> 30 #include <asm/platform.h> 31 32 #undef DEBUG 33 34 #ifdef DEBUG 35 #define DBG(x...) printk(x) 36 #else 37 #define DBG(x...) 38 #endif 39 40 /* PCI Controller */ 41 42 43 /* 44 * pcibios_alloc_controller 45 * pcibios_enable_device 46 * pcibios_fixups 47 * pcibios_align_resource 48 * pcibios_fixup_bus 49 * pci_bus_add_device 50 * pci_mmap_page_range 51 */ 52 53 struct pci_controller* pci_ctrl_head; 54 struct pci_controller** pci_ctrl_tail = &pci_ctrl_head; 55 56 static int pci_bus_count; 57 58 /* 59 * We need to avoid collisions with `mirrored' VGA ports 60 * and other strange ISA hardware, so we always want the 61 * addresses to be allocated in the 0x000-0x0ff region 62 * modulo 0x400. 63 * 64 * Why? Because some silly external IO cards only decode 65 * the low 10 bits of the IO address. The 0x00-0xff region 66 * is reserved for motherboard devices that decode all 16 67 * bits, so it's ok to allocate at, say, 0x2800-0x28ff, 68 * but we want to try to avoid allocating at 0x2900-0x2bff 69 * which might have be mirrored at 0x0100-0x03ff.. 70 */ 71 resource_size_t 72 pcibios_align_resource(void *data, const struct resource *res, 73 resource_size_t size, resource_size_t align) 74 { 75 struct pci_dev *dev = data; 76 resource_size_t start = res->start; 77 78 if (res->flags & IORESOURCE_IO) { 79 if (size > 0x100) { 80 printk(KERN_ERR "PCI: I/O Region %s/%d too large" 81 " (%ld bytes)\n", pci_name(dev), 82 dev->resource - res, size); 83 } 84 85 if (start & 0x300) 86 start = (start + 0x3ff) & ~0x3ff; 87 } 88 89 return start; 90 } 91 92 int 93 pcibios_enable_resources(struct pci_dev *dev, int mask) 94 { 95 u16 cmd, old_cmd; 96 int idx; 97 struct resource *r; 98 99 pci_read_config_word(dev, PCI_COMMAND, &cmd); 100 old_cmd = cmd; 101 for(idx=0; idx<6; idx++) { 102 r = &dev->resource[idx]; 103 if (!r->start && r->end) { 104 printk (KERN_ERR "PCI: Device %s not available because " 105 "of resource collisions\n", pci_name(dev)); 106 return -EINVAL; 107 } 108 if (r->flags & IORESOURCE_IO) 109 cmd |= PCI_COMMAND_IO; 110 if (r->flags & IORESOURCE_MEM) 111 cmd |= PCI_COMMAND_MEMORY; 112 } 113 if (dev->resource[PCI_ROM_RESOURCE].start) 114 cmd |= PCI_COMMAND_MEMORY; 115 if (cmd != old_cmd) { 116 printk("PCI: Enabling device %s (%04x -> %04x)\n", 117 pci_name(dev), old_cmd, cmd); 118 pci_write_config_word(dev, PCI_COMMAND, cmd); 119 } 120 return 0; 121 } 122 123 struct pci_controller * __init pcibios_alloc_controller(void) 124 { 125 struct pci_controller *pci_ctrl; 126 127 pci_ctrl = (struct pci_controller *)alloc_bootmem(sizeof(*pci_ctrl)); 128 memset(pci_ctrl, 0, sizeof(struct pci_controller)); 129 130 *pci_ctrl_tail = pci_ctrl; 131 pci_ctrl_tail = &pci_ctrl->next; 132 133 return pci_ctrl; 134 } 135 136 static void __init pci_controller_apertures(struct pci_controller *pci_ctrl, 137 struct list_head *resources) 138 { 139 struct resource *res; 140 unsigned long io_offset; 141 int i; 142 143 io_offset = (unsigned long)pci_ctrl->io_space.base; 144 res = &pci_ctrl->io_resource; 145 if (!res->flags) { 146 if (io_offset) 147 printk (KERN_ERR "I/O resource not set for host" 148 " bridge %d\n", pci_ctrl->index); 149 res->start = 0; 150 res->end = IO_SPACE_LIMIT; 151 res->flags = IORESOURCE_IO; 152 } 153 res->start += io_offset; 154 res->end += io_offset; 155 pci_add_resource_offset(resources, res, io_offset); 156 157 for (i = 0; i < 3; i++) { 158 res = &pci_ctrl->mem_resources[i]; 159 if (!res->flags) { 160 if (i > 0) 161 continue; 162 printk(KERN_ERR "Memory resource not set for " 163 "host bridge %d\n", pci_ctrl->index); 164 res->start = 0; 165 res->end = ~0U; 166 res->flags = IORESOURCE_MEM; 167 } 168 pci_add_resource(resources, res); 169 } 170 } 171 172 static int __init pcibios_init(void) 173 { 174 struct pci_controller *pci_ctrl; 175 struct list_head resources; 176 struct pci_bus *bus; 177 int next_busno = 0, i; 178 179 printk("PCI: Probing PCI hardware\n"); 180 181 /* Scan all of the recorded PCI controllers. */ 182 for (pci_ctrl = pci_ctrl_head; pci_ctrl; pci_ctrl = pci_ctrl->next) { 183 pci_ctrl->last_busno = 0xff; 184 INIT_LIST_HEAD(&resources); 185 pci_controller_apertures(pci_ctrl, &resources); 186 bus = pci_scan_root_bus(NULL, pci_ctrl->first_busno, 187 pci_ctrl->ops, pci_ctrl, &resources); 188 pci_ctrl->bus = bus; 189 pci_ctrl->last_busno = bus->busn_res.end; 190 if (next_busno <= pci_ctrl->last_busno) 191 next_busno = pci_ctrl->last_busno+1; 192 } 193 pci_bus_count = next_busno; 194 195 return platform_pcibios_fixup(); 196 } 197 198 subsys_initcall(pcibios_init); 199 200 void __init pcibios_fixup_bus(struct pci_bus *bus) 201 { 202 if (bus->parent) { 203 /* This is a subordinate bridge */ 204 pci_read_bridge_bases(bus); 205 } 206 } 207 208 void pcibios_set_master(struct pci_dev *dev) 209 { 210 /* No special bus mastering setup handling */ 211 } 212 213 /* the next one is stolen from the alpha port... */ 214 215 void __init 216 pcibios_update_irq(struct pci_dev *dev, int irq) 217 { 218 pci_write_config_byte(dev, PCI_INTERRUPT_LINE, irq); 219 } 220 221 int pcibios_enable_device(struct pci_dev *dev, int mask) 222 { 223 u16 cmd, old_cmd; 224 int idx; 225 struct resource *r; 226 227 pci_read_config_word(dev, PCI_COMMAND, &cmd); 228 old_cmd = cmd; 229 for (idx=0; idx<6; idx++) { 230 r = &dev->resource[idx]; 231 if (!r->start && r->end) { 232 printk(KERN_ERR "PCI: Device %s not available because " 233 "of resource collisions\n", pci_name(dev)); 234 return -EINVAL; 235 } 236 if (r->flags & IORESOURCE_IO) 237 cmd |= PCI_COMMAND_IO; 238 if (r->flags & IORESOURCE_MEM) 239 cmd |= PCI_COMMAND_MEMORY; 240 } 241 if (cmd != old_cmd) { 242 printk("PCI: Enabling device %s (%04x -> %04x)\n", 243 pci_name(dev), old_cmd, cmd); 244 pci_write_config_word(dev, PCI_COMMAND, cmd); 245 } 246 247 return 0; 248 } 249 250 #ifdef CONFIG_PROC_FS 251 252 /* 253 * Return the index of the PCI controller for device pdev. 254 */ 255 256 int 257 pci_controller_num(struct pci_dev *dev) 258 { 259 struct pci_controller *pci_ctrl = (struct pci_controller*) dev->sysdata; 260 return pci_ctrl->index; 261 } 262 263 #endif /* CONFIG_PROC_FS */ 264 265 /* 266 * Platform support for /proc/bus/pci/X/Y mmap()s, 267 * modelled on the sparc64 implementation by Dave Miller. 268 * -- paulus. 269 */ 270 271 /* 272 * Adjust vm_pgoff of VMA such that it is the physical page offset 273 * corresponding to the 32-bit pci bus offset for DEV requested by the user. 274 * 275 * Basically, the user finds the base address for his device which he wishes 276 * to mmap. They read the 32-bit value from the config space base register, 277 * add whatever PAGE_SIZE multiple offset they wish, and feed this into the 278 * offset parameter of mmap on /proc/bus/pci/XXX for that device. 279 * 280 * Returns negative error code on failure, zero on success. 281 */ 282 static __inline__ int 283 __pci_mmap_make_offset(struct pci_dev *dev, struct vm_area_struct *vma, 284 enum pci_mmap_state mmap_state) 285 { 286 struct pci_controller *pci_ctrl = (struct pci_controller*) dev->sysdata; 287 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; 288 unsigned long io_offset = 0; 289 int i, res_bit; 290 291 if (pci_ctrl == 0) 292 return -EINVAL; /* should never happen */ 293 294 /* If memory, add on the PCI bridge address offset */ 295 if (mmap_state == pci_mmap_mem) { 296 res_bit = IORESOURCE_MEM; 297 } else { 298 io_offset = (unsigned long)pci_ctrl->io_space.base; 299 offset += io_offset; 300 res_bit = IORESOURCE_IO; 301 } 302 303 /* 304 * Check that the offset requested corresponds to one of the 305 * resources of the device. 306 */ 307 for (i = 0; i <= PCI_ROM_RESOURCE; i++) { 308 struct resource *rp = &dev->resource[i]; 309 int flags = rp->flags; 310 311 /* treat ROM as memory (should be already) */ 312 if (i == PCI_ROM_RESOURCE) 313 flags |= IORESOURCE_MEM; 314 315 /* Active and same type? */ 316 if ((flags & res_bit) == 0) 317 continue; 318 319 /* In the range of this resource? */ 320 if (offset < (rp->start & PAGE_MASK) || offset > rp->end) 321 continue; 322 323 /* found it! construct the final physical address */ 324 if (mmap_state == pci_mmap_io) 325 offset += pci_ctrl->io_space.start - io_offset; 326 vma->vm_pgoff = offset >> PAGE_SHIFT; 327 return 0; 328 } 329 330 return -EINVAL; 331 } 332 333 /* 334 * Set vm_page_prot of VMA, as appropriate for this architecture, for a pci 335 * device mapping. 336 */ 337 static __inline__ void 338 __pci_mmap_set_pgprot(struct pci_dev *dev, struct vm_area_struct *vma, 339 enum pci_mmap_state mmap_state, int write_combine) 340 { 341 int prot = pgprot_val(vma->vm_page_prot); 342 343 /* Set to write-through */ 344 prot &= ~_PAGE_NO_CACHE; 345 #if 0 346 if (!write_combine) 347 prot |= _PAGE_WRITETHRU; 348 #endif 349 vma->vm_page_prot = __pgprot(prot); 350 } 351 352 /* 353 * Perform the actual remap of the pages for a PCI device mapping, as 354 * appropriate for this architecture. The region in the process to map 355 * is described by vm_start and vm_end members of VMA, the base physical 356 * address is found in vm_pgoff. 357 * The pci device structure is provided so that architectures may make mapping 358 * decisions on a per-device or per-bus basis. 359 * 360 * Returns a negative error code on failure, zero on success. 361 */ 362 int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma, 363 enum pci_mmap_state mmap_state, 364 int write_combine) 365 { 366 int ret; 367 368 ret = __pci_mmap_make_offset(dev, vma, mmap_state); 369 if (ret < 0) 370 return ret; 371 372 __pci_mmap_set_pgprot(dev, vma, mmap_state, write_combine); 373 374 ret = io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, 375 vma->vm_end - vma->vm_start,vma->vm_page_prot); 376 377 return ret; 378 } 379