1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * linux/arch/alpha/kernel/core_irongate.c
4 *
5 * Based on code written by David A. Rusling (david.rusling@reo.mts.dec.com).
6 *
7 * Copyright (C) 1999 Alpha Processor, Inc.,
8 * (David Daniel, Stig Telfer, Soohoon Lee)
9 *
10 * Code common to all IRONGATE core logic chips.
11 */
12
13 #define __EXTERN_INLINE inline
14 #include <asm/io.h>
15 #include <asm/core_irongate.h>
16 #undef __EXTERN_INLINE
17
18 #include <linux/types.h>
19 #include <linux/pci.h>
20 #include <linux/sched.h>
21 #include <linux/init.h>
22 #include <linux/initrd.h>
23 #include <linux/memblock.h>
24
25 #include <asm/ptrace.h>
26 #include <asm/cacheflush.h>
27 #include <asm/tlbflush.h>
28
29 #include "proto.h"
30 #include "pci_impl.h"
31
32 /*
33 * BIOS32-style PCI interface:
34 */
35
36 #define DEBUG_CONFIG 0
37
38 #if DEBUG_CONFIG
39 # define DBG_CFG(args) printk args
40 #else
41 # define DBG_CFG(args)
42 #endif
43
44 igcsr32 *IronECC;
45
46 /*
47 * Given a bus, device, and function number, compute resulting
48 * configuration space address accordingly. It is therefore not safe
49 * to have concurrent invocations to configuration space access
50 * routines, but there really shouldn't be any need for this.
51 *
52 * addr[31:24] reserved
53 * addr[23:16] bus number (8 bits = 128 possible buses)
54 * addr[15:11] Device number (5 bits)
55 * addr[10: 8] function number
56 * addr[ 7: 2] register number
57 *
58 * For IRONGATE:
59 * if (bus = addr[23:16]) == 0
60 * then
61 * type 0 config cycle:
62 * addr_on_pci[31:11] = id selection for device = addr[15:11]
63 * addr_on_pci[10: 2] = addr[10: 2] ???
64 * addr_on_pci[ 1: 0] = 00
65 * else
66 * type 1 config cycle (pass on with no decoding):
67 * addr_on_pci[31:24] = 0
68 * addr_on_pci[23: 2] = addr[23: 2]
69 * addr_on_pci[ 1: 0] = 01
70 * fi
71 *
72 * Notes:
73 * The function number selects which function of a multi-function device
74 * (e.g., SCSI and Ethernet).
75 *
76 * The register selects a DWORD (32 bit) register offset. Hence it
77 * doesn't get shifted by 2 bits as we want to "drop" the bottom two
78 * bits.
79 */
80
81 static int
mk_conf_addr(struct pci_bus * pbus,unsigned int device_fn,int where,unsigned long * pci_addr,unsigned char * type1)82 mk_conf_addr(struct pci_bus *pbus, unsigned int device_fn, int where,
83 unsigned long *pci_addr, unsigned char *type1)
84 {
85 unsigned long addr;
86 u8 bus = pbus->number;
87
88 DBG_CFG(("mk_conf_addr(bus=%d ,device_fn=0x%x, where=0x%x, "
89 "pci_addr=0x%p, type1=0x%p)\n",
90 bus, device_fn, where, pci_addr, type1));
91
92 *type1 = (bus != 0);
93
94 addr = (bus << 16) | (device_fn << 8) | where;
95 addr |= IRONGATE_CONF;
96
97 *pci_addr = addr;
98 DBG_CFG(("mk_conf_addr: returning pci_addr 0x%lx\n", addr));
99 return 0;
100 }
101
102 static int
irongate_read_config(struct pci_bus * bus,unsigned int devfn,int where,int size,u32 * value)103 irongate_read_config(struct pci_bus *bus, unsigned int devfn, int where,
104 int size, u32 *value)
105 {
106 unsigned long addr;
107 unsigned char type1;
108
109 if (mk_conf_addr(bus, devfn, where, &addr, &type1))
110 return PCIBIOS_DEVICE_NOT_FOUND;
111
112 switch (size) {
113 case 1:
114 *value = __kernel_ldbu(*(vucp)addr);
115 break;
116 case 2:
117 *value = __kernel_ldwu(*(vusp)addr);
118 break;
119 case 4:
120 *value = *(vuip)addr;
121 break;
122 }
123
124 return PCIBIOS_SUCCESSFUL;
125 }
126
127 static int
irongate_write_config(struct pci_bus * bus,unsigned int devfn,int where,int size,u32 value)128 irongate_write_config(struct pci_bus *bus, unsigned int devfn, int where,
129 int size, u32 value)
130 {
131 unsigned long addr;
132 unsigned char type1;
133
134 if (mk_conf_addr(bus, devfn, where, &addr, &type1))
135 return PCIBIOS_DEVICE_NOT_FOUND;
136
137 switch (size) {
138 case 1:
139 __kernel_stb(value, *(vucp)addr);
140 mb();
141 __kernel_ldbu(*(vucp)addr);
142 break;
143 case 2:
144 __kernel_stw(value, *(vusp)addr);
145 mb();
146 __kernel_ldwu(*(vusp)addr);
147 break;
148 case 4:
149 *(vuip)addr = value;
150 mb();
151 *(vuip)addr;
152 break;
153 }
154
155 return PCIBIOS_SUCCESSFUL;
156 }
157
158 struct pci_ops irongate_pci_ops =
159 {
160 .read = irongate_read_config,
161 .write = irongate_write_config,
162 };
163
164 int
irongate_pci_clr_err(void)165 irongate_pci_clr_err(void)
166 {
167 unsigned int nmi_ctl=0;
168 unsigned int IRONGATE_jd;
169
170 again:
171 IRONGATE_jd = IRONGATE0->stat_cmd;
172 printk("Iron stat_cmd %x\n", IRONGATE_jd);
173 IRONGATE0->stat_cmd = IRONGATE_jd; /* write again clears error bits */
174 mb();
175 IRONGATE_jd = IRONGATE0->stat_cmd; /* re-read to force write */
176
177 IRONGATE_jd = *IronECC;
178 printk("Iron ECC %x\n", IRONGATE_jd);
179 *IronECC = IRONGATE_jd; /* write again clears error bits */
180 mb();
181 IRONGATE_jd = *IronECC; /* re-read to force write */
182
183 /* Clear ALI NMI */
184 nmi_ctl = inb(0x61);
185 nmi_ctl |= 0x0c;
186 outb(nmi_ctl, 0x61);
187 nmi_ctl &= ~0x0c;
188 outb(nmi_ctl, 0x61);
189
190 IRONGATE_jd = *IronECC;
191 if (IRONGATE_jd & 0x300) goto again;
192
193 return 0;
194 }
195
196 #define IRONGATE_3GB 0xc0000000UL
197
198 /* On Albacore (aka UP1500) with 4Gb of RAM we have to reserve some
199 memory for PCI. At this point we just reserve memory above 3Gb. Most
200 of this memory will be freed after PCI setup is done. */
201 static void __init
albacore_init_arch(void)202 albacore_init_arch(void)
203 {
204 unsigned long memtop = max_low_pfn << PAGE_SHIFT;
205 unsigned long pci_mem = (memtop + 0x1000000UL) & ~0xffffffUL;
206 struct percpu_struct *cpu;
207 int pal_rev, pal_var;
208
209 cpu = (struct percpu_struct*)((char*)hwrpb + hwrpb->processor_offset);
210 pal_rev = cpu->pal_revision & 0xffff;
211 pal_var = (cpu->pal_revision >> 16) & 0xff;
212
213 /* Consoles earlier than A5.6-18 (OSF PALcode v1.62-2) set up
214 the CPU incorrectly (leave speculative stores enabled),
215 which causes memory corruption under certain conditions.
216 Issue a warning for such consoles. */
217 if (alpha_using_srm &&
218 (pal_rev < 0x13e || (pal_rev == 0x13e && pal_var < 2)))
219 printk(KERN_WARNING "WARNING! Upgrade to SRM A5.6-19 "
220 "or later\n");
221
222 if (pci_mem > IRONGATE_3GB)
223 pci_mem = IRONGATE_3GB;
224 IRONGATE0->pci_mem = pci_mem;
225 alpha_mv.min_mem_address = pci_mem;
226 if (memtop > pci_mem) {
227 #ifdef CONFIG_BLK_DEV_INITRD
228 extern unsigned long initrd_start, initrd_end;
229
230 /* Move the initrd out of the way. */
231 if (initrd_end && __pa(initrd_end) > pci_mem) {
232 unsigned long size;
233
234 size = initrd_end - initrd_start;
235 memblock_free((void *)initrd_start, PAGE_ALIGN(size));
236 if (!move_initrd(pci_mem))
237 printk("irongate_init_arch: initrd too big "
238 "(%ldK)\ndisabling initrd\n",
239 size / 1024);
240 }
241 #endif
242 memblock_reserve(pci_mem, memtop - pci_mem);
243 printk("irongate_init_arch: temporarily reserving "
244 "region %08lx-%08lx for PCI\n", pci_mem, memtop - 1);
245 }
246 }
247
248 static void __init
irongate_setup_agp(void)249 irongate_setup_agp(void)
250 {
251 /* Disable the GART window. AGPGART doesn't work due to yet
252 unresolved memory coherency issues... */
253 IRONGATE0->agpva = IRONGATE0->agpva & ~0xf;
254 alpha_agpgart_size = 0;
255 }
256
257 void __init
irongate_init_arch(void)258 irongate_init_arch(void)
259 {
260 struct pci_controller *hose;
261 int amd761 = (IRONGATE0->dev_vendor >> 16) > 0x7006; /* Albacore? */
262
263 IronECC = amd761 ? &IRONGATE0->bacsr54_eccms761 : &IRONGATE0->dramms;
264
265 irongate_pci_clr_err();
266
267 if (amd761)
268 albacore_init_arch();
269
270 irongate_setup_agp();
271
272 /*
273 * Create our single hose.
274 */
275
276 pci_isa_hose = hose = alloc_pci_controller();
277 hose->io_space = &ioport_resource;
278 hose->mem_space = &iomem_resource;
279 hose->index = 0;
280
281 /* This is for userland consumption. For some reason, the 40-bit
282 PIO bias that we use in the kernel through KSEG didn't work for
283 the page table based user mappings. So make sure we get the
284 43-bit PIO bias. */
285 hose->sparse_mem_base = 0;
286 hose->sparse_io_base = 0;
287 hose->dense_mem_base
288 = (IRONGATE_MEM & 0xffffffffffUL) | 0x80000000000UL;
289 hose->dense_io_base
290 = (IRONGATE_IO & 0xffffffffffUL) | 0x80000000000UL;
291
292 hose->sg_isa = hose->sg_pci = NULL;
293 __direct_map_base = 0;
294 __direct_map_size = 0xffffffff;
295 }
296
297 /*
298 * IO map and AGP support
299 */
300 #include <linux/vmalloc.h>
301 #include <linux/agp_backend.h>
302 #include <linux/agpgart.h>
303 #include <linux/export.h>
304
305 #define GET_PAGE_DIR_OFF(addr) (addr >> 22)
306 #define GET_PAGE_DIR_IDX(addr) (GET_PAGE_DIR_OFF(addr))
307
308 #define GET_GATT_OFF(addr) ((addr & 0x003ff000) >> 12)
309 #define GET_GATT(addr) (gatt_pages[GET_PAGE_DIR_IDX(addr)])
310
311 void __iomem *
irongate_ioremap(unsigned long addr,unsigned long size)312 irongate_ioremap(unsigned long addr, unsigned long size)
313 {
314 struct vm_struct *area;
315 unsigned long vaddr;
316 unsigned long baddr, last;
317 u32 *mmio_regs, *gatt_pages, *cur_gatt, pte;
318 unsigned long gart_bus_addr;
319
320 if (!alpha_agpgart_size)
321 return (void __iomem *)(addr + IRONGATE_MEM);
322
323 gart_bus_addr = (unsigned long)IRONGATE0->bar0 &
324 PCI_BASE_ADDRESS_MEM_MASK;
325
326 /*
327 * Check for within the AGP aperture...
328 */
329 do {
330 /*
331 * Check the AGP area
332 */
333 if (addr >= gart_bus_addr && addr + size - 1 <
334 gart_bus_addr + alpha_agpgart_size)
335 break;
336
337 /*
338 * Not found - assume legacy ioremap
339 */
340 return (void __iomem *)(addr + IRONGATE_MEM);
341 } while(0);
342
343 mmio_regs = (u32 *)(((unsigned long)IRONGATE0->bar1 &
344 PCI_BASE_ADDRESS_MEM_MASK) + IRONGATE_MEM);
345
346 gatt_pages = (u32 *)(phys_to_virt(mmio_regs[1])); /* FIXME */
347
348 /*
349 * Adjust the limits (mappings must be page aligned)
350 */
351 if (addr & ~PAGE_MASK) {
352 printk("AGP ioremap failed... addr not page aligned (0x%lx)\n",
353 addr);
354 return (void __iomem *)(addr + IRONGATE_MEM);
355 }
356 last = addr + size - 1;
357 size = PAGE_ALIGN(last) - addr;
358
359 #if 0
360 printk("irongate_ioremap(0x%lx, 0x%lx)\n", addr, size);
361 printk("irongate_ioremap: gart_bus_addr 0x%lx\n", gart_bus_addr);
362 printk("irongate_ioremap: gart_aper_size 0x%lx\n", gart_aper_size);
363 printk("irongate_ioremap: mmio_regs %p\n", mmio_regs);
364 printk("irongate_ioremap: gatt_pages %p\n", gatt_pages);
365
366 for(baddr = addr; baddr <= last; baddr += PAGE_SIZE)
367 {
368 cur_gatt = phys_to_virt(GET_GATT(baddr) & ~1);
369 pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1;
370 printk("irongate_ioremap: cur_gatt %p pte 0x%x\n",
371 cur_gatt, pte);
372 }
373 #endif
374
375 /*
376 * Map it
377 */
378 area = get_vm_area(size, VM_IOREMAP);
379 if (!area) return NULL;
380
381 for(baddr = addr, vaddr = (unsigned long)area->addr;
382 baddr <= last;
383 baddr += PAGE_SIZE, vaddr += PAGE_SIZE)
384 {
385 cur_gatt = phys_to_virt(GET_GATT(baddr) & ~1);
386 pte = cur_gatt[GET_GATT_OFF(baddr)] & ~1;
387
388 if (__alpha_remap_area_pages(vaddr,
389 pte, PAGE_SIZE, 0)) {
390 printk("AGP ioremap: FAILED to map...\n");
391 vfree(area->addr);
392 return NULL;
393 }
394 }
395
396 flush_tlb_all();
397
398 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK);
399 #if 0
400 printk("irongate_ioremap(0x%lx, 0x%lx) returning 0x%lx\n",
401 addr, size, vaddr);
402 #endif
403 return (void __iomem *)vaddr;
404 }
405 EXPORT_SYMBOL(irongate_ioremap);
406
407 void
irongate_iounmap(volatile void __iomem * xaddr)408 irongate_iounmap(volatile void __iomem *xaddr)
409 {
410 unsigned long addr = (unsigned long) xaddr;
411 if (((long)addr >> 41) == -2)
412 return; /* kseg map, nothing to do */
413 if (addr)
414 return vfree((void *)(PAGE_MASK & addr));
415 }
416 EXPORT_SYMBOL(irongate_iounmap);
417