1 /* 2 * linux/arch/arm/mach-pxa/pxa3xx.c 3 * 4 * code specific to pxa3xx aka Monahans 5 * 6 * Copyright (C) 2006 Marvell International Ltd. 7 * 8 * 2007-09-02: eric miao <eric.miao@marvell.com> 9 * initial version 10 * 11 * This program is free software; you can redistribute it and/or modify 12 * it under the terms of the GNU General Public License version 2 as 13 * published by the Free Software Foundation. 14 */ 15 #include <linux/module.h> 16 #include <linux/kernel.h> 17 #include <linux/init.h> 18 #include <linux/gpio-pxa.h> 19 #include <linux/pm.h> 20 #include <linux/platform_device.h> 21 #include <linux/irq.h> 22 #include <linux/io.h> 23 #include <linux/of.h> 24 #include <linux/syscore_ops.h> 25 #include <linux/i2c/pxa-i2c.h> 26 27 #include <asm/mach/map.h> 28 #include <asm/suspend.h> 29 #include <mach/hardware.h> 30 #include <mach/pxa3xx-regs.h> 31 #include <mach/reset.h> 32 #include <linux/platform_data/usb-ohci-pxa27x.h> 33 #include <mach/pm.h> 34 #include <mach/dma.h> 35 #include <mach/smemc.h> 36 #include <mach/irqs.h> 37 38 #include "generic.h" 39 #include "devices.h" 40 41 #define PECR_IE(n) ((1 << ((n) * 2)) << 28) 42 #define PECR_IS(n) ((1 << ((n) * 2)) << 29) 43 44 extern void __init pxa_dt_irq_init(int (*fn)(struct irq_data *, unsigned int)); 45 46 /* 47 * NAND NFC: DFI bus arbitration subset 48 */ 49 #define NDCR (*(volatile u32 __iomem*)(NAND_VIRT + 0)) 50 #define NDCR_ND_ARB_EN (1 << 12) 51 #define NDCR_ND_ARB_CNTL (1 << 19) 52 53 #ifdef CONFIG_PM 54 55 #define ISRAM_START 0x5c000000 56 #define ISRAM_SIZE SZ_256K 57 58 static void __iomem *sram; 59 static unsigned long wakeup_src; 60 61 /* 62 * Enter a standby mode (S0D1C2 or S0D2C2). Upon wakeup, the dynamic 63 * memory controller has to be reinitialised, so we place some code 64 * in the SRAM to perform this function. 65 * 66 * We disable FIQs across the standby - otherwise, we might receive a 67 * FIQ while the SDRAM is unavailable. 68 */ 69 static void pxa3xx_cpu_standby(unsigned int pwrmode) 70 { 71 extern const char pm_enter_standby_start[], pm_enter_standby_end[]; 72 void (*fn)(unsigned int) = (void __force *)(sram + 0x8000); 73 74 memcpy_toio(sram + 0x8000, pm_enter_standby_start, 75 pm_enter_standby_end - pm_enter_standby_start); 76 77 AD2D0SR = ~0; 78 AD2D1SR = ~0; 79 AD2D0ER = wakeup_src; 80 AD2D1ER = 0; 81 ASCR = ASCR; 82 ARSR = ARSR; 83 84 local_fiq_disable(); 85 fn(pwrmode); 86 local_fiq_enable(); 87 88 AD2D0ER = 0; 89 AD2D1ER = 0; 90 } 91 92 /* 93 * NOTE: currently, the OBM (OEM Boot Module) binary comes along with 94 * PXA3xx development kits assumes that the resuming process continues 95 * with the address stored within the first 4 bytes of SDRAM. The PSPR 96 * register is used privately by BootROM and OBM, and _must_ be set to 97 * 0x5c014000 for the moment. 98 */ 99 static void pxa3xx_cpu_pm_suspend(void) 100 { 101 volatile unsigned long *p = (volatile void *)0xc0000000; 102 unsigned long saved_data = *p; 103 #ifndef CONFIG_IWMMXT 104 u64 acc0; 105 106 asm volatile("mra %Q0, %R0, acc0" : "=r" (acc0)); 107 #endif 108 109 extern int pxa3xx_finish_suspend(unsigned long); 110 111 /* resuming from D2 requires the HSIO2/BOOT/TPM clocks enabled */ 112 CKENA |= (1 << CKEN_BOOT) | (1 << CKEN_TPM); 113 CKENB |= 1 << (CKEN_HSIO2 & 0x1f); 114 115 /* clear and setup wakeup source */ 116 AD3SR = ~0; 117 AD3ER = wakeup_src; 118 ASCR = ASCR; 119 ARSR = ARSR; 120 121 PCFR |= (1u << 13); /* L1_DIS */ 122 PCFR &= ~((1u << 12) | (1u << 1)); /* L0_EN | SL_ROD */ 123 124 PSPR = 0x5c014000; 125 126 /* overwrite with the resume address */ 127 *p = virt_to_phys(cpu_resume); 128 129 cpu_suspend(0, pxa3xx_finish_suspend); 130 131 *p = saved_data; 132 133 AD3ER = 0; 134 135 #ifndef CONFIG_IWMMXT 136 asm volatile("mar acc0, %Q0, %R0" : "=r" (acc0)); 137 #endif 138 } 139 140 static void pxa3xx_cpu_pm_enter(suspend_state_t state) 141 { 142 /* 143 * Don't sleep if no wakeup sources are defined 144 */ 145 if (wakeup_src == 0) { 146 printk(KERN_ERR "Not suspending: no wakeup sources\n"); 147 return; 148 } 149 150 switch (state) { 151 case PM_SUSPEND_STANDBY: 152 pxa3xx_cpu_standby(PXA3xx_PM_S0D2C2); 153 break; 154 155 case PM_SUSPEND_MEM: 156 pxa3xx_cpu_pm_suspend(); 157 break; 158 } 159 } 160 161 static int pxa3xx_cpu_pm_valid(suspend_state_t state) 162 { 163 return state == PM_SUSPEND_MEM || state == PM_SUSPEND_STANDBY; 164 } 165 166 static struct pxa_cpu_pm_fns pxa3xx_cpu_pm_fns = { 167 .valid = pxa3xx_cpu_pm_valid, 168 .enter = pxa3xx_cpu_pm_enter, 169 }; 170 171 static void __init pxa3xx_init_pm(void) 172 { 173 sram = ioremap(ISRAM_START, ISRAM_SIZE); 174 if (!sram) { 175 printk(KERN_ERR "Unable to map ISRAM: disabling standby/suspend\n"); 176 return; 177 } 178 179 /* 180 * Since we copy wakeup code into the SRAM, we need to ensure 181 * that it is preserved over the low power modes. Note: bit 8 182 * is undocumented in the developer manual, but must be set. 183 */ 184 AD1R |= ADXR_L2 | ADXR_R0; 185 AD2R |= ADXR_L2 | ADXR_R0; 186 AD3R |= ADXR_L2 | ADXR_R0; 187 188 /* 189 * Clear the resume enable registers. 190 */ 191 AD1D0ER = 0; 192 AD2D0ER = 0; 193 AD2D1ER = 0; 194 AD3ER = 0; 195 196 pxa_cpu_pm_fns = &pxa3xx_cpu_pm_fns; 197 } 198 199 static int pxa3xx_set_wake(struct irq_data *d, unsigned int on) 200 { 201 unsigned long flags, mask = 0; 202 203 switch (d->irq) { 204 case IRQ_SSP3: 205 mask = ADXER_MFP_WSSP3; 206 break; 207 case IRQ_MSL: 208 mask = ADXER_WMSL0; 209 break; 210 case IRQ_USBH2: 211 case IRQ_USBH1: 212 mask = ADXER_WUSBH; 213 break; 214 case IRQ_KEYPAD: 215 mask = ADXER_WKP; 216 break; 217 case IRQ_AC97: 218 mask = ADXER_MFP_WAC97; 219 break; 220 case IRQ_USIM: 221 mask = ADXER_WUSIM0; 222 break; 223 case IRQ_SSP2: 224 mask = ADXER_MFP_WSSP2; 225 break; 226 case IRQ_I2C: 227 mask = ADXER_MFP_WI2C; 228 break; 229 case IRQ_STUART: 230 mask = ADXER_MFP_WUART3; 231 break; 232 case IRQ_BTUART: 233 mask = ADXER_MFP_WUART2; 234 break; 235 case IRQ_FFUART: 236 mask = ADXER_MFP_WUART1; 237 break; 238 case IRQ_MMC: 239 mask = ADXER_MFP_WMMC1; 240 break; 241 case IRQ_SSP: 242 mask = ADXER_MFP_WSSP1; 243 break; 244 case IRQ_RTCAlrm: 245 mask = ADXER_WRTC; 246 break; 247 case IRQ_SSP4: 248 mask = ADXER_MFP_WSSP4; 249 break; 250 case IRQ_TSI: 251 mask = ADXER_WTSI; 252 break; 253 case IRQ_USIM2: 254 mask = ADXER_WUSIM1; 255 break; 256 case IRQ_MMC2: 257 mask = ADXER_MFP_WMMC2; 258 break; 259 case IRQ_NAND: 260 mask = ADXER_MFP_WFLASH; 261 break; 262 case IRQ_USB2: 263 mask = ADXER_WUSB2; 264 break; 265 case IRQ_WAKEUP0: 266 mask = ADXER_WEXTWAKE0; 267 break; 268 case IRQ_WAKEUP1: 269 mask = ADXER_WEXTWAKE1; 270 break; 271 case IRQ_MMC3: 272 mask = ADXER_MFP_GEN12; 273 break; 274 default: 275 return -EINVAL; 276 } 277 278 local_irq_save(flags); 279 if (on) 280 wakeup_src |= mask; 281 else 282 wakeup_src &= ~mask; 283 local_irq_restore(flags); 284 285 return 0; 286 } 287 #else 288 static inline void pxa3xx_init_pm(void) {} 289 #define pxa3xx_set_wake NULL 290 #endif 291 292 static void pxa_ack_ext_wakeup(struct irq_data *d) 293 { 294 PECR |= PECR_IS(d->irq - IRQ_WAKEUP0); 295 } 296 297 static void pxa_mask_ext_wakeup(struct irq_data *d) 298 { 299 pxa_mask_irq(d); 300 PECR &= ~PECR_IE(d->irq - IRQ_WAKEUP0); 301 } 302 303 static void pxa_unmask_ext_wakeup(struct irq_data *d) 304 { 305 pxa_unmask_irq(d); 306 PECR |= PECR_IE(d->irq - IRQ_WAKEUP0); 307 } 308 309 static int pxa_set_ext_wakeup_type(struct irq_data *d, unsigned int flow_type) 310 { 311 if (flow_type & IRQ_TYPE_EDGE_RISING) 312 PWER |= 1 << (d->irq - IRQ_WAKEUP0); 313 314 if (flow_type & IRQ_TYPE_EDGE_FALLING) 315 PWER |= 1 << (d->irq - IRQ_WAKEUP0 + 2); 316 317 return 0; 318 } 319 320 static struct irq_chip pxa_ext_wakeup_chip = { 321 .name = "WAKEUP", 322 .irq_ack = pxa_ack_ext_wakeup, 323 .irq_mask = pxa_mask_ext_wakeup, 324 .irq_unmask = pxa_unmask_ext_wakeup, 325 .irq_set_type = pxa_set_ext_wakeup_type, 326 }; 327 328 static void __init pxa_init_ext_wakeup_irq(int (*fn)(struct irq_data *, 329 unsigned int)) 330 { 331 int irq; 332 333 for (irq = IRQ_WAKEUP0; irq <= IRQ_WAKEUP1; irq++) { 334 irq_set_chip_and_handler(irq, &pxa_ext_wakeup_chip, 335 handle_edge_irq); 336 irq_clear_status_flags(irq, IRQ_NOREQUEST); 337 } 338 339 pxa_ext_wakeup_chip.irq_set_wake = fn; 340 } 341 342 static void __init __pxa3xx_init_irq(void) 343 { 344 /* enable CP6 access */ 345 u32 value; 346 __asm__ __volatile__("mrc p15, 0, %0, c15, c1, 0\n": "=r"(value)); 347 value |= (1 << 6); 348 __asm__ __volatile__("mcr p15, 0, %0, c15, c1, 0\n": :"r"(value)); 349 350 pxa_init_ext_wakeup_irq(pxa3xx_set_wake); 351 } 352 353 void __init pxa3xx_init_irq(void) 354 { 355 __pxa3xx_init_irq(); 356 pxa_init_irq(56, pxa3xx_set_wake); 357 } 358 359 #ifdef CONFIG_OF 360 void __init pxa3xx_dt_init_irq(void) 361 { 362 __pxa3xx_init_irq(); 363 pxa_dt_irq_init(pxa3xx_set_wake); 364 } 365 #endif /* CONFIG_OF */ 366 367 static struct map_desc pxa3xx_io_desc[] __initdata = { 368 { /* Mem Ctl */ 369 .virtual = (unsigned long)SMEMC_VIRT, 370 .pfn = __phys_to_pfn(PXA3XX_SMEMC_BASE), 371 .length = SMEMC_SIZE, 372 .type = MT_DEVICE 373 }, { 374 .virtual = (unsigned long)NAND_VIRT, 375 .pfn = __phys_to_pfn(NAND_PHYS), 376 .length = NAND_SIZE, 377 .type = MT_DEVICE 378 }, 379 }; 380 381 void __init pxa3xx_map_io(void) 382 { 383 pxa_map_io(); 384 iotable_init(ARRAY_AND_SIZE(pxa3xx_io_desc)); 385 pxa3xx_get_clk_frequency_khz(1); 386 } 387 388 /* 389 * device registration specific to PXA3xx. 390 */ 391 392 void __init pxa3xx_set_i2c_power_info(struct i2c_pxa_platform_data *info) 393 { 394 pxa_register_device(&pxa3xx_device_i2c_power, info); 395 } 396 397 static struct pxa_gpio_platform_data pxa3xx_gpio_pdata = { 398 .irq_base = PXA_GPIO_TO_IRQ(0), 399 }; 400 401 static struct platform_device *devices[] __initdata = { 402 &pxa27x_device_udc, 403 &pxa_device_pmu, 404 &pxa_device_i2s, 405 &pxa_device_asoc_ssp1, 406 &pxa_device_asoc_ssp2, 407 &pxa_device_asoc_ssp3, 408 &pxa_device_asoc_ssp4, 409 &pxa_device_asoc_platform, 410 &pxa_device_rtc, 411 &pxa3xx_device_ssp1, 412 &pxa3xx_device_ssp2, 413 &pxa3xx_device_ssp3, 414 &pxa3xx_device_ssp4, 415 &pxa27x_device_pwm0, 416 &pxa27x_device_pwm1, 417 }; 418 419 static int __init pxa3xx_init(void) 420 { 421 int ret = 0; 422 423 if (cpu_is_pxa3xx()) { 424 425 reset_status = ARSR; 426 427 /* 428 * clear RDH bit every time after reset 429 * 430 * Note: the last 3 bits DxS are write-1-to-clear so carefully 431 * preserve them here in case they will be referenced later 432 */ 433 ASCR &= ~(ASCR_RDH | ASCR_D1S | ASCR_D2S | ASCR_D3S); 434 435 /* 436 * Disable DFI bus arbitration, to prevent a system bus lock if 437 * somebody disables the NAND clock (unused clock) while this 438 * bit remains set. 439 */ 440 NDCR = (NDCR & ~NDCR_ND_ARB_EN) | NDCR_ND_ARB_CNTL; 441 442 if ((ret = pxa_init_dma(IRQ_DMA, 32))) 443 return ret; 444 445 pxa3xx_init_pm(); 446 447 register_syscore_ops(&pxa_irq_syscore_ops); 448 register_syscore_ops(&pxa3xx_mfp_syscore_ops); 449 450 if (of_have_populated_dt()) 451 return 0; 452 453 pxa2xx_set_dmac_info(32); 454 ret = platform_add_devices(devices, ARRAY_SIZE(devices)); 455 if (ret) 456 return ret; 457 if (cpu_is_pxa300() || cpu_is_pxa310() || cpu_is_pxa320()) { 458 platform_device_add_data(&pxa3xx_device_gpio, 459 &pxa3xx_gpio_pdata, 460 sizeof(pxa3xx_gpio_pdata)); 461 ret = platform_device_register(&pxa3xx_device_gpio); 462 } 463 } 464 465 return ret; 466 } 467 468 postcore_initcall(pxa3xx_init); 469