xref: /linux/arch/sh/include/cpu-sh4/cpu/sq.h (revision e58e871becec2d3b04ed91c0c16fe8deac9c9dfa)
1 /*
2  * include/asm-sh/cpu-sh4/sq.h
3  *
4  * Copyright (C) 2001, 2002, 2003  Paul Mundt
5  * Copyright (C) 2001, 2002  M. R. Brown
6  *
7  * This file is subject to the terms and conditions of the GNU General Public
8  * License.  See the file "COPYING" in the main directory of this archive
9  * for more details.
10  */
11 #ifndef __ASM_CPU_SH4_SQ_H
12 #define __ASM_CPU_SH4_SQ_H
13 
14 #include <asm/addrspace.h>
15 #include <asm/page.h>
16 
17 /*
18  * Store queues range from e0000000-e3fffffc, allowing approx. 64MB to be
19  * mapped to any physical address space. Since data is written (and aligned)
20  * to 32-byte boundaries, we need to be sure that all allocations are aligned.
21  */
22 #define SQ_SIZE                 32
23 #define SQ_ALIGN_MASK           (~(SQ_SIZE - 1))
24 #define SQ_ALIGN(addr)          (((addr)+SQ_SIZE-1) & SQ_ALIGN_MASK)
25 
26 #define SQ_QACR0		(P4SEG_REG_BASE  + 0x38)
27 #define SQ_QACR1		(P4SEG_REG_BASE  + 0x3c)
28 #define SQ_ADDRMAX              (P4SEG_STORE_QUE + 0x04000000)
29 
30 /* arch/sh/kernel/cpu/sh4/sq.c */
31 unsigned long sq_remap(unsigned long phys, unsigned int size,
32 		       const char *name, pgprot_t prot);
33 void sq_unmap(unsigned long vaddr);
34 void sq_flush_range(unsigned long start, unsigned int len);
35 
36 #endif /* __ASM_CPU_SH4_SQ_H */
37