/linux/arch/sh/mm/ |
H A D | flush-sh4.c | 19 v = aligned_start & ~(L1_CACHE_BYTES-1); in sh4__flush_wback_region() 20 end = (aligned_start + size + L1_CACHE_BYTES-1) in sh4__flush_wback_region() 21 & ~(L1_CACHE_BYTES-1); in sh4__flush_wback_region() 22 cnt = (end - v) / L1_CACHE_BYTES; in sh4__flush_wback_region() 25 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 26 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 27 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 28 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 29 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 30 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() [all …]
|
H A D | cache-sh2a.c | 57 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh2a__flush_wback_region() 58 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) in sh2a__flush_wback_region() 59 & ~(L1_CACHE_BYTES-1); in sh2a__flush_wback_region() 70 for (v = begin; v < end; v += L1_CACHE_BYTES) { in sh2a__flush_wback_region() 78 for (v = begin; v < end; v += L1_CACHE_BYTES) in sh2a__flush_wback_region() 97 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh2a__flush_purge_region() 98 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) in sh2a__flush_purge_region() 99 & ~(L1_CACHE_BYTES-1); in sh2a__flush_purge_region() 104 for (v = begin; v < end; v+=L1_CACHE_BYTES) { in sh2a__flush_purge_region() 127 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh2a__flush_invalidate_region() [all …]
|
H A D | cache-sh2.c | 23 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh2__flush_wback_region() 24 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) in sh2__flush_wback_region() 25 & ~(L1_CACHE_BYTES-1); in sh2__flush_wback_region() 26 for (v = begin; v < end; v+=L1_CACHE_BYTES) { in sh2__flush_wback_region() 44 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh2__flush_purge_region() 45 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) in sh2__flush_purge_region() 46 & ~(L1_CACHE_BYTES-1); in sh2__flush_purge_region() 48 for (v = begin; v < end; v+=L1_CACHE_BYTES) in sh2__flush_purge_region() 75 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh2__flush_invalidate_region() 76 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) in sh2__flush_invalidate_region() [all …]
|
H A D | cache-sh3.c | 38 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh3__flush_wback_region() 39 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) in sh3__flush_wback_region() 40 & ~(L1_CACHE_BYTES-1); in sh3__flush_wback_region() 42 for (v = begin; v < end; v+=L1_CACHE_BYTES) { in sh3__flush_wback_region() 76 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); in sh3__flush_purge_region() 77 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) in sh3__flush_purge_region() 78 & ~(L1_CACHE_BYTES-1); in sh3__flush_purge_region() 80 for (v = begin; v < end; v+=L1_CACHE_BYTES) { in sh3__flush_purge_region()
|
/linux/arch/csky/mm/ |
H A D | cachev2.c | 26 unsigned long i = start & ~(L1_CACHE_BYTES - 1); in icache_inv_range() 28 for (; i < end; i += L1_CACHE_BYTES) in icache_inv_range() 49 unsigned long i = param->start & ~(L1_CACHE_BYTES - 1); in local_icache_inv_range() 54 for (; i < param->end; i += L1_CACHE_BYTES) in local_icache_inv_range() 81 unsigned long i = start & ~(L1_CACHE_BYTES - 1); in dcache_wb_range() 83 for (; i < end; i += L1_CACHE_BYTES) in dcache_wb_range() 97 unsigned long i = start & ~(L1_CACHE_BYTES - 1); in dma_wbinv_range() 99 for (; i < end; i += L1_CACHE_BYTES) in dma_wbinv_range() 106 unsigned long i = start & ~(L1_CACHE_BYTES - 1); in dma_inv_range() 108 for (; i < end; i += L1_CACHE_BYTES) in dma_inv_range() [all …]
|
/linux/arch/microblaze/include/asm/ |
H A D | cache.h | 17 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) macro 19 #define SMP_CACHE_BYTES L1_CACHE_BYTES 22 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES 24 #define ARCH_SLAB_MINALIGN L1_CACHE_BYTES
|
/linux/arch/arm/lib/ |
H A D | copy_page.S | 14 #define COPY_COUNT (PAGE_SZ / (2 * L1_CACHE_BYTES) PLD( -1 )) 27 PLD( pld [r1, #L1_CACHE_BYTES] ) 30 1: PLD( pld [r1, #2 * L1_CACHE_BYTES]) 31 PLD( pld [r1, #3 * L1_CACHE_BYTES]) 33 .rept (2 * L1_CACHE_BYTES / 16 - 1)
|
/linux/arch/hexagon/include/asm/ |
H A D | cache.h | 13 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) macro 15 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES 17 #define __cacheline_aligned __aligned(L1_CACHE_BYTES) 18 #define ____cacheline_aligned __aligned(L1_CACHE_BYTES)
|
/linux/arch/arc/kernel/ |
H A D | vmlinux.lds.S | 62 INIT_TEXT_SECTION(L1_CACHE_BYTES) 67 INIT_SETUP(L1_CACHE_BYTES) 78 PERCPU_SECTION(L1_CACHE_BYTES) 95 EXCEPTION_TABLE(L1_CACHE_BYTES) 105 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE)
|
/linux/arch/powerpc/include/asm/ |
H A D | cache.h | 30 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) macro 32 #define SMP_CACHE_BYTES L1_CACHE_BYTES 37 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES 89 return L1_CACHE_BYTES; in l1_dcache_bytes() 99 return L1_CACHE_BYTES; in l1_icache_bytes()
|
H A D | page_32.h | 44 WARN_ON((unsigned long)addr & (L1_CACHE_BYTES - 1)); in clear_page() 46 for (i = 0; i < PAGE_SIZE / L1_CACHE_BYTES; i++, addr += L1_CACHE_BYTES) in clear_page()
|
/linux/arch/csky/kernel/ |
H A D | vmlinux.lds.S | 49 PERCPU_SECTION(L1_CACHE_BYTES) 55 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE) 106 EXCEPTION_TABLE(L1_CACHE_BYTES) 107 BSS_SECTION(L1_CACHE_BYTES, PAGE_SIZE, L1_CACHE_BYTES)
|
/linux/arch/alpha/include/asm/ |
H A D | cache.h | 11 # define L1_CACHE_BYTES 64 macro 17 # define L1_CACHE_BYTES 32 macro 21 #define SMP_CACHE_BYTES L1_CACHE_BYTES
|
/linux/arch/xtensa/include/asm/ |
H A D | cache.h | 17 #define L1_CACHE_BYTES XCHAL_DCACHE_LINESIZE macro 18 #define SMP_CACHE_BYTES L1_CACHE_BYTES 32 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES
|
/linux/drivers/md/dm-vdo/ |
H A D | cpu.h | 50 unsigned int offset = ((uintptr_t) address % L1_CACHE_BYTES); in uds_prefetch_range() 51 unsigned int cache_lines = (1 + ((size + offset) / L1_CACHE_BYTES)); in uds_prefetch_range() 55 address += L1_CACHE_BYTES; in uds_prefetch_range()
|
/linux/arch/nios2/kernel/ |
H A D | vmlinux.lds.S | 41 EXCEPTION_TABLE(L1_CACHE_BYTES) 47 PERCPU_SECTION(L1_CACHE_BYTES) 52 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE)
|
/linux/arch/powerpc/lib/ |
H A D | copy_32.S | 61 CACHELINE_BYTES = L1_CACHE_BYTES 63 CACHELINE_MASK = (L1_CACHE_BYTES-1) 211 #if L1_CACHE_BYTES >= 32 213 #if L1_CACHE_BYTES >= 64 216 #if L1_CACHE_BYTES >= 128 390 #if L1_CACHE_BYTES >= 32 392 #if L1_CACHE_BYTES >= 64 395 #if L1_CACHE_BYTES >= 128 448 #if L1_CACHE_BYTES >= 32 450 #if L1_CACHE_BYTES >= 64 [all …]
|
H A D | checksum_32.S | 120 CACHELINE_BYTES = L1_CACHE_BYTES 122 CACHELINE_MASK = (L1_CACHE_BYTES-1) 195 #if L1_CACHE_BYTES >= 32 197 #if L1_CACHE_BYTES >= 64 200 #if L1_CACHE_BYTES >= 128 255 #if L1_CACHE_BYTES >= 32 257 #if L1_CACHE_BYTES >= 64 260 #if L1_CACHE_BYTES >= 128
|
/linux/arch/arc/mm/ |
H A D | cache.c | 213 num_lines = DIV_ROUND_UP(sz, L1_CACHE_BYTES); in __cache_line_loop_v3() 235 paddr += L1_CACHE_BYTES; in __cache_line_loop_v3() 239 vaddr += L1_CACHE_BYTES; in __cache_line_loop_v3() 272 num_lines = DIV_ROUND_UP(sz, L1_CACHE_BYTES); in __cache_line_loop_v4() 292 paddr += L1_CACHE_BYTES; in __cache_line_loop_v4() 325 sz += L1_CACHE_BYTES - 1; in __cache_line_loop_v4() 1003 if (ic->line_len != L1_CACHE_BYTES) in arc_cache_init_master() 1005 ic->line_len, L1_CACHE_BYTES); in arc_cache_init_master() 1023 if (dc->line_len != L1_CACHE_BYTES) in arc_cache_init_master() 1025 dc->line_len, L1_CACHE_BYTES); in arc_cache_init_master() [all …]
|
/linux/drivers/misc/sgi-xp/ |
H A D | xpnet.c | 121 #define XPNET_MAX_MTU (0x800000UL - L1_CACHE_BYTES) 170 skb = dev_alloc_skb(msg->size + L1_CACHE_BYTES); in xpnet_receive() 173 msg->size + L1_CACHE_BYTES); in xpnet_receive() 187 skb_reserve(skb, (L1_CACHE_BYTES - ((u64)skb->data & in xpnet_receive() 188 (L1_CACHE_BYTES - 1)) + in xpnet_receive() 209 dst = (void *)((u64)skb->data & ~(L1_CACHE_BYTES - 1)); in xpnet_receive() 445 start_addr = ((u64)skb->data & ~(L1_CACHE_BYTES - 1)); in xpnet_dev_hard_start_xmit()
|
/linux/arch/riscv/kernel/ |
H A D | vmlinux-xip.lds.S | 52 RO_DATA(L1_CACHE_BYTES) 76 RW_DATA(L1_CACHE_BYTES, PAGE_SIZE, THREAD_SIZE) 122 PERCPU_SECTION(L1_CACHE_BYTES)
|
/linux/include/linux/ |
H A D | cache.h | 9 #define L1_CACHE_ALIGN(x) __ALIGN_KERNEL(x, L1_CACHE_BYTES) 13 #define SMP_CACHE_BYTES L1_CACHE_BYTES 111 #define cache_line_size() L1_CACHE_BYTES
|
/linux/arch/m68k/include/asm/ |
H A D | cache.h | 10 #define L1_CACHE_BYTES (1<< L1_CACHE_SHIFT) macro 12 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES
|
/linux/tools/include/linux/ |
H A D | cache.h | 6 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) macro 8 #define SMP_CACHE_BYTES L1_CACHE_BYTES
|
/linux/arch/arm/include/asm/ |
H A D | cache.h | 9 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) macro 18 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES
|