1 /* 2 * mm/debug.c 3 * 4 * mm/ specific debug routines. 5 * 6 */ 7 8 #include <linux/kernel.h> 9 #include <linux/mm.h> 10 #include <linux/trace_events.h> 11 #include <linux/memcontrol.h> 12 13 static const struct trace_print_flags pageflag_names[] = { 14 {1UL << PG_locked, "locked" }, 15 {1UL << PG_error, "error" }, 16 {1UL << PG_referenced, "referenced" }, 17 {1UL << PG_uptodate, "uptodate" }, 18 {1UL << PG_dirty, "dirty" }, 19 {1UL << PG_lru, "lru" }, 20 {1UL << PG_active, "active" }, 21 {1UL << PG_slab, "slab" }, 22 {1UL << PG_owner_priv_1, "owner_priv_1" }, 23 {1UL << PG_arch_1, "arch_1" }, 24 {1UL << PG_reserved, "reserved" }, 25 {1UL << PG_private, "private" }, 26 {1UL << PG_private_2, "private_2" }, 27 {1UL << PG_writeback, "writeback" }, 28 #ifdef CONFIG_PAGEFLAGS_EXTENDED 29 {1UL << PG_head, "head" }, 30 {1UL << PG_tail, "tail" }, 31 #else 32 {1UL << PG_compound, "compound" }, 33 #endif 34 {1UL << PG_swapcache, "swapcache" }, 35 {1UL << PG_mappedtodisk, "mappedtodisk" }, 36 {1UL << PG_reclaim, "reclaim" }, 37 {1UL << PG_swapbacked, "swapbacked" }, 38 {1UL << PG_unevictable, "unevictable" }, 39 #ifdef CONFIG_MMU 40 {1UL << PG_mlocked, "mlocked" }, 41 #endif 42 #ifdef CONFIG_ARCH_USES_PG_UNCACHED 43 {1UL << PG_uncached, "uncached" }, 44 #endif 45 #ifdef CONFIG_MEMORY_FAILURE 46 {1UL << PG_hwpoison, "hwpoison" }, 47 #endif 48 #ifdef CONFIG_TRANSPARENT_HUGEPAGE 49 {1UL << PG_compound_lock, "compound_lock" }, 50 #endif 51 #if defined(CONFIG_IDLE_PAGE_TRACKING) && defined(CONFIG_64BIT) 52 {1UL << PG_young, "young" }, 53 {1UL << PG_idle, "idle" }, 54 #endif 55 }; 56 57 static void dump_flags(unsigned long flags, 58 const struct trace_print_flags *names, int count) 59 { 60 const char *delim = ""; 61 unsigned long mask; 62 int i; 63 64 pr_emerg("flags: %#lx(", flags); 65 66 /* remove zone id */ 67 flags &= (1UL << NR_PAGEFLAGS) - 1; 68 69 for (i = 0; i < count && flags; i++) { 70 71 mask = names[i].mask; 72 if ((flags & mask) != mask) 73 continue; 74 75 flags &= ~mask; 76 pr_cont("%s%s", delim, names[i].name); 77 delim = "|"; 78 } 79 80 /* check for left over flags */ 81 if (flags) 82 pr_cont("%s%#lx", delim, flags); 83 84 pr_cont(")\n"); 85 } 86 87 void dump_page_badflags(struct page *page, const char *reason, 88 unsigned long badflags) 89 { 90 pr_emerg("page:%p count:%d mapcount:%d mapping:%p index:%#lx\n", 91 page, atomic_read(&page->_count), page_mapcount(page), 92 page->mapping, page->index); 93 BUILD_BUG_ON(ARRAY_SIZE(pageflag_names) != __NR_PAGEFLAGS); 94 dump_flags(page->flags, pageflag_names, ARRAY_SIZE(pageflag_names)); 95 if (reason) 96 pr_alert("page dumped because: %s\n", reason); 97 if (page->flags & badflags) { 98 pr_alert("bad because of flags:\n"); 99 dump_flags(page->flags & badflags, 100 pageflag_names, ARRAY_SIZE(pageflag_names)); 101 } 102 #ifdef CONFIG_MEMCG 103 if (page->mem_cgroup) 104 pr_alert("page->mem_cgroup:%p\n", page->mem_cgroup); 105 #endif 106 } 107 108 void dump_page(struct page *page, const char *reason) 109 { 110 dump_page_badflags(page, reason, 0); 111 } 112 EXPORT_SYMBOL(dump_page); 113 114 #ifdef CONFIG_DEBUG_VM 115 116 static const struct trace_print_flags vmaflags_names[] = { 117 {VM_READ, "read" }, 118 {VM_WRITE, "write" }, 119 {VM_EXEC, "exec" }, 120 {VM_SHARED, "shared" }, 121 {VM_MAYREAD, "mayread" }, 122 {VM_MAYWRITE, "maywrite" }, 123 {VM_MAYEXEC, "mayexec" }, 124 {VM_MAYSHARE, "mayshare" }, 125 {VM_GROWSDOWN, "growsdown" }, 126 {VM_PFNMAP, "pfnmap" }, 127 {VM_DENYWRITE, "denywrite" }, 128 {VM_LOCKED, "locked" }, 129 {VM_IO, "io" }, 130 {VM_SEQ_READ, "seqread" }, 131 {VM_RAND_READ, "randread" }, 132 {VM_DONTCOPY, "dontcopy" }, 133 {VM_DONTEXPAND, "dontexpand" }, 134 {VM_ACCOUNT, "account" }, 135 {VM_NORESERVE, "noreserve" }, 136 {VM_HUGETLB, "hugetlb" }, 137 #if defined(CONFIG_X86) 138 {VM_PAT, "pat" }, 139 #elif defined(CONFIG_PPC) 140 {VM_SAO, "sao" }, 141 #elif defined(CONFIG_PARISC) || defined(CONFIG_METAG) || defined(CONFIG_IA64) 142 {VM_GROWSUP, "growsup" }, 143 #elif !defined(CONFIG_MMU) 144 {VM_MAPPED_COPY, "mappedcopy" }, 145 #else 146 {VM_ARCH_1, "arch_1" }, 147 #endif 148 {VM_DONTDUMP, "dontdump" }, 149 #ifdef CONFIG_MEM_SOFT_DIRTY 150 {VM_SOFTDIRTY, "softdirty" }, 151 #endif 152 {VM_MIXEDMAP, "mixedmap" }, 153 {VM_HUGEPAGE, "hugepage" }, 154 {VM_NOHUGEPAGE, "nohugepage" }, 155 {VM_MERGEABLE, "mergeable" }, 156 }; 157 158 void dump_vma(const struct vm_area_struct *vma) 159 { 160 pr_emerg("vma %p start %p end %p\n" 161 "next %p prev %p mm %p\n" 162 "prot %lx anon_vma %p vm_ops %p\n" 163 "pgoff %lx file %p private_data %p\n", 164 vma, (void *)vma->vm_start, (void *)vma->vm_end, vma->vm_next, 165 vma->vm_prev, vma->vm_mm, 166 (unsigned long)pgprot_val(vma->vm_page_prot), 167 vma->anon_vma, vma->vm_ops, vma->vm_pgoff, 168 vma->vm_file, vma->vm_private_data); 169 dump_flags(vma->vm_flags, vmaflags_names, ARRAY_SIZE(vmaflags_names)); 170 } 171 EXPORT_SYMBOL(dump_vma); 172 173 void dump_mm(const struct mm_struct *mm) 174 { 175 pr_emerg("mm %p mmap %p seqnum %d task_size %lu\n" 176 #ifdef CONFIG_MMU 177 "get_unmapped_area %p\n" 178 #endif 179 "mmap_base %lu mmap_legacy_base %lu highest_vm_end %lu\n" 180 "pgd %p mm_users %d mm_count %d nr_ptes %lu nr_pmds %lu map_count %d\n" 181 "hiwater_rss %lx hiwater_vm %lx total_vm %lx locked_vm %lx\n" 182 "pinned_vm %lx shared_vm %lx exec_vm %lx stack_vm %lx\n" 183 "start_code %lx end_code %lx start_data %lx end_data %lx\n" 184 "start_brk %lx brk %lx start_stack %lx\n" 185 "arg_start %lx arg_end %lx env_start %lx env_end %lx\n" 186 "binfmt %p flags %lx core_state %p\n" 187 #ifdef CONFIG_AIO 188 "ioctx_table %p\n" 189 #endif 190 #ifdef CONFIG_MEMCG 191 "owner %p " 192 #endif 193 "exe_file %p\n" 194 #ifdef CONFIG_MMU_NOTIFIER 195 "mmu_notifier_mm %p\n" 196 #endif 197 #ifdef CONFIG_NUMA_BALANCING 198 "numa_next_scan %lu numa_scan_offset %lu numa_scan_seq %d\n" 199 #endif 200 #if defined(CONFIG_NUMA_BALANCING) || defined(CONFIG_COMPACTION) 201 "tlb_flush_pending %d\n" 202 #endif 203 "%s", /* This is here to hold the comma */ 204 205 mm, mm->mmap, mm->vmacache_seqnum, mm->task_size, 206 #ifdef CONFIG_MMU 207 mm->get_unmapped_area, 208 #endif 209 mm->mmap_base, mm->mmap_legacy_base, mm->highest_vm_end, 210 mm->pgd, atomic_read(&mm->mm_users), 211 atomic_read(&mm->mm_count), 212 atomic_long_read((atomic_long_t *)&mm->nr_ptes), 213 mm_nr_pmds((struct mm_struct *)mm), 214 mm->map_count, 215 mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm, 216 mm->pinned_vm, mm->shared_vm, mm->exec_vm, mm->stack_vm, 217 mm->start_code, mm->end_code, mm->start_data, mm->end_data, 218 mm->start_brk, mm->brk, mm->start_stack, 219 mm->arg_start, mm->arg_end, mm->env_start, mm->env_end, 220 mm->binfmt, mm->flags, mm->core_state, 221 #ifdef CONFIG_AIO 222 mm->ioctx_table, 223 #endif 224 #ifdef CONFIG_MEMCG 225 mm->owner, 226 #endif 227 mm->exe_file, 228 #ifdef CONFIG_MMU_NOTIFIER 229 mm->mmu_notifier_mm, 230 #endif 231 #ifdef CONFIG_NUMA_BALANCING 232 mm->numa_next_scan, mm->numa_scan_offset, mm->numa_scan_seq, 233 #endif 234 #if defined(CONFIG_NUMA_BALANCING) || defined(CONFIG_COMPACTION) 235 mm->tlb_flush_pending, 236 #endif 237 "" /* This is here to not have a comma! */ 238 ); 239 240 dump_flags(mm->def_flags, vmaflags_names, 241 ARRAY_SIZE(vmaflags_names)); 242 } 243 244 #endif /* CONFIG_DEBUG_VM */ 245