1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright 2021 Collabora ltd. */ 3 4 #include <linux/err.h> 5 #include <linux/device.h> 6 #include <linux/devcoredump.h> 7 #include <linux/moduleparam.h> 8 #include <linux/iosys-map.h> 9 #include <drm/panfrost_drm.h> 10 #include <drm/drm_device.h> 11 12 #include "panfrost_job.h" 13 #include "panfrost_gem.h" 14 #include "panfrost_regs.h" 15 #include "panfrost_dump.h" 16 #include "panfrost_device.h" 17 18 static bool panfrost_dump_core = true; 19 module_param_named(dump_core, panfrost_dump_core, bool, 0600); 20 21 struct panfrost_dump_iterator { 22 void *start; 23 struct panfrost_dump_object_header *hdr; 24 void *data; 25 }; 26 27 static const unsigned short panfrost_dump_registers[] = { 28 SHADER_READY_LO, 29 SHADER_READY_HI, 30 TILER_READY_LO, 31 TILER_READY_HI, 32 L2_READY_LO, 33 L2_READY_HI, 34 JOB_INT_MASK, 35 JOB_INT_STAT, 36 JS_HEAD_LO(0), 37 JS_HEAD_HI(0), 38 JS_TAIL_LO(0), 39 JS_TAIL_HI(0), 40 JS_AFFINITY_LO(0), 41 JS_AFFINITY_HI(0), 42 JS_CONFIG(0), 43 JS_STATUS(0), 44 JS_HEAD_NEXT_LO(0), 45 JS_HEAD_NEXT_HI(0), 46 JS_AFFINITY_NEXT_LO(0), 47 JS_AFFINITY_NEXT_HI(0), 48 JS_CONFIG_NEXT(0), 49 MMU_INT_MASK, 50 MMU_INT_STAT, 51 AS_TRANSTAB_LO(0), 52 AS_TRANSTAB_HI(0), 53 AS_MEMATTR_LO(0), 54 AS_MEMATTR_HI(0), 55 AS_FAULTSTATUS(0), 56 AS_FAULTADDRESS_LO(0), 57 AS_FAULTADDRESS_HI(0), 58 AS_STATUS(0), 59 }; 60 61 static void panfrost_core_dump_header(struct panfrost_dump_iterator *iter, 62 u32 type, void *data_end) 63 { 64 struct panfrost_dump_object_header *hdr = iter->hdr; 65 66 hdr->magic = cpu_to_le32(PANFROSTDUMP_MAGIC); 67 hdr->type = cpu_to_le32(type); 68 hdr->file_offset = cpu_to_le32(iter->data - iter->start); 69 hdr->file_size = cpu_to_le32(data_end - iter->data); 70 71 iter->hdr++; 72 iter->data += le32_to_cpu(hdr->file_size); 73 } 74 75 static void 76 panfrost_core_dump_registers(struct panfrost_dump_iterator *iter, 77 struct panfrost_device *pfdev, 78 u32 as_nr, int slot) 79 { 80 struct panfrost_dump_registers *dumpreg = iter->data; 81 unsigned int i; 82 83 for (i = 0; i < ARRAY_SIZE(panfrost_dump_registers); i++, dumpreg++) { 84 unsigned int js_as_offset = 0; 85 unsigned int reg; 86 87 if (panfrost_dump_registers[i] >= JS_BASE && 88 panfrost_dump_registers[i] <= JS_BASE + JS_SLOT_STRIDE) 89 js_as_offset = slot * JS_SLOT_STRIDE; 90 else if (panfrost_dump_registers[i] >= MMU_BASE && 91 panfrost_dump_registers[i] <= MMU_BASE + MMU_AS_STRIDE) 92 js_as_offset = (as_nr << MMU_AS_SHIFT); 93 94 reg = panfrost_dump_registers[i] + js_as_offset; 95 96 dumpreg->reg = cpu_to_le32(reg); 97 dumpreg->value = cpu_to_le32(gpu_read(pfdev, reg)); 98 } 99 100 panfrost_core_dump_header(iter, PANFROSTDUMP_BUF_REG, dumpreg); 101 } 102 103 void panfrost_core_dump(struct panfrost_job *job) 104 { 105 struct panfrost_device *pfdev = job->pfdev; 106 struct panfrost_dump_iterator iter; 107 struct drm_gem_object *dbo; 108 unsigned int n_obj, n_bomap_pages; 109 __le64 *bomap, *bomap_start; 110 size_t file_size; 111 u32 as_nr; 112 int slot; 113 int ret, i; 114 115 as_nr = job->mmu->as; 116 slot = panfrost_job_get_slot(job); 117 118 /* Only catch the first event, or when manually re-armed */ 119 if (!panfrost_dump_core) 120 return; 121 panfrost_dump_core = false; 122 123 /* At least, we dump registers and end marker */ 124 n_obj = 2; 125 n_bomap_pages = 0; 126 file_size = ARRAY_SIZE(panfrost_dump_registers) * 127 sizeof(struct panfrost_dump_registers); 128 129 /* Add in the active buffer objects */ 130 for (i = 0; i < job->bo_count; i++) { 131 /* 132 * Even though the CPU could be configured to use 16K or 64K pages, this 133 * is a very unusual situation for most kernel setups on SoCs that have 134 * a Panfrost device. Also many places across the driver make the somewhat 135 * arbitrary assumption that Panfrost's MMU page size is the same as the CPU's, 136 * so let's have a sanity check to ensure that's always the case 137 */ 138 dbo = job->bos[i]; 139 WARN_ON(!IS_ALIGNED(dbo->size, PAGE_SIZE)); 140 141 file_size += dbo->size; 142 n_bomap_pages += dbo->size >> PAGE_SHIFT; 143 n_obj++; 144 } 145 146 /* If we have any buffer objects, add a bomap object */ 147 if (n_bomap_pages) { 148 file_size += n_bomap_pages * sizeof(*bomap); 149 n_obj++; 150 } 151 152 /* Add the size of the headers */ 153 file_size += sizeof(*iter.hdr) * n_obj; 154 155 /* 156 * Allocate the file in vmalloc memory, it's likely to be big. 157 * The reason behind these GFP flags is that we don't want to trigger the 158 * OOM killer in the event that not enough memory could be found for our 159 * dump file. We also don't want the allocator to do any error reporting, 160 * as the right behaviour is failing gracefully if a big enough buffer 161 * could not be allocated. 162 */ 163 iter.start = __vmalloc(file_size, GFP_KERNEL | __GFP_NOWARN | 164 __GFP_NORETRY); 165 if (!iter.start) { 166 dev_warn(pfdev->dev, "failed to allocate devcoredump file\n"); 167 return; 168 } 169 170 /* Point the data member after the headers */ 171 iter.hdr = iter.start; 172 iter.data = &iter.hdr[n_obj]; 173 174 memset(iter.hdr, 0, iter.data - iter.start); 175 176 /* 177 * For now, we write the job identifier in the register dump header, 178 * so that we can decode the entire dump later with pandecode 179 */ 180 iter.hdr->reghdr.jc = cpu_to_le64(job->jc); 181 iter.hdr->reghdr.major = cpu_to_le32(PANFROSTDUMP_MAJOR); 182 iter.hdr->reghdr.minor = cpu_to_le32(PANFROSTDUMP_MINOR); 183 iter.hdr->reghdr.gpu_id = cpu_to_le32(pfdev->features.id); 184 iter.hdr->reghdr.nbos = cpu_to_le64(job->bo_count); 185 186 panfrost_core_dump_registers(&iter, pfdev, as_nr, slot); 187 188 /* Reserve space for the bomap */ 189 if (job->bo_count) { 190 bomap_start = bomap = iter.data; 191 memset(bomap, 0, sizeof(*bomap) * n_bomap_pages); 192 panfrost_core_dump_header(&iter, PANFROSTDUMP_BUF_BOMAP, 193 bomap + n_bomap_pages); 194 } 195 196 for (i = 0; i < job->bo_count; i++) { 197 struct iosys_map map; 198 struct panfrost_gem_mapping *mapping; 199 struct panfrost_gem_object *bo; 200 struct sg_page_iter page_iter; 201 void *vaddr; 202 203 bo = to_panfrost_bo(job->bos[i]); 204 mapping = job->mappings[i]; 205 206 if (!bo->base.sgt) { 207 dev_err(pfdev->dev, "Panfrost Dump: BO has no sgt, cannot dump\n"); 208 iter.hdr->bomap.valid = 0; 209 goto dump_header; 210 } 211 212 ret = drm_gem_shmem_vmap(&bo->base, &map); 213 if (ret) { 214 dev_err(pfdev->dev, "Panfrost Dump: couldn't map Buffer Object\n"); 215 iter.hdr->bomap.valid = 0; 216 goto dump_header; 217 } 218 219 WARN_ON(!mapping->active); 220 221 iter.hdr->bomap.data[0] = cpu_to_le32((bomap - bomap_start)); 222 223 for_each_sgtable_page(bo->base.sgt, &page_iter, 0) { 224 struct page *page = sg_page_iter_page(&page_iter); 225 226 if (!IS_ERR(page)) { 227 *bomap++ = cpu_to_le64(page_to_phys(page)); 228 } else { 229 dev_err(pfdev->dev, "Panfrost Dump: wrong page\n"); 230 *bomap++ = ~cpu_to_le64(0); 231 } 232 } 233 234 iter.hdr->bomap.iova = cpu_to_le64(mapping->mmnode.start << PAGE_SHIFT); 235 236 vaddr = map.vaddr; 237 memcpy(iter.data, vaddr, bo->base.base.size); 238 239 drm_gem_shmem_vunmap(&bo->base, &map); 240 241 iter.hdr->bomap.valid = cpu_to_le32(1); 242 243 dump_header: panfrost_core_dump_header(&iter, PANFROSTDUMP_BUF_BO, iter.data + 244 bo->base.base.size); 245 } 246 panfrost_core_dump_header(&iter, PANFROSTDUMP_BUF_TRAILER, iter.data); 247 248 dev_coredumpv(pfdev->dev, iter.start, iter.data - iter.start, GFP_KERNEL); 249 } 250