1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * MIPS cacheinfo support 4 */ 5 #include <linux/cacheinfo.h> 6 7 /* Populates leaf and increments to next leaf */ 8 #define populate_cache(cache, leaf, c_level, c_type) \ 9 do { \ 10 leaf->type = c_type; \ 11 leaf->level = c_level; \ 12 leaf->coherency_line_size = c->cache.linesz; \ 13 leaf->number_of_sets = c->cache.sets; \ 14 leaf->ways_of_associativity = c->cache.ways; \ 15 leaf->size = c->cache.linesz * c->cache.sets * \ 16 c->cache.ways; \ 17 leaf++; \ 18 } while (0) 19 20 int init_cache_level(unsigned int cpu) 21 { 22 struct cpuinfo_mips *c = ¤t_cpu_data; 23 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); 24 int levels = 0, leaves = 0; 25 26 /* 27 * If Dcache is not set, we assume the cache structures 28 * are not properly initialized. 29 */ 30 if (c->dcache.waysize) 31 levels += 1; 32 else 33 return -ENOENT; 34 35 36 leaves += (c->icache.waysize) ? 2 : 1; 37 38 if (c->vcache.waysize) { 39 levels++; 40 leaves++; 41 } 42 43 if (c->scache.waysize) { 44 levels++; 45 leaves++; 46 } 47 48 if (c->tcache.waysize) { 49 levels++; 50 leaves++; 51 } 52 53 this_cpu_ci->num_levels = levels; 54 this_cpu_ci->num_leaves = leaves; 55 return 0; 56 } 57 58 static void fill_cpumask_siblings(int cpu, cpumask_t *cpu_map) 59 { 60 int cpu1; 61 62 for_each_possible_cpu(cpu1) 63 if (cpus_are_siblings(cpu, cpu1)) 64 cpumask_set_cpu(cpu1, cpu_map); 65 } 66 67 static void fill_cpumask_cluster(int cpu, cpumask_t *cpu_map) 68 { 69 int cpu1; 70 int cluster = cpu_cluster(&cpu_data[cpu]); 71 72 for_each_possible_cpu(cpu1) 73 if (cpu_cluster(&cpu_data[cpu1]) == cluster) 74 cpumask_set_cpu(cpu1, cpu_map); 75 } 76 77 int populate_cache_leaves(unsigned int cpu) 78 { 79 struct cpuinfo_mips *c = ¤t_cpu_data; 80 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); 81 struct cacheinfo *this_leaf = this_cpu_ci->info_list; 82 int level = 1; 83 84 if (c->icache.waysize) { 85 /* I/D caches are per core */ 86 fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map); 87 populate_cache(dcache, this_leaf, level, CACHE_TYPE_DATA); 88 fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map); 89 populate_cache(icache, this_leaf, level, CACHE_TYPE_INST); 90 level++; 91 } else { 92 populate_cache(dcache, this_leaf, level, CACHE_TYPE_UNIFIED); 93 level++; 94 } 95 96 if (c->vcache.waysize) { 97 /* Vcache is per core as well */ 98 fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map); 99 populate_cache(vcache, this_leaf, level, CACHE_TYPE_UNIFIED); 100 level++; 101 } 102 103 if (c->scache.waysize) { 104 /* Scache is per cluster */ 105 fill_cpumask_cluster(cpu, &this_leaf->shared_cpu_map); 106 populate_cache(scache, this_leaf, level, CACHE_TYPE_UNIFIED); 107 level++; 108 } 109 110 if (c->tcache.waysize) 111 populate_cache(tcache, this_leaf, level, CACHE_TYPE_UNIFIED); 112 113 this_cpu_ci->cpu_map_populated = true; 114 115 return 0; 116 } 117