xref: /linux/arch/powerpc/kernel/hw_breakpoint_constraints.c (revision 0b8061c340b643e01da431dd60c75a41bb1d31ec)
1 // SPDX-License-Identifier: GPL-2.0+
2 #include <linux/kernel.h>
3 #include <linux/uaccess.h>
4 #include <linux/sched.h>
5 #include <asm/hw_breakpoint.h>
6 #include <asm/sstep.h>
7 #include <asm/cache.h>
8 
9 static bool dar_in_user_range(unsigned long dar, struct arch_hw_breakpoint *info)
10 {
11 	return ((info->address <= dar) && (dar - info->address < info->len));
12 }
13 
14 static bool ea_user_range_overlaps(unsigned long ea, int size,
15 				   struct arch_hw_breakpoint *info)
16 {
17 	return ((ea < info->address + info->len) &&
18 		(ea + size > info->address));
19 }
20 
21 static bool dar_in_hw_range(unsigned long dar, struct arch_hw_breakpoint *info)
22 {
23 	unsigned long hw_start_addr, hw_end_addr;
24 
25 	hw_start_addr = ALIGN_DOWN(info->address, HW_BREAKPOINT_SIZE);
26 	hw_end_addr = ALIGN(info->address + info->len, HW_BREAKPOINT_SIZE);
27 
28 	return ((hw_start_addr <= dar) && (hw_end_addr > dar));
29 }
30 
31 static bool ea_hw_range_overlaps(unsigned long ea, int size,
32 				 struct arch_hw_breakpoint *info)
33 {
34 	unsigned long hw_start_addr, hw_end_addr;
35 	unsigned long align_size = HW_BREAKPOINT_SIZE;
36 
37 	/*
38 	 * On p10 predecessors, quadword is handle differently then
39 	 * other instructions.
40 	 */
41 	if (!cpu_has_feature(CPU_FTR_ARCH_31) && size == 16)
42 		align_size = HW_BREAKPOINT_SIZE_QUADWORD;
43 
44 	hw_start_addr = ALIGN_DOWN(info->address, align_size);
45 	hw_end_addr = ALIGN(info->address + info->len, align_size);
46 
47 	return ((ea < hw_end_addr) && (ea + size > hw_start_addr));
48 }
49 
50 /*
51  * If hw has multiple DAWR registers, we also need to check all
52  * dawrx constraint bits to confirm this is _really_ a valid event.
53  * If type is UNKNOWN, but privilege level matches, consider it as
54  * a positive match.
55  */
56 static bool check_dawrx_constraints(struct pt_regs *regs, int type,
57 				    struct arch_hw_breakpoint *info)
58 {
59 	if (OP_IS_LOAD(type) && !(info->type & HW_BRK_TYPE_READ))
60 		return false;
61 
62 	/*
63 	 * The Cache Management instructions other than dcbz never
64 	 * cause a match. i.e. if type is CACHEOP, the instruction
65 	 * is dcbz, and dcbz is treated as Store.
66 	 */
67 	if ((OP_IS_STORE(type) || type == CACHEOP) && !(info->type & HW_BRK_TYPE_WRITE))
68 		return false;
69 
70 	if (is_kernel_addr(regs->nip) && !(info->type & HW_BRK_TYPE_KERNEL))
71 		return false;
72 
73 	if (user_mode(regs) && !(info->type & HW_BRK_TYPE_USER))
74 		return false;
75 
76 	return true;
77 }
78 
79 /*
80  * Return true if the event is valid wrt dawr configuration,
81  * including extraneous exception. Otherwise return false.
82  */
83 bool wp_check_constraints(struct pt_regs *regs, struct ppc_inst instr,
84 			  unsigned long ea, int type, int size,
85 			  struct arch_hw_breakpoint *info)
86 {
87 	bool in_user_range = dar_in_user_range(regs->dar, info);
88 	bool dawrx_constraints;
89 
90 	/*
91 	 * 8xx supports only one breakpoint and thus we can
92 	 * unconditionally return true.
93 	 */
94 	if (IS_ENABLED(CONFIG_PPC_8xx)) {
95 		if (!in_user_range)
96 			info->type |= HW_BRK_TYPE_EXTRANEOUS_IRQ;
97 		return true;
98 	}
99 
100 	if (unlikely(ppc_inst_equal(instr, ppc_inst(0)))) {
101 		if (cpu_has_feature(CPU_FTR_ARCH_31) &&
102 		    !dar_in_hw_range(regs->dar, info))
103 			return false;
104 
105 		return true;
106 	}
107 
108 	dawrx_constraints = check_dawrx_constraints(regs, type, info);
109 
110 	if (type == UNKNOWN) {
111 		if (cpu_has_feature(CPU_FTR_ARCH_31) &&
112 		    !dar_in_hw_range(regs->dar, info))
113 			return false;
114 
115 		return dawrx_constraints;
116 	}
117 
118 	if (ea_user_range_overlaps(ea, size, info))
119 		return dawrx_constraints;
120 
121 	if (ea_hw_range_overlaps(ea, size, info)) {
122 		if (dawrx_constraints) {
123 			info->type |= HW_BRK_TYPE_EXTRANEOUS_IRQ;
124 			return true;
125 		}
126 	}
127 	return false;
128 }
129 
130 static int cache_op_size(void)
131 {
132 #ifdef __powerpc64__
133 	return ppc64_caches.l1d.block_size;
134 #else
135 	return L1_CACHE_BYTES;
136 #endif
137 }
138 
139 void wp_get_instr_detail(struct pt_regs *regs, struct ppc_inst *instr,
140 			 int *type, int *size, unsigned long *ea)
141 {
142 	struct instruction_op op;
143 
144 	if (__get_user_instr_inatomic(*instr, (void __user *)regs->nip))
145 		return;
146 
147 	analyse_instr(&op, regs, *instr);
148 	*type = GETTYPE(op.type);
149 	*ea = op.ea;
150 #ifdef __powerpc64__
151 	if (!(regs->msr & MSR_64BIT))
152 		*ea &= 0xffffffffUL;
153 #endif
154 
155 	*size = GETSIZE(op.type);
156 	if (*type == CACHEOP) {
157 		*size = cache_op_size();
158 		*ea &= ~(*size - 1);
159 	} else if (*type == LOAD_VMX || *type == STORE_VMX) {
160 		*ea &= ~(*size - 1);
161 	}
162 }
163