1 /* SPDX-License-Identifier: GPL-2.0 */
2 /*
3 * definition for kernel virtual machines on s390
4 *
5 * Copyright IBM Corp. 2008, 2018
6 *
7 * Author(s): Carsten Otte <cotte@de.ibm.com>
8 */
9
10
11 #ifndef ASM_KVM_HOST_H
12 #define ASM_KVM_HOST_H
13
14 #include <linux/types.h>
15 #include <linux/hrtimer.h>
16 #include <linux/interrupt.h>
17 #include <linux/kvm_types.h>
18 #include <linux/kvm.h>
19 #include <linux/seqlock.h>
20 #include <linux/module.h>
21 #include <linux/pci.h>
22 #include <linux/mmu_notifier.h>
23 #include <asm/debug.h>
24 #include <asm/cpu.h>
25 #include <asm/fpu.h>
26 #include <asm/isc.h>
27 #include <asm/guarded_storage.h>
28
29 #define KVM_S390_BSCA_CPU_SLOTS 64
30 #define KVM_S390_ESCA_CPU_SLOTS 248
31 #define KVM_MAX_VCPUS 255
32
33 #define KVM_INTERNAL_MEM_SLOTS 1
34
35 /*
36 * These seem to be used for allocating ->chip in the routing table, which we
37 * don't use. 1 is as small as we can get to reduce the needed memory. If we
38 * need to look at ->chip later on, we'll need to revisit this.
39 */
40 #define KVM_NR_IRQCHIPS 1
41 #define KVM_IRQCHIP_NUM_PINS 1
42 #define KVM_HALT_POLL_NS_DEFAULT 50000
43
44 /* s390-specific vcpu->requests bit members */
45 #define KVM_REQ_ENABLE_IBS KVM_ARCH_REQ(0)
46 #define KVM_REQ_DISABLE_IBS KVM_ARCH_REQ(1)
47 #define KVM_REQ_ICPT_OPEREXC KVM_ARCH_REQ(2)
48 #define KVM_REQ_START_MIGRATION KVM_ARCH_REQ(3)
49 #define KVM_REQ_STOP_MIGRATION KVM_ARCH_REQ(4)
50 #define KVM_REQ_VSIE_RESTART KVM_ARCH_REQ(5)
51 #define KVM_REQ_REFRESH_GUEST_PREFIX \
52 KVM_ARCH_REQ_FLAGS(6, KVM_REQUEST_WAIT | KVM_REQUEST_NO_WAKEUP)
53
54 #define SIGP_CTRL_C 0x80
55 #define SIGP_CTRL_SCN_MASK 0x3f
56
57 union bsca_sigp_ctrl {
58 __u8 value;
59 struct {
60 __u8 c : 1;
61 __u8 r : 1;
62 __u8 scn : 6;
63 };
64 };
65
66 union esca_sigp_ctrl {
67 __u16 value;
68 struct {
69 __u8 c : 1;
70 __u8 reserved: 7;
71 __u8 scn;
72 };
73 };
74
75 struct esca_entry {
76 union esca_sigp_ctrl sigp_ctrl;
77 __u16 reserved1[3];
78 __u64 sda;
79 __u64 reserved2[6];
80 };
81
82 struct bsca_entry {
83 __u8 reserved0;
84 union bsca_sigp_ctrl sigp_ctrl;
85 __u16 reserved[3];
86 __u64 sda;
87 __u64 reserved2[2];
88 };
89
90 union ipte_control {
91 unsigned long val;
92 struct {
93 unsigned long k : 1;
94 unsigned long kh : 31;
95 unsigned long kg : 32;
96 };
97 };
98
99 /*
100 * Utility is defined as two bytes but having it four bytes wide
101 * generates more efficient code. Since the following bytes are
102 * reserved this makes no functional difference.
103 */
104 union sca_utility {
105 __u32 val;
106 struct {
107 __u32 mtcr : 1;
108 __u32 : 31;
109 };
110 };
111
112 struct bsca_block {
113 union ipte_control ipte_control;
114 __u64 reserved[5];
115 __u64 mcn;
116 union sca_utility utility;
117 __u8 reserved2[4];
118 struct bsca_entry cpu[KVM_S390_BSCA_CPU_SLOTS];
119 };
120
121 struct esca_block {
122 union ipte_control ipte_control;
123 __u64 reserved1[6];
124 union sca_utility utility;
125 __u8 reserved2[4];
126 __u64 mcn[4];
127 __u64 reserved3[20];
128 struct esca_entry cpu[KVM_S390_ESCA_CPU_SLOTS];
129 };
130
131 /*
132 * This struct is used to store some machine check info from lowcore
133 * for machine checks that happen while the guest is running.
134 * This info in host's lowcore might be overwritten by a second machine
135 * check from host when host is in the machine check's high-level handling.
136 * The size is 24 bytes.
137 */
138 struct mcck_volatile_info {
139 __u64 mcic;
140 __u64 failing_storage_address;
141 __u32 ext_damage_code;
142 __u32 reserved;
143 };
144
145 #define CR0_INITIAL_MASK (CR0_UNUSED_56 | CR0_INTERRUPT_KEY_SUBMASK | \
146 CR0_MEASUREMENT_ALERT_SUBMASK)
147 #define CR14_INITIAL_MASK (CR14_UNUSED_32 | CR14_UNUSED_33 | \
148 CR14_EXTERNAL_DAMAGE_SUBMASK)
149
150 #define SIDAD_SIZE_MASK 0xff
151 #define sida_addr(sie_block) phys_to_virt((sie_block)->sidad & PAGE_MASK)
152 #define sida_size(sie_block) \
153 ((((sie_block)->sidad & SIDAD_SIZE_MASK) + 1) * PAGE_SIZE)
154
155 #define CPUSTAT_STOPPED 0x80000000
156 #define CPUSTAT_WAIT 0x10000000
157 #define CPUSTAT_ECALL_PEND 0x08000000
158 #define CPUSTAT_STOP_INT 0x04000000
159 #define CPUSTAT_IO_INT 0x02000000
160 #define CPUSTAT_EXT_INT 0x01000000
161 #define CPUSTAT_RUNNING 0x00800000
162 #define CPUSTAT_RETAINED 0x00400000
163 #define CPUSTAT_TIMING_SUB 0x00020000
164 #define CPUSTAT_SIE_SUB 0x00010000
165 #define CPUSTAT_RRF 0x00008000
166 #define CPUSTAT_SLSV 0x00004000
167 #define CPUSTAT_SLSR 0x00002000
168 #define CPUSTAT_ZARCH 0x00000800
169 #define CPUSTAT_MCDS 0x00000100
170 #define CPUSTAT_KSS 0x00000200
171 #define CPUSTAT_SM 0x00000080
172 #define CPUSTAT_IBS 0x00000040
173 #define CPUSTAT_GED2 0x00000010
174 #define CPUSTAT_G 0x00000008
175 #define CPUSTAT_GED 0x00000004
176 #define CPUSTAT_J 0x00000002
177 #define CPUSTAT_P 0x00000001
178
179 struct kvm_s390_sie_block {
180 atomic_t cpuflags; /* 0x0000 */
181 __u32 : 1; /* 0x0004 */
182 __u32 prefix : 18;
183 __u32 : 1;
184 __u32 ibc : 12;
185 __u8 reserved08[4]; /* 0x0008 */
186 #define PROG_IN_SIE (1<<0)
187 __u32 prog0c; /* 0x000c */
188 union {
189 __u8 reserved10[16]; /* 0x0010 */
190 struct {
191 __u64 pv_handle_cpu;
192 __u64 pv_handle_config;
193 };
194 };
195 #define PROG_BLOCK_SIE (1<<0)
196 #define PROG_REQUEST (1<<1)
197 atomic_t prog20; /* 0x0020 */
198 __u8 reserved24[4]; /* 0x0024 */
199 __u64 cputm; /* 0x0028 */
200 __u64 ckc; /* 0x0030 */
201 __u64 epoch; /* 0x0038 */
202 __u32 svcc; /* 0x0040 */
203 #define LCTL_CR0 0x8000
204 #define LCTL_CR6 0x0200
205 #define LCTL_CR9 0x0040
206 #define LCTL_CR10 0x0020
207 #define LCTL_CR11 0x0010
208 #define LCTL_CR14 0x0002
209 __u16 lctl; /* 0x0044 */
210 __s16 icpua; /* 0x0046 */
211 #define ICTL_OPEREXC 0x80000000
212 #define ICTL_PINT 0x20000000
213 #define ICTL_LPSW 0x00400000
214 #define ICTL_STCTL 0x00040000
215 #define ICTL_ISKE 0x00004000
216 #define ICTL_SSKE 0x00002000
217 #define ICTL_RRBE 0x00001000
218 #define ICTL_TPROT 0x00000200
219 __u32 ictl; /* 0x0048 */
220 #define ECA_CEI 0x80000000
221 #define ECA_IB 0x40000000
222 #define ECA_SIGPI 0x10000000
223 #define ECA_MVPGI 0x01000000
224 #define ECA_AIV 0x00200000
225 #define ECA_VX 0x00020000
226 #define ECA_PROTEXCI 0x00002000
227 #define ECA_APIE 0x00000008
228 #define ECA_SII 0x00000001
229 __u32 eca; /* 0x004c */
230 #define ICPT_INST 0x04
231 #define ICPT_PROGI 0x08
232 #define ICPT_INSTPROGI 0x0C
233 #define ICPT_EXTREQ 0x10
234 #define ICPT_EXTINT 0x14
235 #define ICPT_IOREQ 0x18
236 #define ICPT_WAIT 0x1c
237 #define ICPT_VALIDITY 0x20
238 #define ICPT_STOP 0x28
239 #define ICPT_OPEREXC 0x2C
240 #define ICPT_PARTEXEC 0x38
241 #define ICPT_IOINST 0x40
242 #define ICPT_KSS 0x5c
243 #define ICPT_MCHKREQ 0x60
244 #define ICPT_INT_ENABLE 0x64
245 #define ICPT_PV_INSTR 0x68
246 #define ICPT_PV_NOTIFY 0x6c
247 #define ICPT_PV_PREF 0x70
248 __u8 icptcode; /* 0x0050 */
249 __u8 icptstatus; /* 0x0051 */
250 __u16 ihcpu; /* 0x0052 */
251 __u8 reserved54; /* 0x0054 */
252 #define IICTL_CODE_NONE 0x00
253 #define IICTL_CODE_MCHK 0x01
254 #define IICTL_CODE_EXT 0x02
255 #define IICTL_CODE_IO 0x03
256 #define IICTL_CODE_RESTART 0x04
257 #define IICTL_CODE_SPECIFICATION 0x10
258 #define IICTL_CODE_OPERAND 0x11
259 __u8 iictl; /* 0x0055 */
260 __u16 ipa; /* 0x0056 */
261 __u32 ipb; /* 0x0058 */
262 __u32 scaoh; /* 0x005c */
263 #define FPF_BPBC 0x20
264 __u8 fpf; /* 0x0060 */
265 #define ECB_GS 0x40
266 #define ECB_TE 0x10
267 #define ECB_SPECI 0x08
268 #define ECB_SRSI 0x04
269 #define ECB_HOSTPROTINT 0x02
270 #define ECB_PTF 0x01
271 __u8 ecb; /* 0x0061 */
272 #define ECB2_CMMA 0x80
273 #define ECB2_IEP 0x20
274 #define ECB2_PFMFI 0x08
275 #define ECB2_ESCA 0x04
276 #define ECB2_ZPCI_LSI 0x02
277 __u8 ecb2; /* 0x0062 */
278 #define ECB3_AISI 0x20
279 #define ECB3_AISII 0x10
280 #define ECB3_DEA 0x08
281 #define ECB3_AES 0x04
282 #define ECB3_RI 0x01
283 __u8 ecb3; /* 0x0063 */
284 #define ESCA_SCAOL_MASK ~0x3fU
285 __u32 scaol; /* 0x0064 */
286 __u8 sdf; /* 0x0068 */
287 __u8 epdx; /* 0x0069 */
288 __u8 cpnc; /* 0x006a */
289 __u8 reserved6b; /* 0x006b */
290 __u32 todpr; /* 0x006c */
291 #define GISA_FORMAT1 0x00000001
292 __u32 gd; /* 0x0070 */
293 __u8 reserved74[12]; /* 0x0074 */
294 __u64 mso; /* 0x0080 */
295 __u64 msl; /* 0x0088 */
296 psw_t gpsw; /* 0x0090 */
297 __u64 gg14; /* 0x00a0 */
298 __u64 gg15; /* 0x00a8 */
299 __u8 reservedb0[8]; /* 0x00b0 */
300 #define HPID_KVM 0x4
301 #define HPID_VSIE 0x5
302 __u8 hpid; /* 0x00b8 */
303 __u8 reservedb9[7]; /* 0x00b9 */
304 union {
305 struct {
306 __u32 eiparams; /* 0x00c0 */
307 __u16 extcpuaddr; /* 0x00c4 */
308 __u16 eic; /* 0x00c6 */
309 };
310 __u64 mcic; /* 0x00c0 */
311 } __packed;
312 __u32 reservedc8; /* 0x00c8 */
313 union {
314 struct {
315 __u16 pgmilc; /* 0x00cc */
316 __u16 iprcc; /* 0x00ce */
317 };
318 __u32 edc; /* 0x00cc */
319 } __packed;
320 union {
321 struct {
322 __u32 dxc; /* 0x00d0 */
323 __u16 mcn; /* 0x00d4 */
324 __u8 perc; /* 0x00d6 */
325 __u8 peratmid; /* 0x00d7 */
326 };
327 __u64 faddr; /* 0x00d0 */
328 } __packed;
329 __u64 peraddr; /* 0x00d8 */
330 __u8 eai; /* 0x00e0 */
331 __u8 peraid; /* 0x00e1 */
332 __u8 oai; /* 0x00e2 */
333 __u8 armid; /* 0x00e3 */
334 __u8 reservede4[4]; /* 0x00e4 */
335 union {
336 __u64 tecmc; /* 0x00e8 */
337 struct {
338 __u16 subchannel_id; /* 0x00e8 */
339 __u16 subchannel_nr; /* 0x00ea */
340 __u32 io_int_parm; /* 0x00ec */
341 __u32 io_int_word; /* 0x00f0 */
342 };
343 } __packed;
344 __u8 reservedf4[8]; /* 0x00f4 */
345 #define CRYCB_FORMAT_MASK 0x00000003
346 #define CRYCB_FORMAT0 0x00000000
347 #define CRYCB_FORMAT1 0x00000001
348 #define CRYCB_FORMAT2 0x00000003
349 __u32 crycbd; /* 0x00fc */
350 __u64 gcr[16]; /* 0x0100 */
351 union {
352 __u64 gbea; /* 0x0180 */
353 __u64 sidad;
354 };
355 __u8 reserved188[8]; /* 0x0188 */
356 __u64 sdnxo; /* 0x0190 */
357 __u8 reserved198[8]; /* 0x0198 */
358 __u32 fac; /* 0x01a0 */
359 __u8 reserved1a4[20]; /* 0x01a4 */
360 __u64 cbrlo; /* 0x01b8 */
361 __u8 reserved1c0[8]; /* 0x01c0 */
362 #define ECD_HOSTREGMGMT 0x20000000
363 #define ECD_MEF 0x08000000
364 #define ECD_ETOKENF 0x02000000
365 #define ECD_ECC 0x00200000
366 #define ECD_HMAC 0x00004000
367 __u32 ecd; /* 0x01c8 */
368 __u8 reserved1cc[18]; /* 0x01cc */
369 __u64 pp; /* 0x01de */
370 __u8 reserved1e6[2]; /* 0x01e6 */
371 __u64 itdba; /* 0x01e8 */
372 __u64 riccbd; /* 0x01f0 */
373 __u64 gvrd; /* 0x01f8 */
374 } __packed __aligned(512);
375
376 struct kvm_s390_itdb {
377 __u8 data[256];
378 };
379
380 struct sie_page {
381 struct kvm_s390_sie_block sie_block;
382 struct mcck_volatile_info mcck_info; /* 0x0200 */
383 __u8 reserved218[360]; /* 0x0218 */
384 __u64 pv_grregs[16]; /* 0x0380 */
385 __u8 reserved400[512]; /* 0x0400 */
386 struct kvm_s390_itdb itdb; /* 0x0600 */
387 __u8 reserved700[2304]; /* 0x0700 */
388 };
389
390 struct kvm_vcpu_stat {
391 struct kvm_vcpu_stat_generic generic;
392 u64 exit_userspace;
393 u64 exit_null;
394 u64 exit_external_request;
395 u64 exit_io_request;
396 u64 exit_external_interrupt;
397 u64 exit_stop_request;
398 u64 exit_validity;
399 u64 exit_instruction;
400 u64 exit_pei;
401 u64 halt_no_poll_steal;
402 u64 instruction_lctl;
403 u64 instruction_lctlg;
404 u64 instruction_stctl;
405 u64 instruction_stctg;
406 u64 exit_program_interruption;
407 u64 exit_instr_and_program;
408 u64 exit_operation_exception;
409 u64 deliver_ckc;
410 u64 deliver_cputm;
411 u64 deliver_external_call;
412 u64 deliver_emergency_signal;
413 u64 deliver_service_signal;
414 u64 deliver_virtio;
415 u64 deliver_stop_signal;
416 u64 deliver_prefix_signal;
417 u64 deliver_restart_signal;
418 u64 deliver_program;
419 u64 deliver_io;
420 u64 deliver_machine_check;
421 u64 exit_wait_state;
422 u64 inject_ckc;
423 u64 inject_cputm;
424 u64 inject_external_call;
425 u64 inject_emergency_signal;
426 u64 inject_mchk;
427 u64 inject_pfault_init;
428 u64 inject_program;
429 u64 inject_restart;
430 u64 inject_set_prefix;
431 u64 inject_stop_signal;
432 u64 instruction_epsw;
433 u64 instruction_gs;
434 u64 instruction_io_other;
435 u64 instruction_lpsw;
436 u64 instruction_lpswe;
437 u64 instruction_lpswey;
438 u64 instruction_pfmf;
439 u64 instruction_ptff;
440 u64 instruction_sck;
441 u64 instruction_sckpf;
442 u64 instruction_stidp;
443 u64 instruction_spx;
444 u64 instruction_stpx;
445 u64 instruction_stap;
446 u64 instruction_iske;
447 u64 instruction_ri;
448 u64 instruction_rrbe;
449 u64 instruction_sske;
450 u64 instruction_ipte_interlock;
451 u64 instruction_stsi;
452 u64 instruction_stfl;
453 u64 instruction_tb;
454 u64 instruction_tpi;
455 u64 instruction_tprot;
456 u64 instruction_tsch;
457 u64 instruction_sie;
458 u64 instruction_essa;
459 u64 instruction_sthyi;
460 u64 instruction_sigp_sense;
461 u64 instruction_sigp_sense_running;
462 u64 instruction_sigp_external_call;
463 u64 instruction_sigp_emergency;
464 u64 instruction_sigp_cond_emergency;
465 u64 instruction_sigp_start;
466 u64 instruction_sigp_stop;
467 u64 instruction_sigp_stop_store_status;
468 u64 instruction_sigp_store_status;
469 u64 instruction_sigp_store_adtl_status;
470 u64 instruction_sigp_arch;
471 u64 instruction_sigp_prefix;
472 u64 instruction_sigp_restart;
473 u64 instruction_sigp_init_cpu_reset;
474 u64 instruction_sigp_cpu_reset;
475 u64 instruction_sigp_unknown;
476 u64 instruction_diagnose_10;
477 u64 instruction_diagnose_44;
478 u64 instruction_diagnose_9c;
479 u64 diag_9c_ignored;
480 u64 diag_9c_forward;
481 u64 instruction_diagnose_258;
482 u64 instruction_diagnose_308;
483 u64 instruction_diagnose_500;
484 u64 instruction_diagnose_other;
485 u64 pfault_sync;
486 };
487
488 #define PGM_OPERATION 0x01
489 #define PGM_PRIVILEGED_OP 0x02
490 #define PGM_EXECUTE 0x03
491 #define PGM_PROTECTION 0x04
492 #define PGM_ADDRESSING 0x05
493 #define PGM_SPECIFICATION 0x06
494 #define PGM_DATA 0x07
495 #define PGM_FIXED_POINT_OVERFLOW 0x08
496 #define PGM_FIXED_POINT_DIVIDE 0x09
497 #define PGM_DECIMAL_OVERFLOW 0x0a
498 #define PGM_DECIMAL_DIVIDE 0x0b
499 #define PGM_HFP_EXPONENT_OVERFLOW 0x0c
500 #define PGM_HFP_EXPONENT_UNDERFLOW 0x0d
501 #define PGM_HFP_SIGNIFICANCE 0x0e
502 #define PGM_HFP_DIVIDE 0x0f
503 #define PGM_SEGMENT_TRANSLATION 0x10
504 #define PGM_PAGE_TRANSLATION 0x11
505 #define PGM_TRANSLATION_SPEC 0x12
506 #define PGM_SPECIAL_OPERATION 0x13
507 #define PGM_OPERAND 0x15
508 #define PGM_TRACE_TABEL 0x16
509 #define PGM_VECTOR_PROCESSING 0x1b
510 #define PGM_SPACE_SWITCH 0x1c
511 #define PGM_HFP_SQUARE_ROOT 0x1d
512 #define PGM_PC_TRANSLATION_SPEC 0x1f
513 #define PGM_AFX_TRANSLATION 0x20
514 #define PGM_ASX_TRANSLATION 0x21
515 #define PGM_LX_TRANSLATION 0x22
516 #define PGM_EX_TRANSLATION 0x23
517 #define PGM_PRIMARY_AUTHORITY 0x24
518 #define PGM_SECONDARY_AUTHORITY 0x25
519 #define PGM_LFX_TRANSLATION 0x26
520 #define PGM_LSX_TRANSLATION 0x27
521 #define PGM_ALET_SPECIFICATION 0x28
522 #define PGM_ALEN_TRANSLATION 0x29
523 #define PGM_ALE_SEQUENCE 0x2a
524 #define PGM_ASTE_VALIDITY 0x2b
525 #define PGM_ASTE_SEQUENCE 0x2c
526 #define PGM_EXTENDED_AUTHORITY 0x2d
527 #define PGM_LSTE_SEQUENCE 0x2e
528 #define PGM_ASTE_INSTANCE 0x2f
529 #define PGM_STACK_FULL 0x30
530 #define PGM_STACK_EMPTY 0x31
531 #define PGM_STACK_SPECIFICATION 0x32
532 #define PGM_STACK_TYPE 0x33
533 #define PGM_STACK_OPERATION 0x34
534 #define PGM_ASCE_TYPE 0x38
535 #define PGM_REGION_FIRST_TRANS 0x39
536 #define PGM_REGION_SECOND_TRANS 0x3a
537 #define PGM_REGION_THIRD_TRANS 0x3b
538 #define PGM_SECURE_STORAGE_ACCESS 0x3d
539 #define PGM_NON_SECURE_STORAGE_ACCESS 0x3e
540 #define PGM_SECURE_STORAGE_VIOLATION 0x3f
541 #define PGM_MONITOR 0x40
542 #define PGM_PER 0x80
543 #define PGM_CRYPTO_OPERATION 0x119
544
545 /* irq types in ascend order of priorities */
546 enum irq_types {
547 IRQ_PEND_SET_PREFIX = 0,
548 IRQ_PEND_RESTART,
549 IRQ_PEND_SIGP_STOP,
550 IRQ_PEND_IO_ISC_7,
551 IRQ_PEND_IO_ISC_6,
552 IRQ_PEND_IO_ISC_5,
553 IRQ_PEND_IO_ISC_4,
554 IRQ_PEND_IO_ISC_3,
555 IRQ_PEND_IO_ISC_2,
556 IRQ_PEND_IO_ISC_1,
557 IRQ_PEND_IO_ISC_0,
558 IRQ_PEND_VIRTIO,
559 IRQ_PEND_PFAULT_DONE,
560 IRQ_PEND_PFAULT_INIT,
561 IRQ_PEND_EXT_HOST,
562 IRQ_PEND_EXT_SERVICE,
563 IRQ_PEND_EXT_SERVICE_EV,
564 IRQ_PEND_EXT_TIMING,
565 IRQ_PEND_EXT_CPU_TIMER,
566 IRQ_PEND_EXT_CLOCK_COMP,
567 IRQ_PEND_EXT_EXTERNAL,
568 IRQ_PEND_EXT_EMERGENCY,
569 IRQ_PEND_EXT_MALFUNC,
570 IRQ_PEND_EXT_IRQ_KEY,
571 IRQ_PEND_MCHK_REP,
572 IRQ_PEND_PROG,
573 IRQ_PEND_SVC,
574 IRQ_PEND_MCHK_EX,
575 IRQ_PEND_COUNT
576 };
577
578 /* We have 2M for virtio device descriptor pages. Smallest amount of
579 * memory per page is 24 bytes (1 queue), so (2048*1024) / 24 = 87381
580 */
581 #define KVM_S390_MAX_VIRTIO_IRQS 87381
582
583 /*
584 * Repressible (non-floating) machine check interrupts
585 * subclass bits in MCIC
586 */
587 #define MCHK_EXTD_BIT 58
588 #define MCHK_DEGR_BIT 56
589 #define MCHK_WARN_BIT 55
590 #define MCHK_REP_MASK ((1UL << MCHK_DEGR_BIT) | \
591 (1UL << MCHK_EXTD_BIT) | \
592 (1UL << MCHK_WARN_BIT))
593
594 /* Exigent machine check interrupts subclass bits in MCIC */
595 #define MCHK_SD_BIT 63
596 #define MCHK_PD_BIT 62
597 #define MCHK_EX_MASK ((1UL << MCHK_SD_BIT) | (1UL << MCHK_PD_BIT))
598
599 #define IRQ_PEND_EXT_MASK ((1UL << IRQ_PEND_EXT_IRQ_KEY) | \
600 (1UL << IRQ_PEND_EXT_CLOCK_COMP) | \
601 (1UL << IRQ_PEND_EXT_CPU_TIMER) | \
602 (1UL << IRQ_PEND_EXT_MALFUNC) | \
603 (1UL << IRQ_PEND_EXT_EMERGENCY) | \
604 (1UL << IRQ_PEND_EXT_EXTERNAL) | \
605 (1UL << IRQ_PEND_EXT_TIMING) | \
606 (1UL << IRQ_PEND_EXT_HOST) | \
607 (1UL << IRQ_PEND_EXT_SERVICE) | \
608 (1UL << IRQ_PEND_EXT_SERVICE_EV) | \
609 (1UL << IRQ_PEND_VIRTIO) | \
610 (1UL << IRQ_PEND_PFAULT_INIT) | \
611 (1UL << IRQ_PEND_PFAULT_DONE))
612
613 #define IRQ_PEND_IO_MASK ((1UL << IRQ_PEND_IO_ISC_0) | \
614 (1UL << IRQ_PEND_IO_ISC_1) | \
615 (1UL << IRQ_PEND_IO_ISC_2) | \
616 (1UL << IRQ_PEND_IO_ISC_3) | \
617 (1UL << IRQ_PEND_IO_ISC_4) | \
618 (1UL << IRQ_PEND_IO_ISC_5) | \
619 (1UL << IRQ_PEND_IO_ISC_6) | \
620 (1UL << IRQ_PEND_IO_ISC_7))
621
622 #define IRQ_PEND_MCHK_MASK ((1UL << IRQ_PEND_MCHK_REP) | \
623 (1UL << IRQ_PEND_MCHK_EX))
624
625 #define IRQ_PEND_EXT_II_MASK ((1UL << IRQ_PEND_EXT_CPU_TIMER) | \
626 (1UL << IRQ_PEND_EXT_CLOCK_COMP) | \
627 (1UL << IRQ_PEND_EXT_EMERGENCY) | \
628 (1UL << IRQ_PEND_EXT_EXTERNAL) | \
629 (1UL << IRQ_PEND_EXT_SERVICE) | \
630 (1UL << IRQ_PEND_EXT_SERVICE_EV))
631
632 struct kvm_s390_interrupt_info {
633 struct list_head list;
634 u64 type;
635 union {
636 struct kvm_s390_io_info io;
637 struct kvm_s390_ext_info ext;
638 struct kvm_s390_pgm_info pgm;
639 struct kvm_s390_emerg_info emerg;
640 struct kvm_s390_extcall_info extcall;
641 struct kvm_s390_prefix_info prefix;
642 struct kvm_s390_stop_info stop;
643 struct kvm_s390_mchk_info mchk;
644 };
645 };
646
647 struct kvm_s390_irq_payload {
648 struct kvm_s390_io_info io;
649 struct kvm_s390_ext_info ext;
650 struct kvm_s390_pgm_info pgm;
651 struct kvm_s390_emerg_info emerg;
652 struct kvm_s390_extcall_info extcall;
653 struct kvm_s390_prefix_info prefix;
654 struct kvm_s390_stop_info stop;
655 struct kvm_s390_mchk_info mchk;
656 };
657
658 struct kvm_s390_local_interrupt {
659 spinlock_t lock;
660 DECLARE_BITMAP(sigp_emerg_pending, KVM_MAX_VCPUS);
661 struct kvm_s390_irq_payload irq;
662 unsigned long pending_irqs;
663 };
664
665 #define FIRQ_LIST_IO_ISC_0 0
666 #define FIRQ_LIST_IO_ISC_1 1
667 #define FIRQ_LIST_IO_ISC_2 2
668 #define FIRQ_LIST_IO_ISC_3 3
669 #define FIRQ_LIST_IO_ISC_4 4
670 #define FIRQ_LIST_IO_ISC_5 5
671 #define FIRQ_LIST_IO_ISC_6 6
672 #define FIRQ_LIST_IO_ISC_7 7
673 #define FIRQ_LIST_PFAULT 8
674 #define FIRQ_LIST_VIRTIO 9
675 #define FIRQ_LIST_COUNT 10
676 #define FIRQ_CNTR_IO 0
677 #define FIRQ_CNTR_SERVICE 1
678 #define FIRQ_CNTR_VIRTIO 2
679 #define FIRQ_CNTR_PFAULT 3
680 #define FIRQ_MAX_COUNT 4
681
682 /* mask the AIS mode for a given ISC */
683 #define AIS_MODE_MASK(isc) (0x80 >> isc)
684
685 #define KVM_S390_AIS_MODE_ALL 0
686 #define KVM_S390_AIS_MODE_SINGLE 1
687
688 struct kvm_s390_float_interrupt {
689 unsigned long pending_irqs;
690 unsigned long masked_irqs;
691 spinlock_t lock;
692 struct list_head lists[FIRQ_LIST_COUNT];
693 int counters[FIRQ_MAX_COUNT];
694 struct kvm_s390_mchk_info mchk;
695 struct kvm_s390_ext_info srv_signal;
696 int next_rr_cpu;
697 struct mutex ais_lock;
698 u8 simm;
699 u8 nimm;
700 };
701
702 struct kvm_hw_wp_info_arch {
703 unsigned long addr;
704 unsigned long phys_addr;
705 int len;
706 char *old_data;
707 };
708
709 struct kvm_hw_bp_info_arch {
710 unsigned long addr;
711 int len;
712 };
713
714 /*
715 * Only the upper 16 bits of kvm_guest_debug->control are arch specific.
716 * Further KVM_GUESTDBG flags which an be used from userspace can be found in
717 * arch/s390/include/uapi/asm/kvm.h
718 */
719 #define KVM_GUESTDBG_EXIT_PENDING 0x10000000
720
721 #define guestdbg_enabled(vcpu) \
722 (vcpu->guest_debug & KVM_GUESTDBG_ENABLE)
723 #define guestdbg_sstep_enabled(vcpu) \
724 (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP)
725 #define guestdbg_hw_bp_enabled(vcpu) \
726 (vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP)
727 #define guestdbg_exit_pending(vcpu) (guestdbg_enabled(vcpu) && \
728 (vcpu->guest_debug & KVM_GUESTDBG_EXIT_PENDING))
729
730 #define KVM_GUESTDBG_VALID_MASK \
731 (KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_SINGLESTEP |\
732 KVM_GUESTDBG_USE_HW_BP | KVM_GUESTDBG_EXIT_PENDING)
733
734 struct kvm_guestdbg_info_arch {
735 unsigned long cr0;
736 unsigned long cr9;
737 unsigned long cr10;
738 unsigned long cr11;
739 struct kvm_hw_bp_info_arch *hw_bp_info;
740 struct kvm_hw_wp_info_arch *hw_wp_info;
741 int nr_hw_bp;
742 int nr_hw_wp;
743 unsigned long last_bp;
744 };
745
746 struct kvm_s390_pv_vcpu {
747 u64 handle;
748 unsigned long stor_base;
749 };
750
751 struct kvm_vcpu_arch {
752 struct kvm_s390_sie_block *sie_block;
753 /* if vsie is active, currently executed shadow sie control block */
754 struct kvm_s390_sie_block *vsie_block;
755 unsigned int host_acrs[NUM_ACRS];
756 struct gs_cb *host_gscb;
757 struct kvm_s390_local_interrupt local_int;
758 struct hrtimer ckc_timer;
759 struct kvm_s390_pgm_info pgm;
760 struct gmap *gmap;
761 struct kvm_guestdbg_info_arch guestdbg;
762 unsigned long pfault_token;
763 unsigned long pfault_select;
764 unsigned long pfault_compare;
765 bool cputm_enabled;
766 /*
767 * The seqcount protects updates to cputm_start and sie_block.cputm,
768 * this way we can have non-blocking reads with consistent values.
769 * Only the owning VCPU thread (vcpu->cpu) is allowed to change these
770 * values and to start/stop/enable/disable cpu timer accounting.
771 */
772 seqcount_t cputm_seqcount;
773 __u64 cputm_start;
774 bool gs_enabled;
775 bool skey_enabled;
776 /* Indicator if the access registers have been loaded from guest */
777 bool acrs_loaded;
778 struct kvm_s390_pv_vcpu pv;
779 union diag318_info diag318_info;
780 };
781
782 struct kvm_vm_stat {
783 struct kvm_vm_stat_generic generic;
784 u64 inject_io;
785 u64 inject_float_mchk;
786 u64 inject_pfault_done;
787 u64 inject_service_signal;
788 u64 inject_virtio;
789 u64 aen_forward;
790 u64 gmap_shadow_create;
791 u64 gmap_shadow_reuse;
792 u64 gmap_shadow_r1_entry;
793 u64 gmap_shadow_r2_entry;
794 u64 gmap_shadow_r3_entry;
795 u64 gmap_shadow_sg_entry;
796 u64 gmap_shadow_pg_entry;
797 };
798
799 struct kvm_arch_memory_slot {
800 };
801
802 struct s390_map_info {
803 struct list_head list;
804 __u64 guest_addr;
805 __u64 addr;
806 struct page *page;
807 };
808
809 struct s390_io_adapter {
810 unsigned int id;
811 int isc;
812 bool maskable;
813 bool masked;
814 bool swap;
815 bool suppressible;
816 };
817
818 #define MAX_S390_IO_ADAPTERS ((MAX_ISC + 1) * 8)
819 #define MAX_S390_ADAPTER_MAPS 256
820
821 /* maximum size of facilities and facility mask is 2k bytes */
822 #define S390_ARCH_FAC_LIST_SIZE_BYTE (1<<11)
823 #define S390_ARCH_FAC_LIST_SIZE_U64 \
824 (S390_ARCH_FAC_LIST_SIZE_BYTE / sizeof(u64))
825 #define S390_ARCH_FAC_MASK_SIZE_BYTE S390_ARCH_FAC_LIST_SIZE_BYTE
826 #define S390_ARCH_FAC_MASK_SIZE_U64 \
827 (S390_ARCH_FAC_MASK_SIZE_BYTE / sizeof(u64))
828
829 struct kvm_s390_cpu_model {
830 /* facility mask supported by kvm & hosting machine */
831 __u64 fac_mask[S390_ARCH_FAC_MASK_SIZE_U64];
832 struct kvm_s390_vm_cpu_subfunc subfuncs;
833 /* facility list requested by guest (in dma page) */
834 __u64 *fac_list;
835 u64 cpuid;
836 unsigned short ibc;
837 /* subset of available UV-features for pv-guests enabled by user space */
838 struct kvm_s390_vm_cpu_uv_feat uv_feat_guest;
839 };
840
841 typedef int (*crypto_hook)(struct kvm_vcpu *vcpu);
842
843 struct kvm_s390_crypto {
844 struct kvm_s390_crypto_cb *crycb;
845 struct rw_semaphore pqap_hook_rwsem;
846 crypto_hook *pqap_hook;
847 __u32 crycbd;
848 __u8 aes_kw;
849 __u8 dea_kw;
850 __u8 apie;
851 };
852
853 #define APCB0_MASK_SIZE 1
854 struct kvm_s390_apcb0 {
855 __u64 apm[APCB0_MASK_SIZE]; /* 0x0000 */
856 __u64 aqm[APCB0_MASK_SIZE]; /* 0x0008 */
857 __u64 adm[APCB0_MASK_SIZE]; /* 0x0010 */
858 __u64 reserved18; /* 0x0018 */
859 };
860
861 #define APCB1_MASK_SIZE 4
862 struct kvm_s390_apcb1 {
863 __u64 apm[APCB1_MASK_SIZE]; /* 0x0000 */
864 __u64 aqm[APCB1_MASK_SIZE]; /* 0x0020 */
865 __u64 adm[APCB1_MASK_SIZE]; /* 0x0040 */
866 __u64 reserved60[4]; /* 0x0060 */
867 };
868
869 struct kvm_s390_crypto_cb {
870 struct kvm_s390_apcb0 apcb0; /* 0x0000 */
871 __u8 reserved20[0x0048 - 0x0020]; /* 0x0020 */
872 __u8 dea_wrapping_key_mask[24]; /* 0x0048 */
873 __u8 aes_wrapping_key_mask[32]; /* 0x0060 */
874 struct kvm_s390_apcb1 apcb1; /* 0x0080 */
875 };
876
877 struct kvm_s390_gisa {
878 union {
879 struct { /* common to all formats */
880 u32 next_alert;
881 u8 ipm;
882 u8 reserved01[2];
883 u8 iam;
884 };
885 struct { /* format 0 */
886 u32 next_alert;
887 u8 ipm;
888 u8 reserved01;
889 u8 : 6;
890 u8 g : 1;
891 u8 c : 1;
892 u8 iam;
893 u8 reserved02[4];
894 u32 airq_count;
895 } g0;
896 struct { /* format 1 */
897 u32 next_alert;
898 u8 ipm;
899 u8 simm;
900 u8 nimm;
901 u8 iam;
902 u8 aism[8];
903 u8 : 6;
904 u8 g : 1;
905 u8 c : 1;
906 u8 reserved03[11];
907 u32 airq_count;
908 } g1;
909 struct {
910 u64 word[4];
911 } u64;
912 };
913 };
914
915 struct kvm_s390_gib {
916 u32 alert_list_origin;
917 u32 reserved01;
918 u8:5;
919 u8 nisc:3;
920 u8 reserved03[3];
921 u32 reserved04[5];
922 };
923
924 /*
925 * sie_page2 has to be allocated as DMA because fac_list, crycb and
926 * gisa need 31bit addresses in the sie control block.
927 */
928 struct sie_page2 {
929 __u64 fac_list[S390_ARCH_FAC_LIST_SIZE_U64]; /* 0x0000 */
930 struct kvm_s390_crypto_cb crycb; /* 0x0800 */
931 struct kvm_s390_gisa gisa; /* 0x0900 */
932 struct kvm *kvm; /* 0x0920 */
933 u8 reserved928[0x1000 - 0x928]; /* 0x0928 */
934 };
935
936 struct vsie_page;
937
938 struct kvm_s390_vsie {
939 struct mutex mutex;
940 struct radix_tree_root addr_to_page;
941 int page_count;
942 int next;
943 struct vsie_page *pages[KVM_MAX_VCPUS];
944 };
945
946 struct kvm_s390_gisa_iam {
947 u8 mask;
948 spinlock_t ref_lock;
949 u32 ref_count[MAX_ISC + 1];
950 };
951
952 struct kvm_s390_gisa_interrupt {
953 struct kvm_s390_gisa *origin;
954 struct kvm_s390_gisa_iam alert;
955 struct hrtimer timer;
956 u64 expires;
957 DECLARE_BITMAP(kicked_mask, KVM_MAX_VCPUS);
958 };
959
960 struct kvm_s390_pv {
961 u64 handle;
962 u64 guest_len;
963 unsigned long stor_base;
964 void *stor_var;
965 bool dumping;
966 void *set_aside;
967 struct list_head need_cleanup;
968 struct mmu_notifier mmu_notifier;
969 };
970
971 struct kvm_arch{
972 void *sca;
973 int use_esca;
974 rwlock_t sca_lock;
975 debug_info_t *dbf;
976 struct kvm_s390_float_interrupt float_int;
977 struct kvm_device *flic;
978 struct gmap *gmap;
979 unsigned long mem_limit;
980 int css_support;
981 int use_irqchip;
982 int use_cmma;
983 int use_pfmfi;
984 int use_skf;
985 int use_zpci_interp;
986 int user_cpu_state_ctrl;
987 int user_sigp;
988 int user_stsi;
989 int user_instr0;
990 struct s390_io_adapter *adapters[MAX_S390_IO_ADAPTERS];
991 wait_queue_head_t ipte_wq;
992 int ipte_lock_count;
993 struct mutex ipte_mutex;
994 spinlock_t start_stop_lock;
995 struct sie_page2 *sie_page2;
996 struct kvm_s390_cpu_model model;
997 struct kvm_s390_crypto crypto;
998 struct kvm_s390_vsie vsie;
999 u8 epdx;
1000 u64 epoch;
1001 int migration_mode;
1002 atomic64_t cmma_dirty_pages;
1003 /* subset of available cpu features enabled by user space */
1004 DECLARE_BITMAP(cpu_feat, KVM_S390_VM_CPU_FEAT_NR_BITS);
1005 /* indexed by vcpu_idx */
1006 DECLARE_BITMAP(idle_mask, KVM_MAX_VCPUS);
1007 struct kvm_s390_gisa_interrupt gisa_int;
1008 struct kvm_s390_pv pv;
1009 struct list_head kzdev_list;
1010 spinlock_t kzdev_list_lock;
1011 };
1012
1013 #define KVM_HVA_ERR_BAD (-1UL)
1014 #define KVM_HVA_ERR_RO_BAD (-2UL)
1015
kvm_is_error_hva(unsigned long addr)1016 static inline bool kvm_is_error_hva(unsigned long addr)
1017 {
1018 return IS_ERR_VALUE(addr);
1019 }
1020
1021 #define ASYNC_PF_PER_VCPU 64
1022 struct kvm_arch_async_pf {
1023 unsigned long pfault_token;
1024 };
1025
1026 bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu);
1027
1028 void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu,
1029 struct kvm_async_pf *work);
1030
1031 bool kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu,
1032 struct kvm_async_pf *work);
1033
1034 void kvm_arch_async_page_present(struct kvm_vcpu *vcpu,
1035 struct kvm_async_pf *work);
1036
kvm_arch_async_page_present_queued(struct kvm_vcpu * vcpu)1037 static inline void kvm_arch_async_page_present_queued(struct kvm_vcpu *vcpu) {}
1038
1039 void kvm_arch_crypto_clear_masks(struct kvm *kvm);
1040 void kvm_arch_crypto_set_masks(struct kvm *kvm, unsigned long *apm,
1041 unsigned long *aqm, unsigned long *adm);
1042
1043 int __sie64a(phys_addr_t sie_block_phys, struct kvm_s390_sie_block *sie_block, u64 *rsa,
1044 unsigned long gasce);
1045
sie64a(struct kvm_s390_sie_block * sie_block,u64 * rsa,unsigned long gasce)1046 static inline int sie64a(struct kvm_s390_sie_block *sie_block, u64 *rsa, unsigned long gasce)
1047 {
1048 return __sie64a(virt_to_phys(sie_block), sie_block, rsa, gasce);
1049 }
1050
1051 extern char sie_exit;
1052
1053 bool kvm_s390_pv_is_protected(struct kvm *kvm);
1054 bool kvm_s390_pv_cpu_is_protected(struct kvm_vcpu *vcpu);
1055
1056 extern int kvm_s390_gisc_register(struct kvm *kvm, u32 gisc);
1057 extern int kvm_s390_gisc_unregister(struct kvm *kvm, u32 gisc);
1058
kvm_arch_sync_events(struct kvm * kvm)1059 static inline void kvm_arch_sync_events(struct kvm *kvm) {}
kvm_arch_free_memslot(struct kvm * kvm,struct kvm_memory_slot * slot)1060 static inline void kvm_arch_free_memslot(struct kvm *kvm,
1061 struct kvm_memory_slot *slot) {}
kvm_arch_memslots_updated(struct kvm * kvm,u64 gen)1062 static inline void kvm_arch_memslots_updated(struct kvm *kvm, u64 gen) {}
kvm_arch_flush_shadow_all(struct kvm * kvm)1063 static inline void kvm_arch_flush_shadow_all(struct kvm *kvm) {}
kvm_arch_flush_shadow_memslot(struct kvm * kvm,struct kvm_memory_slot * slot)1064 static inline void kvm_arch_flush_shadow_memslot(struct kvm *kvm,
1065 struct kvm_memory_slot *slot) {}
kvm_arch_vcpu_blocking(struct kvm_vcpu * vcpu)1066 static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {}
kvm_arch_vcpu_unblocking(struct kvm_vcpu * vcpu)1067 static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {}
1068
1069 #define __KVM_HAVE_ARCH_VM_FREE
1070 void kvm_arch_free_vm(struct kvm *kvm);
1071
1072 struct zpci_kvm_hook {
1073 int (*kvm_register)(void *opaque, struct kvm *kvm);
1074 void (*kvm_unregister)(void *opaque);
1075 };
1076
1077 extern struct zpci_kvm_hook zpci_kvm_hook;
1078
1079 #endif
1080