xref: /linux/arch/powerpc/include/asm/lppaca.h (revision c4ee0af3fa0dc65f690fc908f02b8355f9576ea0)
1 /*
2  * lppaca.h
3  * Copyright (C) 2001  Mike Corrigan IBM Corporation
4  *
5  * This program is free software; you can redistribute it and/or modify
6  * it under the terms of the GNU General Public License as published by
7  * the Free Software Foundation; either version 2 of the License, or
8  * (at your option) any later version.
9  *
10  * This program is distributed in the hope that it will be useful,
11  * but WITHOUT ANY WARRANTY; without even the implied warranty of
12  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
13  * GNU General Public License for more details.
14  *
15  * You should have received a copy of the GNU General Public License
16  * along with this program; if not, write to the Free Software
17  * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307 USA
18  */
19 #ifndef _ASM_POWERPC_LPPACA_H
20 #define _ASM_POWERPC_LPPACA_H
21 #ifdef __KERNEL__
22 
23 /*
24  * These definitions relate to hypervisors that only exist when using
25  * a server type processor
26  */
27 #ifdef CONFIG_PPC_BOOK3S
28 
29 /*
30  * This control block contains the data that is shared between the
31  * hypervisor and the OS.
32  */
33 #include <linux/cache.h>
34 #include <linux/threads.h>
35 #include <asm/types.h>
36 #include <asm/mmu.h>
37 
38 /*
39  * We only have to have statically allocated lppaca structs on
40  * legacy iSeries, which supports at most 64 cpus.
41  */
42 #define NR_LPPACAS	1
43 
44 /*
45  * The Hypervisor barfs if the lppaca crosses a page boundary.  A 1k
46  * alignment is sufficient to prevent this
47  */
48 struct lppaca {
49 	/* cacheline 1 contains read-only data */
50 
51 	__be32	desc;			/* Eye catcher 0xD397D781 */
52 	__be16	size;			/* Size of this struct */
53 	u8	reserved1[3];
54 	u8	__old_status;		/* Old status, including shared proc */
55 	u8	reserved3[14];
56 	volatile __be32 dyn_hw_node_id;	/* Dynamic hardware node id */
57 	volatile __be32 dyn_hw_proc_id;	/* Dynamic hardware proc id */
58 	u8	reserved4[56];
59 	volatile u8 vphn_assoc_counts[8]; /* Virtual processor home node */
60 					  /* associativity change counters */
61 	u8	reserved5[32];
62 
63 	/* cacheline 2 contains local read-write data */
64 
65 	u8	reserved6[48];
66 	u8	cede_latency_hint;
67 	u8	ebb_regs_in_use;
68 	u8	reserved7[6];
69 	u8	dtl_enable_mask;	/* Dispatch Trace Log mask */
70 	u8	donate_dedicated_cpu;	/* Donate dedicated CPU cycles */
71 	u8	fpregs_in_use;
72 	u8	pmcregs_in_use;
73 	u8	reserved8[28];
74 	__be64	wait_state_cycles;	/* Wait cycles for this proc */
75 	u8	reserved9[28];
76 	__be16	slb_count;		/* # of SLBs to maintain */
77 	u8	idle;			/* Indicate OS is idle */
78 	u8	vmxregs_in_use;
79 
80 	/* cacheline 3 is shared with other processors */
81 
82 	/*
83 	 * This is the yield_count.  An "odd" value (low bit on) means that
84 	 * the processor is yielded (either because of an OS yield or a
85 	 * hypervisor preempt).  An even value implies that the processor is
86 	 * currently executing.
87 	 * NOTE: Even dedicated processor partitions can yield so this
88 	 * field cannot be used to determine if we are shared or dedicated.
89 	 */
90 	volatile __be32 yield_count;
91 	volatile __be32 dispersion_count; /* dispatch changed physical cpu */
92 	volatile __be64 cmo_faults;	/* CMO page fault count */
93 	volatile __be64 cmo_fault_time;	/* CMO page fault time */
94 	u8	reserved10[104];
95 
96 	/* cacheline 4-5 */
97 
98 	__be32	page_ins;		/* CMO Hint - # page ins by OS */
99 	u8	reserved11[148];
100 	volatile __be64 dtl_idx;		/* Dispatch Trace Log head index */
101 	u8	reserved12[96];
102 } __attribute__((__aligned__(0x400)));
103 
104 extern struct lppaca lppaca[];
105 
106 #define lppaca_of(cpu)	(*paca[cpu].lppaca_ptr)
107 
108 /*
109  * We are using a non architected field to determine if a partition is
110  * shared or dedicated. This currently works on both KVM and PHYP, but
111  * we will have to transition to something better.
112  */
113 #define LPPACA_OLD_SHARED_PROC		2
114 
115 static inline bool lppaca_shared_proc(struct lppaca *l)
116 {
117 	return !!(l->__old_status & LPPACA_OLD_SHARED_PROC);
118 }
119 
120 /*
121  * SLB shadow buffer structure as defined in the PAPR.  The save_area
122  * contains adjacent ESID and VSID pairs for each shadowed SLB.  The
123  * ESID is stored in the lower 64bits, then the VSID.
124  */
125 struct slb_shadow {
126 	__be32	persistent;		/* Number of persistent SLBs */
127 	__be32	buffer_length;		/* Total shadow buffer length */
128 	__be64	reserved;
129 	struct	{
130 		__be64     esid;
131 		__be64	vsid;
132 	} save_area[SLB_NUM_BOLTED];
133 } ____cacheline_aligned;
134 
135 extern struct slb_shadow slb_shadow[];
136 
137 /*
138  * Layout of entries in the hypervisor's dispatch trace log buffer.
139  */
140 struct dtl_entry {
141 	u8	dispatch_reason;
142 	u8	preempt_reason;
143 	__be16	processor_id;
144 	__be32	enqueue_to_dispatch_time;
145 	__be32	ready_to_enqueue_time;
146 	__be32	waiting_to_ready_time;
147 	__be64	timebase;
148 	__be64	fault_addr;
149 	__be64	srr0;
150 	__be64	srr1;
151 };
152 
153 #define DISPATCH_LOG_BYTES	4096	/* bytes per cpu */
154 #define N_DISPATCH_LOG		(DISPATCH_LOG_BYTES / sizeof(struct dtl_entry))
155 
156 extern struct kmem_cache *dtl_cache;
157 
158 /*
159  * When CONFIG_VIRT_CPU_ACCOUNTING_NATIVE = y, the cpu accounting code controls
160  * reading from the dispatch trace log.  If other code wants to consume
161  * DTL entries, it can set this pointer to a function that will get
162  * called once for each DTL entry that gets processed.
163  */
164 extern void (*dtl_consumer)(struct dtl_entry *entry, u64 index);
165 
166 #endif /* CONFIG_PPC_BOOK3S */
167 #endif /* __KERNEL__ */
168 #endif /* _ASM_POWERPC_LPPACA_H */
169