1 /****************************************************************************** 2 * sysctl.h 3 * 4 * System management operations. For use by node control stack. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a copy 7 * of this software and associated documentation files (the "Software"), to 8 * deal in the Software without restriction, including without limitation the 9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or 10 * sell copies of the Software, and to permit persons to whom the Software is 11 * furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 21 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER 22 * DEALINGS IN THE SOFTWARE. 23 * 24 * Copyright (c) 2002-2006, K Fraser 25 */ 26 27 #ifndef __XEN_PUBLIC_SYSCTL_H__ 28 #define __XEN_PUBLIC_SYSCTL_H__ 29 30 #if !defined(__XEN__) && !defined(__XEN_TOOLS__) 31 #error "sysctl operations are intended for use by node control tools only" 32 #endif 33 34 #include "xen.h" 35 #include "domctl.h" 36 #include "physdev.h" 37 38 #define XEN_SYSCTL_INTERFACE_VERSION 0x00000014 39 40 /* 41 * Read console content from Xen buffer ring. 42 */ 43 /* XEN_SYSCTL_readconsole */ 44 struct xen_sysctl_readconsole { 45 /* IN: Non-zero -> clear after reading. */ 46 uint8_t clear; 47 /* IN: Non-zero -> start index specified by @index field. */ 48 uint8_t incremental; 49 uint8_t pad0, pad1; 50 /* 51 * IN: Start index for consuming from ring buffer (if @incremental); 52 * OUT: End index after consuming from ring buffer. 53 */ 54 uint32_t index; 55 /* IN: Virtual address to write console data. */ 56 XEN_GUEST_HANDLE_64(char) buffer; 57 /* IN: Size of buffer; OUT: Bytes written to buffer. */ 58 uint32_t count; 59 }; 60 61 /* Get trace buffers machine base address */ 62 /* XEN_SYSCTL_tbuf_op */ 63 struct xen_sysctl_tbuf_op { 64 /* IN variables */ 65 #define XEN_SYSCTL_TBUFOP_get_info 0 66 #define XEN_SYSCTL_TBUFOP_set_cpu_mask 1 67 #define XEN_SYSCTL_TBUFOP_set_evt_mask 2 68 #define XEN_SYSCTL_TBUFOP_set_size 3 69 #define XEN_SYSCTL_TBUFOP_enable 4 70 #define XEN_SYSCTL_TBUFOP_disable 5 71 uint32_t cmd; 72 /* IN/OUT variables */ 73 struct xenctl_bitmap cpu_mask; 74 uint32_t evt_mask; 75 /* OUT variables */ 76 uint64_aligned_t buffer_mfn; 77 uint32_t size; /* Also an IN variable! */ 78 }; 79 80 /* 81 * Get physical information about the host machine 82 */ 83 /* XEN_SYSCTL_physinfo */ 84 /* The platform supports HVM guests. */ 85 #define _XEN_SYSCTL_PHYSCAP_hvm 0 86 #define XEN_SYSCTL_PHYSCAP_hvm (1u<<_XEN_SYSCTL_PHYSCAP_hvm) 87 /* The platform supports PV guests. */ 88 #define _XEN_SYSCTL_PHYSCAP_pv 1 89 #define XEN_SYSCTL_PHYSCAP_pv (1u<<_XEN_SYSCTL_PHYSCAP_pv) 90 /* The platform supports direct access to I/O devices with IOMMU. */ 91 #define _XEN_SYSCTL_PHYSCAP_directio 2 92 #define XEN_SYSCTL_PHYSCAP_directio (1u<<_XEN_SYSCTL_PHYSCAP_directio) 93 /* The platform supports Hardware Assisted Paging. */ 94 #define _XEN_SYSCTL_PHYSCAP_hap 3 95 #define XEN_SYSCTL_PHYSCAP_hap (1u<<_XEN_SYSCTL_PHYSCAP_hap) 96 /* The platform supports software paging. */ 97 #define _XEN_SYSCTL_PHYSCAP_shadow 4 98 #define XEN_SYSCTL_PHYSCAP_shadow (1u<<_XEN_SYSCTL_PHYSCAP_shadow) 99 /* The platform supports sharing of HAP page tables with the IOMMU. */ 100 #define _XEN_SYSCTL_PHYSCAP_iommu_hap_pt_share 5 101 #define XEN_SYSCTL_PHYSCAP_iommu_hap_pt_share \ 102 (1u << _XEN_SYSCTL_PHYSCAP_iommu_hap_pt_share) 103 #define XEN_SYSCTL_PHYSCAP_vmtrace (1u << 6) 104 /* The platform supports vPMU. */ 105 #define XEN_SYSCTL_PHYSCAP_vpmu (1u << 7) 106 107 /* Xen supports the Grant v1 and/or v2 ABIs. */ 108 #define XEN_SYSCTL_PHYSCAP_gnttab_v1 (1u << 8) 109 #define XEN_SYSCTL_PHYSCAP_gnttab_v2 (1u << 9) 110 111 /* Max XEN_SYSCTL_PHYSCAP_* constant. Used for ABI checking. */ 112 #define XEN_SYSCTL_PHYSCAP_MAX XEN_SYSCTL_PHYSCAP_gnttab_v2 113 114 struct xen_sysctl_physinfo { 115 uint32_t threads_per_core; 116 uint32_t cores_per_socket; 117 uint32_t nr_cpus; /* # CPUs currently online */ 118 uint32_t max_cpu_id; /* Largest possible CPU ID on this host */ 119 uint32_t nr_nodes; /* # nodes currently online */ 120 uint32_t max_node_id; /* Largest possible node ID on this host */ 121 uint32_t cpu_khz; 122 uint32_t capabilities;/* XEN_SYSCTL_PHYSCAP_??? */ 123 uint64_aligned_t total_pages; 124 uint64_aligned_t free_pages; 125 uint64_aligned_t scrub_pages; 126 uint64_aligned_t outstanding_pages; 127 uint64_aligned_t max_mfn; /* Largest possible MFN on this host */ 128 uint32_t hw_cap[8]; 129 }; 130 131 /* 132 * Get the ID of the current scheduler. 133 */ 134 /* XEN_SYSCTL_sched_id */ 135 struct xen_sysctl_sched_id { 136 /* OUT variable */ 137 uint32_t sched_id; 138 }; 139 140 /* Interface for controlling Xen software performance counters. */ 141 /* XEN_SYSCTL_perfc_op */ 142 /* Sub-operations: */ 143 #define XEN_SYSCTL_PERFCOP_reset 1 /* Reset all counters to zero. */ 144 #define XEN_SYSCTL_PERFCOP_query 2 /* Get perfctr information. */ 145 struct xen_sysctl_perfc_desc { 146 char name[80]; /* name of perf counter */ 147 uint32_t nr_vals; /* number of values for this counter */ 148 }; 149 typedef struct xen_sysctl_perfc_desc xen_sysctl_perfc_desc_t; 150 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_perfc_desc_t); 151 typedef uint32_t xen_sysctl_perfc_val_t; 152 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_perfc_val_t); 153 154 struct xen_sysctl_perfc_op { 155 /* IN variables. */ 156 uint32_t cmd; /* XEN_SYSCTL_PERFCOP_??? */ 157 /* OUT variables. */ 158 uint32_t nr_counters; /* number of counters description */ 159 uint32_t nr_vals; /* number of values */ 160 /* counter information (or NULL) */ 161 XEN_GUEST_HANDLE_64(xen_sysctl_perfc_desc_t) desc; 162 /* counter values (or NULL) */ 163 XEN_GUEST_HANDLE_64(xen_sysctl_perfc_val_t) val; 164 }; 165 166 /* XEN_SYSCTL_getdomaininfolist */ 167 struct xen_sysctl_getdomaininfolist { 168 /* IN variables. */ 169 domid_t first_domain; 170 uint32_t max_domains; 171 XEN_GUEST_HANDLE_64(xen_domctl_getdomaininfo_t) buffer; 172 /* OUT variables. */ 173 uint32_t num_domains; 174 }; 175 176 /* Inject debug keys into Xen. */ 177 /* XEN_SYSCTL_debug_keys */ 178 struct xen_sysctl_debug_keys { 179 /* IN variables. */ 180 XEN_GUEST_HANDLE_64(const_char) keys; 181 uint32_t nr_keys; 182 }; 183 184 /* Get physical CPU information. */ 185 /* XEN_SYSCTL_getcpuinfo */ 186 struct xen_sysctl_cpuinfo { 187 uint64_aligned_t idletime; 188 }; 189 typedef struct xen_sysctl_cpuinfo xen_sysctl_cpuinfo_t; 190 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cpuinfo_t); 191 struct xen_sysctl_getcpuinfo { 192 /* IN variables. */ 193 uint32_t max_cpus; 194 XEN_GUEST_HANDLE_64(xen_sysctl_cpuinfo_t) info; 195 /* OUT variables. */ 196 uint32_t nr_cpus; 197 }; 198 199 /* XEN_SYSCTL_availheap */ 200 struct xen_sysctl_availheap { 201 /* IN variables. */ 202 uint32_t min_bitwidth; /* Smallest address width (zero if don't care). */ 203 uint32_t max_bitwidth; /* Largest address width (zero if don't care). */ 204 int32_t node; /* NUMA node of interest (-1 for all nodes). */ 205 /* OUT variables. */ 206 uint64_aligned_t avail_bytes;/* Bytes available in the specified region. */ 207 }; 208 209 /* XEN_SYSCTL_get_pmstat */ 210 struct pm_px_val { 211 uint64_aligned_t freq; /* Px core frequency */ 212 uint64_aligned_t residency; /* Px residency time */ 213 uint64_aligned_t count; /* Px transition count */ 214 }; 215 typedef struct pm_px_val pm_px_val_t; 216 DEFINE_XEN_GUEST_HANDLE(pm_px_val_t); 217 218 struct pm_px_stat { 219 uint8_t total; /* total Px states */ 220 uint8_t usable; /* usable Px states */ 221 uint8_t last; /* last Px state */ 222 uint8_t cur; /* current Px state */ 223 XEN_GUEST_HANDLE_64(uint64) trans_pt; /* Px transition table */ 224 XEN_GUEST_HANDLE_64(pm_px_val_t) pt; 225 }; 226 227 struct pm_cx_stat { 228 uint32_t nr; /* entry nr in triggers & residencies, including C0 */ 229 uint32_t last; /* last Cx state */ 230 uint64_aligned_t idle_time; /* idle time from boot */ 231 XEN_GUEST_HANDLE_64(uint64) triggers; /* Cx trigger counts */ 232 XEN_GUEST_HANDLE_64(uint64) residencies; /* Cx residencies */ 233 uint32_t nr_pc; /* entry nr in pc[] */ 234 uint32_t nr_cc; /* entry nr in cc[] */ 235 /* 236 * These two arrays may (and generally will) have unused slots; slots not 237 * having a corresponding hardware register will not be written by the 238 * hypervisor. It is therefore up to the caller to put a suitable sentinel 239 * into all slots before invoking the function. 240 * Indexing is 1-biased (PC1/CC1 being at index 0). 241 */ 242 XEN_GUEST_HANDLE_64(uint64) pc; 243 XEN_GUEST_HANDLE_64(uint64) cc; 244 }; 245 246 struct xen_sysctl_get_pmstat { 247 #define PMSTAT_CATEGORY_MASK 0xf0 248 #define PMSTAT_PX 0x10 249 #define PMSTAT_CX 0x20 250 #define PMSTAT_get_max_px (PMSTAT_PX | 0x1) 251 #define PMSTAT_get_pxstat (PMSTAT_PX | 0x2) 252 #define PMSTAT_reset_pxstat (PMSTAT_PX | 0x3) 253 #define PMSTAT_get_max_cx (PMSTAT_CX | 0x1) 254 #define PMSTAT_get_cxstat (PMSTAT_CX | 0x2) 255 #define PMSTAT_reset_cxstat (PMSTAT_CX | 0x3) 256 uint32_t type; 257 uint32_t cpuid; 258 union { 259 struct pm_px_stat getpx; 260 struct pm_cx_stat getcx; 261 /* other struct for tx, etc */ 262 } u; 263 }; 264 265 /* XEN_SYSCTL_cpu_hotplug */ 266 struct xen_sysctl_cpu_hotplug { 267 /* IN variables */ 268 uint32_t cpu; /* Physical cpu. */ 269 270 /* Single CPU enable/disable. */ 271 #define XEN_SYSCTL_CPU_HOTPLUG_ONLINE 0 272 #define XEN_SYSCTL_CPU_HOTPLUG_OFFLINE 1 273 274 /* 275 * SMT enable/disable. 276 * 277 * These two ops loop over all present CPUs, and either online or offline 278 * every non-primary sibling thread (those with a thread id which is not 279 * 0). This behaviour is chosen to simplify the implementation. 280 * 281 * They are intended as a shorthand for identifying and feeding the cpu 282 * numbers individually to HOTPLUG_{ON,OFF}LINE. 283 * 284 * These are not expected to be used in conjunction with debugging options 285 * such as `maxcpus=` or when other manual configuration of offline cpus 286 * is in use. 287 */ 288 #define XEN_SYSCTL_CPU_HOTPLUG_SMT_ENABLE 2 289 #define XEN_SYSCTL_CPU_HOTPLUG_SMT_DISABLE 3 290 uint32_t op; /* hotplug opcode */ 291 }; 292 293 /* 294 * Get/set xen power management, include 295 * 1. cpufreq governors and related parameters 296 */ 297 /* XEN_SYSCTL_pm_op */ 298 struct xen_userspace { 299 uint32_t scaling_setspeed; 300 }; 301 302 struct xen_ondemand { 303 uint32_t sampling_rate_max; 304 uint32_t sampling_rate_min; 305 306 uint32_t sampling_rate; 307 uint32_t up_threshold; 308 }; 309 310 /* 311 * cpufreq para name of this structure named 312 * same as sysfs file name of native linux 313 */ 314 #define CPUFREQ_NAME_LEN 16 315 struct xen_get_cpufreq_para { 316 /* IN/OUT variable */ 317 uint32_t cpu_num; 318 uint32_t freq_num; 319 uint32_t gov_num; 320 321 /* for all governors */ 322 /* OUT variable */ 323 XEN_GUEST_HANDLE_64(uint32) affected_cpus; 324 XEN_GUEST_HANDLE_64(uint32) scaling_available_frequencies; 325 XEN_GUEST_HANDLE_64(char) scaling_available_governors; 326 char scaling_driver[CPUFREQ_NAME_LEN]; 327 328 uint32_t cpuinfo_cur_freq; 329 uint32_t cpuinfo_max_freq; 330 uint32_t cpuinfo_min_freq; 331 uint32_t scaling_cur_freq; 332 333 char scaling_governor[CPUFREQ_NAME_LEN]; 334 uint32_t scaling_max_freq; 335 uint32_t scaling_min_freq; 336 337 /* for specific governor */ 338 union { 339 struct xen_userspace userspace; 340 struct xen_ondemand ondemand; 341 } u; 342 343 int32_t turbo_enabled; 344 }; 345 346 struct xen_set_cpufreq_gov { 347 char scaling_governor[CPUFREQ_NAME_LEN]; 348 }; 349 350 struct xen_set_cpufreq_para { 351 #define SCALING_MAX_FREQ 1 352 #define SCALING_MIN_FREQ 2 353 #define SCALING_SETSPEED 3 354 #define SAMPLING_RATE 4 355 #define UP_THRESHOLD 5 356 357 uint32_t ctrl_type; 358 uint32_t ctrl_value; 359 }; 360 361 struct xen_sysctl_pm_op { 362 #define PM_PARA_CATEGORY_MASK 0xf0 363 #define CPUFREQ_PARA 0x10 364 365 /* cpufreq command type */ 366 #define GET_CPUFREQ_PARA (CPUFREQ_PARA | 0x01) 367 #define SET_CPUFREQ_GOV (CPUFREQ_PARA | 0x02) 368 #define SET_CPUFREQ_PARA (CPUFREQ_PARA | 0x03) 369 #define GET_CPUFREQ_AVGFREQ (CPUFREQ_PARA | 0x04) 370 371 /* set/reset scheduler power saving option */ 372 #define XEN_SYSCTL_pm_op_set_sched_opt_smt 0x21 373 374 /* 375 * cpuidle max C-state and max C-sub-state access command: 376 * Set cpuid to 0 for max C-state. 377 * Set cpuid to 1 for max C-sub-state. 378 */ 379 #define XEN_SYSCTL_pm_op_get_max_cstate 0x22 380 #define XEN_SYSCTL_pm_op_set_max_cstate 0x23 381 382 /* set scheduler migration cost value */ 383 #define XEN_SYSCTL_pm_op_set_vcpu_migration_delay 0x24 384 #define XEN_SYSCTL_pm_op_get_vcpu_migration_delay 0x25 385 386 /* enable/disable turbo mode when in dbs governor */ 387 #define XEN_SYSCTL_pm_op_enable_turbo 0x26 388 #define XEN_SYSCTL_pm_op_disable_turbo 0x27 389 390 uint32_t cmd; 391 uint32_t cpuid; 392 union { 393 struct xen_get_cpufreq_para get_para; 394 struct xen_set_cpufreq_gov set_gov; 395 struct xen_set_cpufreq_para set_para; 396 uint64_aligned_t get_avgfreq; 397 uint32_t set_sched_opt_smt; 398 #define XEN_SYSCTL_CX_UNLIMITED 0xffffffff 399 uint32_t get_max_cstate; 400 uint32_t set_max_cstate; 401 } u; 402 }; 403 404 /* XEN_SYSCTL_page_offline_op */ 405 struct xen_sysctl_page_offline_op { 406 /* IN: range of page to be offlined */ 407 #define sysctl_page_offline 1 408 #define sysctl_page_online 2 409 #define sysctl_query_page_offline 3 410 uint32_t cmd; 411 uint32_t start; 412 uint32_t end; 413 /* OUT: result of page offline request */ 414 /* 415 * bit 0~15: result flags 416 * bit 16~31: owner 417 */ 418 XEN_GUEST_HANDLE(uint32) status; 419 }; 420 421 #define PG_OFFLINE_STATUS_MASK (0xFFUL) 422 423 /* The result is invalid, i.e. HV does not handle it */ 424 #define PG_OFFLINE_INVALID (0x1UL << 0) 425 426 #define PG_OFFLINE_OFFLINED (0x1UL << 1) 427 #define PG_OFFLINE_PENDING (0x1UL << 2) 428 #define PG_OFFLINE_FAILED (0x1UL << 3) 429 #define PG_OFFLINE_AGAIN (0x1UL << 4) 430 431 #define PG_ONLINE_FAILED PG_OFFLINE_FAILED 432 #define PG_ONLINE_ONLINED PG_OFFLINE_OFFLINED 433 434 #define PG_OFFLINE_STATUS_OFFLINED (0x1UL << 1) 435 #define PG_OFFLINE_STATUS_ONLINE (0x1UL << 2) 436 #define PG_OFFLINE_STATUS_OFFLINE_PENDING (0x1UL << 3) 437 #define PG_OFFLINE_STATUS_BROKEN (0x1UL << 4) 438 439 #define PG_OFFLINE_MISC_MASK (0xFFUL << 4) 440 441 /* valid when PG_OFFLINE_FAILED or PG_OFFLINE_PENDING */ 442 #define PG_OFFLINE_XENPAGE (0x1UL << 8) 443 #define PG_OFFLINE_DOM0PAGE (0x1UL << 9) 444 #define PG_OFFLINE_ANONYMOUS (0x1UL << 10) 445 #define PG_OFFLINE_NOT_CONV_RAM (0x1UL << 11) 446 #define PG_OFFLINE_OWNED (0x1UL << 12) 447 448 #define PG_OFFLINE_BROKEN (0x1UL << 13) 449 #define PG_ONLINE_BROKEN PG_OFFLINE_BROKEN 450 451 #define PG_OFFLINE_OWNER_SHIFT 16 452 453 /* XEN_SYSCTL_lockprof_op */ 454 /* Sub-operations: */ 455 #define XEN_SYSCTL_LOCKPROF_reset 1 /* Reset all profile data to zero. */ 456 #define XEN_SYSCTL_LOCKPROF_query 2 /* Get lock profile information. */ 457 /* Record-type: */ 458 #define LOCKPROF_TYPE_GLOBAL 0 /* global lock, idx meaningless */ 459 #define LOCKPROF_TYPE_PERDOM 1 /* per-domain lock, idx is domid */ 460 #define LOCKPROF_TYPE_N 2 /* number of types */ 461 struct xen_sysctl_lockprof_data { 462 char name[40]; /* lock name (may include up to 2 %d specifiers) */ 463 int32_t type; /* LOCKPROF_TYPE_??? */ 464 int32_t idx; /* index (e.g. domain id) */ 465 uint64_aligned_t lock_cnt; /* # of locking succeeded */ 466 uint64_aligned_t block_cnt; /* # of wait for lock */ 467 uint64_aligned_t lock_time; /* nsecs lock held */ 468 uint64_aligned_t block_time; /* nsecs waited for lock */ 469 }; 470 typedef struct xen_sysctl_lockprof_data xen_sysctl_lockprof_data_t; 471 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_lockprof_data_t); 472 struct xen_sysctl_lockprof_op { 473 /* IN variables. */ 474 uint32_t cmd; /* XEN_SYSCTL_LOCKPROF_??? */ 475 uint32_t max_elem; /* size of output buffer */ 476 /* OUT variables (query only). */ 477 uint32_t nr_elem; /* number of elements available */ 478 uint64_aligned_t time; /* nsecs of profile measurement */ 479 /* profile information (or NULL) */ 480 XEN_GUEST_HANDLE_64(xen_sysctl_lockprof_data_t) data; 481 }; 482 483 /* XEN_SYSCTL_cputopoinfo */ 484 #define XEN_INVALID_CORE_ID (~0U) 485 #define XEN_INVALID_SOCKET_ID (~0U) 486 #define XEN_INVALID_NODE_ID (~0U) 487 488 struct xen_sysctl_cputopo { 489 uint32_t core; 490 uint32_t socket; 491 uint32_t node; 492 }; 493 typedef struct xen_sysctl_cputopo xen_sysctl_cputopo_t; 494 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cputopo_t); 495 496 /* 497 * IN: 498 * - a NULL 'cputopo' handle is a request for maximun 'num_cpus'. 499 * - otherwise it's the number of entries in 'cputopo' 500 * 501 * OUT: 502 * - If 'num_cpus' is less than the number Xen wants to write but the handle 503 * handle is not a NULL one, partial data gets returned and 'num_cpus' gets 504 * updated to reflect the intended number. 505 * - Otherwise, 'num_cpus' shall indicate the number of entries written, which 506 * may be less than the input value. 507 */ 508 struct xen_sysctl_cputopoinfo { 509 uint32_t num_cpus; 510 XEN_GUEST_HANDLE_64(xen_sysctl_cputopo_t) cputopo; 511 }; 512 513 /* XEN_SYSCTL_numainfo */ 514 #define XEN_INVALID_MEM_SZ (~0U) 515 #define XEN_INVALID_NODE_DIST (~0U) 516 517 struct xen_sysctl_meminfo { 518 uint64_t memsize; 519 uint64_t memfree; 520 }; 521 typedef struct xen_sysctl_meminfo xen_sysctl_meminfo_t; 522 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_meminfo_t); 523 524 /* 525 * IN: 526 * - Both 'meminfo' and 'distance' handles being null is a request 527 * for maximum value of 'num_nodes'. 528 * - Otherwise it's the number of entries in 'meminfo' and square root 529 * of number of entries in 'distance' (when corresponding handle is 530 * non-null) 531 * 532 * OUT: 533 * - If 'num_nodes' is less than the number Xen wants to write but either 534 * handle is not a NULL one, partial data gets returned and 'num_nodes' 535 * gets updated to reflect the intended number. 536 * - Otherwise, 'num_nodes' shall indicate the number of entries written, which 537 * may be less than the input value. 538 */ 539 540 struct xen_sysctl_numainfo { 541 uint32_t num_nodes; 542 543 XEN_GUEST_HANDLE_64(xen_sysctl_meminfo_t) meminfo; 544 545 /* 546 * Distance between nodes 'i' and 'j' is stored in index 'i*N + j', 547 * where N is the number of nodes that will be returned in 'num_nodes' 548 * (i.e. not 'num_nodes' provided by the caller) 549 */ 550 XEN_GUEST_HANDLE_64(uint32) distance; 551 }; 552 553 /* XEN_SYSCTL_cpupool_op */ 554 #define XEN_SYSCTL_CPUPOOL_OP_CREATE 1 /* C */ 555 #define XEN_SYSCTL_CPUPOOL_OP_DESTROY 2 /* D */ 556 #define XEN_SYSCTL_CPUPOOL_OP_INFO 3 /* I */ 557 #define XEN_SYSCTL_CPUPOOL_OP_ADDCPU 4 /* A */ 558 #define XEN_SYSCTL_CPUPOOL_OP_RMCPU 5 /* R */ 559 #define XEN_SYSCTL_CPUPOOL_OP_MOVEDOMAIN 6 /* M */ 560 #define XEN_SYSCTL_CPUPOOL_OP_FREEINFO 7 /* F */ 561 #define XEN_SYSCTL_CPUPOOL_PAR_ANY 0xFFFFFFFF 562 struct xen_sysctl_cpupool_op { 563 uint32_t op; /* IN */ 564 uint32_t cpupool_id; /* IN: CDIARM OUT: CI */ 565 uint32_t sched_id; /* IN: C OUT: I */ 566 uint32_t domid; /* IN: M */ 567 uint32_t cpu; /* IN: AR */ 568 uint32_t n_dom; /* OUT: I */ 569 struct xenctl_bitmap cpumap; /* OUT: IF */ 570 }; 571 572 /* 573 * Error return values of cpupool operations: 574 * 575 * -EADDRINUSE: 576 * XEN_SYSCTL_CPUPOOL_OP_RMCPU: A vcpu is temporarily pinned to the cpu 577 * which is to be removed from a cpupool. 578 * -EADDRNOTAVAIL: 579 * XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: A previous 580 * request to remove a cpu from a cpupool was terminated with -EAGAIN 581 * and has not been retried using the same parameters. 582 * -EAGAIN: 583 * XEN_SYSCTL_CPUPOOL_OP_RMCPU: The cpu can't be removed from the cpupool 584 * as it is active in the hypervisor. A retry will succeed soon. 585 * -EBUSY: 586 * XEN_SYSCTL_CPUPOOL_OP_DESTROY, XEN_SYSCTL_CPUPOOL_OP_RMCPU: A cpupool 587 * can't be destroyed or the last cpu can't be removed as there is still 588 * a running domain in that cpupool. 589 * -EEXIST: 590 * XEN_SYSCTL_CPUPOOL_OP_CREATE: A cpupool_id was specified and is already 591 * existing. 592 * -EINVAL: 593 * XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: An illegal 594 * cpu was specified (cpu does not exist). 595 * XEN_SYSCTL_CPUPOOL_OP_MOVEDOMAIN: An illegal domain was specified 596 * (domain id illegal or not suitable for operation). 597 * -ENODEV: 598 * XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: The specified 599 * cpu is either not free (add) or not member of the specified cpupool 600 * (remove). 601 * -ENOENT: 602 * all: The cpupool with the specified cpupool_id doesn't exist. 603 * 604 * Some common error return values like -ENOMEM and -EFAULT are possible for 605 * all the operations. 606 */ 607 608 #define ARINC653_MAX_DOMAINS_PER_SCHEDULE 64 609 /* 610 * This structure is used to pass a new ARINC653 schedule from a 611 * privileged domain (ie dom0) to Xen. 612 */ 613 struct xen_sysctl_arinc653_schedule { 614 /* major_frame holds the time for the new schedule's major frame 615 * in nanoseconds. */ 616 uint64_aligned_t major_frame; 617 /* num_sched_entries holds how many of the entries in the 618 * sched_entries[] array are valid. */ 619 uint8_t num_sched_entries; 620 /* The sched_entries array holds the actual schedule entries. */ 621 struct { 622 /* dom_handle must match a domain's UUID */ 623 xen_domain_handle_t dom_handle; 624 /* If a domain has multiple VCPUs, vcpu_id specifies which one 625 * this schedule entry applies to. It should be set to 0 if 626 * there is only one VCPU for the domain. */ 627 unsigned int vcpu_id; 628 /* runtime specifies the amount of time that should be allocated 629 * to this VCPU per major frame. It is specified in nanoseconds */ 630 uint64_aligned_t runtime; 631 } sched_entries[ARINC653_MAX_DOMAINS_PER_SCHEDULE]; 632 }; 633 typedef struct xen_sysctl_arinc653_schedule xen_sysctl_arinc653_schedule_t; 634 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_arinc653_schedule_t); 635 636 /* 637 * Valid range for context switch rate limit (in microseconds). 638 * Applicable to Credit and Credit2 schedulers. 639 */ 640 #define XEN_SYSCTL_SCHED_RATELIMIT_MAX 500000 641 #define XEN_SYSCTL_SCHED_RATELIMIT_MIN 100 642 643 struct xen_sysctl_credit_schedule { 644 /* Length of timeslice in milliseconds */ 645 #define XEN_SYSCTL_CSCHED_TSLICE_MAX 1000 646 #define XEN_SYSCTL_CSCHED_TSLICE_MIN 1 647 unsigned tslice_ms; 648 unsigned ratelimit_us; 649 /* 650 * How long we consider a vCPU to be cache-hot on the 651 * CPU where it has run (max 100ms, in microseconds) 652 */ 653 #define XEN_SYSCTL_CSCHED_MGR_DLY_MAX_US (100 * 1000) 654 unsigned vcpu_migr_delay_us; 655 }; 656 657 struct xen_sysctl_credit2_schedule { 658 unsigned ratelimit_us; 659 }; 660 661 /* XEN_SYSCTL_scheduler_op */ 662 /* Set or get info? */ 663 #define XEN_SYSCTL_SCHEDOP_putinfo 0 664 #define XEN_SYSCTL_SCHEDOP_getinfo 1 665 struct xen_sysctl_scheduler_op { 666 uint32_t cpupool_id; /* Cpupool whose scheduler is to be targetted. */ 667 uint32_t sched_id; /* XEN_SCHEDULER_* (domctl.h) */ 668 uint32_t cmd; /* XEN_SYSCTL_SCHEDOP_* */ 669 union { 670 struct xen_sysctl_sched_arinc653 { 671 XEN_GUEST_HANDLE_64(xen_sysctl_arinc653_schedule_t) schedule; 672 } sched_arinc653; 673 struct xen_sysctl_credit_schedule sched_credit; 674 struct xen_sysctl_credit2_schedule sched_credit2; 675 } u; 676 }; 677 678 /* 679 * Output format of gcov data: 680 * 681 * XEN_GCOV_FORMAT_MAGIC XEN_GCOV_RECORD ... XEN_GCOV_RECORD 682 * 683 * That is, one magic number followed by 0 or more record. 684 * 685 * The magic number is stored as an uint32_t field. 686 * 687 * The record is packed and variable in length. It has the form: 688 * 689 * filename: a NULL terminated path name extracted from gcov, used to 690 * create the name of gcda file. 691 * size: a uint32_t field indicating the size of the payload, the 692 * unit is byte. 693 * payload: the actual payload, length is `size' bytes. 694 * 695 * Userspace tool will split the record to different files. 696 */ 697 698 #define XEN_GCOV_FORMAT_MAGIC 0x58434f56 /* XCOV */ 699 700 /* 701 * Ouput format of LLVM coverage data is just a raw stream, as would be 702 * written by the compiler_rt run time library into a .profraw file. There 703 * are no special Xen tags or delimiters because none are needed. 704 */ 705 706 #define XEN_SYSCTL_COVERAGE_get_size 0 /* Get total size of output data */ 707 #define XEN_SYSCTL_COVERAGE_read 1 /* Read output data */ 708 #define XEN_SYSCTL_COVERAGE_reset 2 /* Reset all counters */ 709 710 struct xen_sysctl_coverage_op { 711 uint32_t cmd; 712 uint32_t size; /* IN/OUT: size of the buffer */ 713 XEN_GUEST_HANDLE_64(char) buffer; /* OUT */ 714 }; 715 716 #define XEN_SYSCTL_PSR_CMT_get_total_rmid 0 717 #define XEN_SYSCTL_PSR_CMT_get_l3_upscaling_factor 1 718 /* The L3 cache size is returned in KB unit */ 719 #define XEN_SYSCTL_PSR_CMT_get_l3_cache_size 2 720 #define XEN_SYSCTL_PSR_CMT_enabled 3 721 #define XEN_SYSCTL_PSR_CMT_get_l3_event_mask 4 722 struct xen_sysctl_psr_cmt_op { 723 uint32_t cmd; /* IN: XEN_SYSCTL_PSR_CMT_* */ 724 uint32_t flags; /* padding variable, may be extended for future use */ 725 union { 726 uint64_t data; /* OUT */ 727 struct { 728 uint32_t cpu; /* IN */ 729 uint32_t rsvd; 730 } l3_cache; 731 } u; 732 }; 733 734 /* XEN_SYSCTL_pcitopoinfo */ 735 #define XEN_INVALID_DEV (XEN_INVALID_NODE_ID - 1) 736 struct xen_sysctl_pcitopoinfo { 737 /* 738 * IN: Number of elements in 'devs' and 'nodes' arrays. 739 * OUT: Number of processed elements of those arrays. 740 */ 741 uint32_t num_devs; 742 743 /* IN: list of devices for which node IDs are requested. */ 744 XEN_GUEST_HANDLE_64(physdev_pci_device_t) devs; 745 746 /* 747 * OUT: node identifier for each device. 748 * If information for a particular device is not available then 749 * corresponding entry will be set to XEN_INVALID_NODE_ID. If 750 * device is not known to the hypervisor then XEN_INVALID_DEV 751 * will be provided. 752 */ 753 XEN_GUEST_HANDLE_64(uint32) nodes; 754 }; 755 756 #define XEN_SYSCTL_PSR_get_l3_info 0 757 #define XEN_SYSCTL_PSR_get_l2_info 1 758 #define XEN_SYSCTL_PSR_get_mba_info 2 759 struct xen_sysctl_psr_alloc { 760 uint32_t cmd; /* IN: XEN_SYSCTL_PSR_* */ 761 uint32_t target; /* IN */ 762 union { 763 struct { 764 uint32_t cbm_len; /* OUT: CBM length */ 765 uint32_t cos_max; /* OUT: Maximum COS */ 766 #define XEN_SYSCTL_PSR_CAT_L3_CDP (1u << 0) 767 uint32_t flags; /* OUT: CAT flags */ 768 } cat_info; 769 770 struct { 771 uint32_t thrtl_max; /* OUT: Maximum throttle */ 772 uint32_t cos_max; /* OUT: Maximum COS */ 773 #define XEN_SYSCTL_PSR_MBA_LINEAR (1u << 0) 774 uint32_t flags; /* OUT: MBA flags */ 775 } mba_info; 776 } u; 777 }; 778 779 /* 780 * XEN_SYSCTL_get_cpu_levelling_caps (x86 specific) 781 * 782 * Return hardware capabilities concerning masking or faulting of the cpuid 783 * instruction for PV guests. 784 */ 785 struct xen_sysctl_cpu_levelling_caps { 786 #define XEN_SYSCTL_CPU_LEVELCAP_faulting (1ul << 0) /* CPUID faulting */ 787 #define XEN_SYSCTL_CPU_LEVELCAP_ecx (1ul << 1) /* 0x00000001.ecx */ 788 #define XEN_SYSCTL_CPU_LEVELCAP_edx (1ul << 2) /* 0x00000001.edx */ 789 #define XEN_SYSCTL_CPU_LEVELCAP_extd_ecx (1ul << 3) /* 0x80000001.ecx */ 790 #define XEN_SYSCTL_CPU_LEVELCAP_extd_edx (1ul << 4) /* 0x80000001.edx */ 791 #define XEN_SYSCTL_CPU_LEVELCAP_xsave_eax (1ul << 5) /* 0x0000000D:1.eax */ 792 #define XEN_SYSCTL_CPU_LEVELCAP_thermal_ecx (1ul << 6) /* 0x00000006.ecx */ 793 #define XEN_SYSCTL_CPU_LEVELCAP_l7s0_eax (1ul << 7) /* 0x00000007:0.eax */ 794 #define XEN_SYSCTL_CPU_LEVELCAP_l7s0_ebx (1ul << 8) /* 0x00000007:0.ebx */ 795 uint32_t caps; 796 }; 797 798 /* 799 * XEN_SYSCTL_get_cpu_featureset (x86 specific) 800 * 801 * Return information about featuresets available on this host. 802 * - Raw: The real cpuid values. 803 * - Host: The values Xen is using, (after command line overrides, etc). 804 * - PV: Maximum set of features which can be given to a PV guest. 805 * - HVM: Maximum set of features which can be given to a HVM guest. 806 * May fail with -EOPNOTSUPP if querying for PV or HVM data when support is 807 * compiled out of Xen. 808 */ 809 struct xen_sysctl_cpu_featureset { 810 #define XEN_SYSCTL_cpu_featureset_raw 0 811 #define XEN_SYSCTL_cpu_featureset_host 1 812 #define XEN_SYSCTL_cpu_featureset_pv 2 813 #define XEN_SYSCTL_cpu_featureset_hvm 3 814 uint32_t index; /* IN: Which featureset to query? */ 815 uint32_t nr_features; /* IN/OUT: Number of entries in/written to 816 * 'features', or the maximum number of features if 817 * the guest handle is NULL. NB. All featuresets 818 * come from the same numberspace, so have the same 819 * maximum length. */ 820 XEN_GUEST_HANDLE_64(uint32) features; /* OUT: */ 821 }; 822 823 /* 824 * XEN_SYSCTL_LIVEPATCH_op 825 * 826 * Refer to the docs/unstable/misc/livepatch.markdown 827 * for the design details of this hypercall. 828 * 829 * There are four sub-ops: 830 * XEN_SYSCTL_LIVEPATCH_UPLOAD (0) 831 * XEN_SYSCTL_LIVEPATCH_GET (1) 832 * XEN_SYSCTL_LIVEPATCH_LIST (2) 833 * XEN_SYSCTL_LIVEPATCH_ACTION (3) 834 * 835 * The normal sequence of sub-ops is to: 836 * 1) XEN_SYSCTL_LIVEPATCH_UPLOAD to upload the payload. If errors STOP. 837 * 2) XEN_SYSCTL_LIVEPATCH_GET to check the `->rc`. If -XEN_EAGAIN spin. 838 * If zero go to next step. 839 * 3) XEN_SYSCTL_LIVEPATCH_ACTION with LIVEPATCH_ACTION_APPLY to apply the patch. 840 * 4) XEN_SYSCTL_LIVEPATCH_GET to check the `->rc`. If in -XEN_EAGAIN spin. 841 * If zero exit with success. 842 */ 843 844 #define LIVEPATCH_PAYLOAD_VERSION 2 845 /* 846 * .livepatch.funcs structure layout defined in the `Payload format` 847 * section in the Live Patch design document. 848 * 849 * We guard this with __XEN__ as toolstacks SHOULD not use it. 850 */ 851 #ifdef __XEN__ 852 #define LIVEPATCH_OPAQUE_SIZE 31 853 854 struct livepatch_expectation { 855 uint8_t enabled : 1; 856 uint8_t len : 5; /* Length of data up to LIVEPATCH_OPAQUE_SIZE 857 (5 bits is enough for now) */ 858 uint8_t rsv : 2; /* Reserved. Zero value */ 859 uint8_t data[LIVEPATCH_OPAQUE_SIZE]; /* Same size as opaque[] buffer of 860 struct livepatch_func. This is the 861 max number of bytes to be patched */ 862 }; 863 typedef struct livepatch_expectation livepatch_expectation_t; 864 865 typedef enum livepatch_func_state { 866 LIVEPATCH_FUNC_NOT_APPLIED, 867 LIVEPATCH_FUNC_APPLIED 868 } livepatch_func_state_t; 869 870 struct livepatch_func { 871 const char *name; /* Name of function to be patched. */ 872 void *new_addr; 873 void *old_addr; 874 uint32_t new_size; 875 uint32_t old_size; 876 uint8_t version; /* MUST be LIVEPATCH_PAYLOAD_VERSION. */ 877 uint8_t opaque[LIVEPATCH_OPAQUE_SIZE]; 878 uint8_t applied; 879 uint8_t _pad[7]; 880 livepatch_expectation_t expect; 881 }; 882 typedef struct livepatch_func livepatch_func_t; 883 #endif 884 885 /* 886 * Structure describing an ELF payload. Uniquely identifies the 887 * payload. Should be human readable. 888 * Recommended length is upto XEN_LIVEPATCH_NAME_SIZE. 889 * Includes the NUL terminator. 890 */ 891 #define XEN_LIVEPATCH_NAME_SIZE 128 892 struct xen_livepatch_name { 893 XEN_GUEST_HANDLE_64(char) name; /* IN: pointer to name. */ 894 uint16_t size; /* IN: size of name. May be upto 895 XEN_LIVEPATCH_NAME_SIZE. */ 896 uint16_t pad[3]; /* IN: MUST be zero. */ 897 }; 898 899 /* 900 * Upload a payload to the hypervisor. The payload is verified 901 * against basic checks and if there are any issues the proper return code 902 * will be returned. The payload is not applied at this time - that is 903 * controlled by XEN_SYSCTL_LIVEPATCH_ACTION. 904 * 905 * The return value is zero if the payload was succesfully uploaded. 906 * Otherwise an EXX return value is provided. Duplicate `name` are not 907 * supported. 908 * 909 * The payload at this point is verified against basic checks. 910 * 911 * The `payload` is the ELF payload as mentioned in the `Payload format` 912 * section in the Live Patch design document. 913 */ 914 #define XEN_SYSCTL_LIVEPATCH_UPLOAD 0 915 struct xen_sysctl_livepatch_upload { 916 struct xen_livepatch_name name; /* IN, name of the patch. */ 917 uint64_t size; /* IN, size of the ELF file. */ 918 XEN_GUEST_HANDLE_64(uint8) payload; /* IN, the ELF file. */ 919 }; 920 921 /* 922 * Retrieve an status of an specific payload. 923 * 924 * Upon completion the `struct xen_livepatch_status` is updated. 925 * 926 * The return value is zero on success and XEN_EXX on failure. This operation 927 * is synchronous and does not require preemption. 928 */ 929 #define XEN_SYSCTL_LIVEPATCH_GET 1 930 931 struct xen_livepatch_status { 932 #define LIVEPATCH_STATE_CHECKED 1 933 #define LIVEPATCH_STATE_APPLIED 2 934 uint32_t state; /* OUT: LIVEPATCH_STATE_*. */ 935 int32_t rc; /* OUT: 0 if no error, otherwise -XEN_EXX. */ 936 }; 937 typedef struct xen_livepatch_status xen_livepatch_status_t; 938 DEFINE_XEN_GUEST_HANDLE(xen_livepatch_status_t); 939 940 struct xen_sysctl_livepatch_get { 941 struct xen_livepatch_name name; /* IN, name of the payload. */ 942 struct xen_livepatch_status status; /* IN/OUT, state of it. */ 943 }; 944 945 /* 946 * Retrieve an array of abbreviated status, names and metadata of payloads that 947 * are loaded in the hypervisor. 948 * 949 * If the hypercall returns an positive number, it is the number (up to `nr`) 950 * of the payloads returned, along with `nr` updated with the number of remaining 951 * payloads, `version` updated (it may be the same across hypercalls. If it varies 952 * the data is stale and further calls could fail), `name_total_size` and 953 * `metadata_total_size` containing total sizes of transferred data for both the 954 * arrays. 955 * The `status`, `name`, `len`, `metadata` and `metadata_len` are updated at their 956 * designed index value (`idx`) with the returned value of data. 957 * 958 * If the hypercall returns E2BIG the `nr` is too big and should be 959 * lowered. The upper limit of `nr` is left to the implemention. 960 * 961 * Note that due to the asynchronous nature of hypercalls the domain might have 962 * added or removed the number of payloads making this information stale. It is 963 * the responsibility of the toolstack to use the `version` field to check 964 * between each invocation. if the version differs it should discard the stale 965 * data and start from scratch. It is OK for the toolstack to use the new 966 * `version` field. 967 */ 968 #define XEN_SYSCTL_LIVEPATCH_LIST 2 969 struct xen_sysctl_livepatch_list { 970 uint32_t version; /* OUT: Hypervisor stamps value. 971 If varies between calls, we are 972 * getting stale data. */ 973 uint32_t idx; /* IN: Index into hypervisor list. */ 974 uint32_t nr; /* IN: How many status, name, and len 975 should fill out. Can be zero to get 976 amount of payloads and version. 977 OUT: How many payloads left. */ 978 uint32_t pad; /* IN: Must be zero. */ 979 uint32_t name_total_size; /* OUT: Total size of all transfer names */ 980 uint32_t metadata_total_size; /* OUT: Total size of all transfer metadata */ 981 XEN_GUEST_HANDLE_64(xen_livepatch_status_t) status; /* OUT. Must have enough 982 space allocate for nr of them. */ 983 XEN_GUEST_HANDLE_64(char) name; /* OUT: Array of names. Each member 984 may have an arbitrary length up to 985 XEN_LIVEPATCH_NAME_SIZE bytes. Must have 986 nr of them. */ 987 XEN_GUEST_HANDLE_64(uint32) len; /* OUT: Array of lengths of name's. 988 Must have nr of them. */ 989 XEN_GUEST_HANDLE_64(char) metadata; /* OUT: Array of metadata strings. Each 990 member may have an arbitrary length. 991 Must have nr of them. */ 992 XEN_GUEST_HANDLE_64(uint32) metadata_len; /* OUT: Array of lengths of metadata's. 993 Must have nr of them. */ 994 }; 995 996 /* 997 * Perform an operation on the payload structure referenced by the `name` field. 998 * The operation request is asynchronous and the status should be retrieved 999 * by using either XEN_SYSCTL_LIVEPATCH_GET or XEN_SYSCTL_LIVEPATCH_LIST hypercall. 1000 */ 1001 #define XEN_SYSCTL_LIVEPATCH_ACTION 3 1002 struct xen_sysctl_livepatch_action { 1003 struct xen_livepatch_name name; /* IN, name of the patch. */ 1004 #define LIVEPATCH_ACTION_UNLOAD 1 1005 #define LIVEPATCH_ACTION_REVERT 2 1006 #define LIVEPATCH_ACTION_APPLY 3 1007 #define LIVEPATCH_ACTION_REPLACE 4 1008 uint32_t cmd; /* IN: LIVEPATCH_ACTION_*. */ 1009 uint32_t timeout; /* IN: If zero then uses */ 1010 /* hypervisor default. */ 1011 /* Or upper bound of time (ns) */ 1012 /* for operation to take. */ 1013 1014 /* 1015 * Override default inter-module buildid dependency chain enforcement. 1016 * Check only if module is built for given hypervisor by comparing buildid. 1017 */ 1018 #define LIVEPATCH_ACTION_APPLY_NODEPS (1 << 0) 1019 uint32_t flags; /* IN: action flags. */ 1020 /* Provide additional parameters */ 1021 /* for an action. */ 1022 uint32_t pad; /* IN: Always zero. */ 1023 }; 1024 1025 struct xen_sysctl_livepatch_op { 1026 uint32_t cmd; /* IN: XEN_SYSCTL_LIVEPATCH_*. */ 1027 uint32_t pad; /* IN: Always zero. */ 1028 union { 1029 struct xen_sysctl_livepatch_upload upload; 1030 struct xen_sysctl_livepatch_list list; 1031 struct xen_sysctl_livepatch_get get; 1032 struct xen_sysctl_livepatch_action action; 1033 } u; 1034 }; 1035 1036 #if defined(__i386__) || defined(__x86_64__) 1037 /* 1038 * XEN_SYSCTL_get_cpu_policy (x86 specific) 1039 * 1040 * Return information about CPUID and MSR policies available on this host. 1041 * - Raw: The real H/W values. 1042 * - Host: The values Xen is using, (after command line overrides, etc). 1043 * - Max_*: Maximum set of features a PV or HVM guest can use. Includes 1044 * experimental features outside of security support. 1045 * - Default_*: Default set of features a PV or HVM guest can use. This is 1046 * the security supported set. 1047 * May fail with -EOPNOTSUPP if querying for PV or HVM data when support is 1048 * compiled out of Xen. 1049 */ 1050 struct xen_sysctl_cpu_policy { 1051 #define XEN_SYSCTL_cpu_policy_raw 0 1052 #define XEN_SYSCTL_cpu_policy_host 1 1053 #define XEN_SYSCTL_cpu_policy_pv_max 2 1054 #define XEN_SYSCTL_cpu_policy_hvm_max 3 1055 #define XEN_SYSCTL_cpu_policy_pv_default 4 1056 #define XEN_SYSCTL_cpu_policy_hvm_default 5 1057 uint32_t index; /* IN: Which policy to query? */ 1058 uint32_t nr_leaves; /* IN/OUT: Number of leaves in/written to 1059 * 'cpuid_policy', or the maximum number of leaves 1060 * if the guest handle is NULL. */ 1061 uint32_t nr_msrs; /* IN/OUT: Number of MSRs in/written to 1062 * 'msr_policy', or the maximum number of MSRs if 1063 * the guest handle is NULL. */ 1064 uint32_t _rsvd; /* Must be zero. */ 1065 XEN_GUEST_HANDLE_64(xen_cpuid_leaf_t) cpuid_policy; /* OUT */ 1066 XEN_GUEST_HANDLE_64(xen_msr_entry_t) msr_policy; /* OUT */ 1067 }; 1068 typedef struct xen_sysctl_cpu_policy xen_sysctl_cpu_policy_t; 1069 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cpu_policy_t); 1070 #endif 1071 1072 struct xen_sysctl { 1073 uint32_t cmd; 1074 #define XEN_SYSCTL_readconsole 1 1075 #define XEN_SYSCTL_tbuf_op 2 1076 #define XEN_SYSCTL_physinfo 3 1077 #define XEN_SYSCTL_sched_id 4 1078 #define XEN_SYSCTL_perfc_op 5 1079 #define XEN_SYSCTL_getdomaininfolist 6 1080 #define XEN_SYSCTL_debug_keys 7 1081 #define XEN_SYSCTL_getcpuinfo 8 1082 #define XEN_SYSCTL_availheap 9 1083 #define XEN_SYSCTL_get_pmstat 10 1084 #define XEN_SYSCTL_cpu_hotplug 11 1085 #define XEN_SYSCTL_pm_op 12 1086 #define XEN_SYSCTL_page_offline_op 14 1087 #define XEN_SYSCTL_lockprof_op 15 1088 #define XEN_SYSCTL_cputopoinfo 16 1089 #define XEN_SYSCTL_numainfo 17 1090 #define XEN_SYSCTL_cpupool_op 18 1091 #define XEN_SYSCTL_scheduler_op 19 1092 #define XEN_SYSCTL_coverage_op 20 1093 #define XEN_SYSCTL_psr_cmt_op 21 1094 #define XEN_SYSCTL_pcitopoinfo 22 1095 #define XEN_SYSCTL_psr_alloc 23 1096 /* #define XEN_SYSCTL_tmem_op 24 */ 1097 #define XEN_SYSCTL_get_cpu_levelling_caps 25 1098 #define XEN_SYSCTL_get_cpu_featureset 26 1099 #define XEN_SYSCTL_livepatch_op 27 1100 /* #define XEN_SYSCTL_set_parameter 28 */ 1101 #define XEN_SYSCTL_get_cpu_policy 29 1102 uint32_t interface_version; /* XEN_SYSCTL_INTERFACE_VERSION */ 1103 union { 1104 struct xen_sysctl_readconsole readconsole; 1105 struct xen_sysctl_tbuf_op tbuf_op; 1106 struct xen_sysctl_physinfo physinfo; 1107 struct xen_sysctl_cputopoinfo cputopoinfo; 1108 struct xen_sysctl_pcitopoinfo pcitopoinfo; 1109 struct xen_sysctl_numainfo numainfo; 1110 struct xen_sysctl_sched_id sched_id; 1111 struct xen_sysctl_perfc_op perfc_op; 1112 struct xen_sysctl_getdomaininfolist getdomaininfolist; 1113 struct xen_sysctl_debug_keys debug_keys; 1114 struct xen_sysctl_getcpuinfo getcpuinfo; 1115 struct xen_sysctl_availheap availheap; 1116 struct xen_sysctl_get_pmstat get_pmstat; 1117 struct xen_sysctl_cpu_hotplug cpu_hotplug; 1118 struct xen_sysctl_pm_op pm_op; 1119 struct xen_sysctl_page_offline_op page_offline; 1120 struct xen_sysctl_lockprof_op lockprof_op; 1121 struct xen_sysctl_cpupool_op cpupool_op; 1122 struct xen_sysctl_scheduler_op scheduler_op; 1123 struct xen_sysctl_coverage_op coverage_op; 1124 struct xen_sysctl_psr_cmt_op psr_cmt_op; 1125 struct xen_sysctl_psr_alloc psr_alloc; 1126 struct xen_sysctl_cpu_levelling_caps cpu_levelling_caps; 1127 struct xen_sysctl_cpu_featureset cpu_featureset; 1128 struct xen_sysctl_livepatch_op livepatch; 1129 #if defined(__i386__) || defined(__x86_64__) 1130 struct xen_sysctl_cpu_policy cpu_policy; 1131 #endif 1132 uint8_t pad[128]; 1133 } u; 1134 }; 1135 typedef struct xen_sysctl xen_sysctl_t; 1136 DEFINE_XEN_GUEST_HANDLE(xen_sysctl_t); 1137 1138 #endif /* __XEN_PUBLIC_SYSCTL_H__ */ 1139 1140 /* 1141 * Local variables: 1142 * mode: C 1143 * c-file-style: "BSD" 1144 * c-basic-offset: 4 1145 * tab-width: 4 1146 * indent-tabs-mode: nil 1147 * End: 1148 */ 1149