Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/arch/x86/kvm/vmx/vmcs12.h
26489 views
1
/* SPDX-License-Identifier: GPL-2.0 */
2
#ifndef __KVM_X86_VMX_VMCS12_H
3
#define __KVM_X86_VMX_VMCS12_H
4
5
#include <linux/build_bug.h>
6
7
#include "vmcs.h"
8
9
/*
10
* struct vmcs12 describes the state that our guest hypervisor (L1) keeps for a
11
* single nested guest (L2), hence the name vmcs12. Any VMX implementation has
12
* a VMCS structure, and vmcs12 is our emulated VMX's VMCS. This structure is
13
* stored in guest memory specified by VMPTRLD, but is opaque to the guest,
14
* which must access it using VMREAD/VMWRITE/VMCLEAR instructions.
15
* More than one of these structures may exist, if L1 runs multiple L2 guests.
16
* nested_vmx_run() will use the data here to build the vmcs02: a VMCS for the
17
* underlying hardware which will be used to run L2.
18
* This structure is packed to ensure that its layout is identical across
19
* machines (necessary for live migration).
20
*
21
* IMPORTANT: Changing the layout of existing fields in this structure
22
* will break save/restore compatibility with older kvm releases. When
23
* adding new fields, either use space in the reserved padding* arrays
24
* or add the new fields to the end of the structure.
25
*/
26
typedef u64 natural_width;
27
struct __packed vmcs12 {
28
/* According to the Intel spec, a VMCS region must start with the
29
* following two fields. Then follow implementation-specific data.
30
*/
31
struct vmcs_hdr hdr;
32
u32 abort;
33
34
u32 launch_state; /* set to 0 by VMCLEAR, to 1 by VMLAUNCH */
35
u32 padding[7]; /* room for future expansion */
36
37
u64 io_bitmap_a;
38
u64 io_bitmap_b;
39
u64 msr_bitmap;
40
u64 vm_exit_msr_store_addr;
41
u64 vm_exit_msr_load_addr;
42
u64 vm_entry_msr_load_addr;
43
u64 tsc_offset;
44
u64 virtual_apic_page_addr;
45
u64 apic_access_addr;
46
u64 posted_intr_desc_addr;
47
u64 ept_pointer;
48
u64 eoi_exit_bitmap0;
49
u64 eoi_exit_bitmap1;
50
u64 eoi_exit_bitmap2;
51
u64 eoi_exit_bitmap3;
52
u64 xss_exit_bitmap;
53
u64 guest_physical_address;
54
u64 vmcs_link_pointer;
55
u64 guest_ia32_debugctl;
56
u64 guest_ia32_pat;
57
u64 guest_ia32_efer;
58
u64 guest_ia32_perf_global_ctrl;
59
u64 guest_pdptr0;
60
u64 guest_pdptr1;
61
u64 guest_pdptr2;
62
u64 guest_pdptr3;
63
u64 guest_bndcfgs;
64
u64 host_ia32_pat;
65
u64 host_ia32_efer;
66
u64 host_ia32_perf_global_ctrl;
67
u64 vmread_bitmap;
68
u64 vmwrite_bitmap;
69
u64 vm_function_control;
70
u64 eptp_list_address;
71
u64 pml_address;
72
u64 encls_exiting_bitmap;
73
u64 tsc_multiplier;
74
u64 padding64[1]; /* room for future expansion */
75
/*
76
* To allow migration of L1 (complete with its L2 guests) between
77
* machines of different natural widths (32 or 64 bit), we cannot have
78
* unsigned long fields with no explicit size. We use u64 (aliased
79
* natural_width) instead. Luckily, x86 is little-endian.
80
*/
81
natural_width cr0_guest_host_mask;
82
natural_width cr4_guest_host_mask;
83
natural_width cr0_read_shadow;
84
natural_width cr4_read_shadow;
85
natural_width dead_space[4]; /* Last remnants of cr3_target_value[0-3]. */
86
natural_width exit_qualification;
87
natural_width guest_linear_address;
88
natural_width guest_cr0;
89
natural_width guest_cr3;
90
natural_width guest_cr4;
91
natural_width guest_es_base;
92
natural_width guest_cs_base;
93
natural_width guest_ss_base;
94
natural_width guest_ds_base;
95
natural_width guest_fs_base;
96
natural_width guest_gs_base;
97
natural_width guest_ldtr_base;
98
natural_width guest_tr_base;
99
natural_width guest_gdtr_base;
100
natural_width guest_idtr_base;
101
natural_width guest_dr7;
102
natural_width guest_rsp;
103
natural_width guest_rip;
104
natural_width guest_rflags;
105
natural_width guest_pending_dbg_exceptions;
106
natural_width guest_sysenter_esp;
107
natural_width guest_sysenter_eip;
108
natural_width host_cr0;
109
natural_width host_cr3;
110
natural_width host_cr4;
111
natural_width host_fs_base;
112
natural_width host_gs_base;
113
natural_width host_tr_base;
114
natural_width host_gdtr_base;
115
natural_width host_idtr_base;
116
natural_width host_ia32_sysenter_esp;
117
natural_width host_ia32_sysenter_eip;
118
natural_width host_rsp;
119
natural_width host_rip;
120
natural_width paddingl[8]; /* room for future expansion */
121
u32 pin_based_vm_exec_control;
122
u32 cpu_based_vm_exec_control;
123
u32 exception_bitmap;
124
u32 page_fault_error_code_mask;
125
u32 page_fault_error_code_match;
126
u32 cr3_target_count;
127
u32 vm_exit_controls;
128
u32 vm_exit_msr_store_count;
129
u32 vm_exit_msr_load_count;
130
u32 vm_entry_controls;
131
u32 vm_entry_msr_load_count;
132
u32 vm_entry_intr_info_field;
133
u32 vm_entry_exception_error_code;
134
u32 vm_entry_instruction_len;
135
u32 tpr_threshold;
136
u32 secondary_vm_exec_control;
137
u32 vm_instruction_error;
138
u32 vm_exit_reason;
139
u32 vm_exit_intr_info;
140
u32 vm_exit_intr_error_code;
141
u32 idt_vectoring_info_field;
142
u32 idt_vectoring_error_code;
143
u32 vm_exit_instruction_len;
144
u32 vmx_instruction_info;
145
u32 guest_es_limit;
146
u32 guest_cs_limit;
147
u32 guest_ss_limit;
148
u32 guest_ds_limit;
149
u32 guest_fs_limit;
150
u32 guest_gs_limit;
151
u32 guest_ldtr_limit;
152
u32 guest_tr_limit;
153
u32 guest_gdtr_limit;
154
u32 guest_idtr_limit;
155
u32 guest_es_ar_bytes;
156
u32 guest_cs_ar_bytes;
157
u32 guest_ss_ar_bytes;
158
u32 guest_ds_ar_bytes;
159
u32 guest_fs_ar_bytes;
160
u32 guest_gs_ar_bytes;
161
u32 guest_ldtr_ar_bytes;
162
u32 guest_tr_ar_bytes;
163
u32 guest_interruptibility_info;
164
u32 guest_activity_state;
165
u32 guest_sysenter_cs;
166
u32 host_ia32_sysenter_cs;
167
u32 vmx_preemption_timer_value;
168
u32 padding32[7]; /* room for future expansion */
169
u16 virtual_processor_id;
170
u16 posted_intr_nv;
171
u16 guest_es_selector;
172
u16 guest_cs_selector;
173
u16 guest_ss_selector;
174
u16 guest_ds_selector;
175
u16 guest_fs_selector;
176
u16 guest_gs_selector;
177
u16 guest_ldtr_selector;
178
u16 guest_tr_selector;
179
u16 guest_intr_status;
180
u16 host_es_selector;
181
u16 host_cs_selector;
182
u16 host_ss_selector;
183
u16 host_ds_selector;
184
u16 host_fs_selector;
185
u16 host_gs_selector;
186
u16 host_tr_selector;
187
u16 guest_pml_index;
188
};
189
190
/*
191
* VMCS12_REVISION is KVM's arbitrary ID for the layout of struct vmcs12. KVM
192
* enumerates this value to L1 via MSR_IA32_VMX_BASIC, and checks the revision
193
* ID during nested VMPTRLD to verify that L1 is loading a VMCS that adhere's
194
* to KVM's virtual CPU definition.
195
*
196
* DO NOT change this value, as it will break save/restore compatibility with
197
* older KVM releases.
198
*/
199
#define VMCS12_REVISION 0x11e57ed0
200
201
/*
202
* VMCS12_SIZE is the number of bytes L1 should allocate for the VMXON region
203
* and any VMCS region. Although only sizeof(struct vmcs12) are used by the
204
* current implementation, 4K are reserved to avoid future complications and
205
* to preserve userspace ABI.
206
*/
207
#define VMCS12_SIZE KVM_STATE_NESTED_VMX_VMCS_SIZE
208
209
/*
210
* For save/restore compatibility, the vmcs12 field offsets must not change,
211
* although appending fields and/or filling gaps is obviously allowed.
212
*/
213
#define CHECK_OFFSET(field, loc) \
214
ASSERT_STRUCT_OFFSET(struct vmcs12, field, loc)
215
216
static inline void vmx_check_vmcs12_offsets(void)
217
{
218
CHECK_OFFSET(hdr, 0);
219
CHECK_OFFSET(abort, 4);
220
CHECK_OFFSET(launch_state, 8);
221
CHECK_OFFSET(io_bitmap_a, 40);
222
CHECK_OFFSET(io_bitmap_b, 48);
223
CHECK_OFFSET(msr_bitmap, 56);
224
CHECK_OFFSET(vm_exit_msr_store_addr, 64);
225
CHECK_OFFSET(vm_exit_msr_load_addr, 72);
226
CHECK_OFFSET(vm_entry_msr_load_addr, 80);
227
CHECK_OFFSET(tsc_offset, 88);
228
CHECK_OFFSET(virtual_apic_page_addr, 96);
229
CHECK_OFFSET(apic_access_addr, 104);
230
CHECK_OFFSET(posted_intr_desc_addr, 112);
231
CHECK_OFFSET(ept_pointer, 120);
232
CHECK_OFFSET(eoi_exit_bitmap0, 128);
233
CHECK_OFFSET(eoi_exit_bitmap1, 136);
234
CHECK_OFFSET(eoi_exit_bitmap2, 144);
235
CHECK_OFFSET(eoi_exit_bitmap3, 152);
236
CHECK_OFFSET(xss_exit_bitmap, 160);
237
CHECK_OFFSET(guest_physical_address, 168);
238
CHECK_OFFSET(vmcs_link_pointer, 176);
239
CHECK_OFFSET(guest_ia32_debugctl, 184);
240
CHECK_OFFSET(guest_ia32_pat, 192);
241
CHECK_OFFSET(guest_ia32_efer, 200);
242
CHECK_OFFSET(guest_ia32_perf_global_ctrl, 208);
243
CHECK_OFFSET(guest_pdptr0, 216);
244
CHECK_OFFSET(guest_pdptr1, 224);
245
CHECK_OFFSET(guest_pdptr2, 232);
246
CHECK_OFFSET(guest_pdptr3, 240);
247
CHECK_OFFSET(guest_bndcfgs, 248);
248
CHECK_OFFSET(host_ia32_pat, 256);
249
CHECK_OFFSET(host_ia32_efer, 264);
250
CHECK_OFFSET(host_ia32_perf_global_ctrl, 272);
251
CHECK_OFFSET(vmread_bitmap, 280);
252
CHECK_OFFSET(vmwrite_bitmap, 288);
253
CHECK_OFFSET(vm_function_control, 296);
254
CHECK_OFFSET(eptp_list_address, 304);
255
CHECK_OFFSET(pml_address, 312);
256
CHECK_OFFSET(encls_exiting_bitmap, 320);
257
CHECK_OFFSET(tsc_multiplier, 328);
258
CHECK_OFFSET(cr0_guest_host_mask, 344);
259
CHECK_OFFSET(cr4_guest_host_mask, 352);
260
CHECK_OFFSET(cr0_read_shadow, 360);
261
CHECK_OFFSET(cr4_read_shadow, 368);
262
CHECK_OFFSET(dead_space, 376);
263
CHECK_OFFSET(exit_qualification, 408);
264
CHECK_OFFSET(guest_linear_address, 416);
265
CHECK_OFFSET(guest_cr0, 424);
266
CHECK_OFFSET(guest_cr3, 432);
267
CHECK_OFFSET(guest_cr4, 440);
268
CHECK_OFFSET(guest_es_base, 448);
269
CHECK_OFFSET(guest_cs_base, 456);
270
CHECK_OFFSET(guest_ss_base, 464);
271
CHECK_OFFSET(guest_ds_base, 472);
272
CHECK_OFFSET(guest_fs_base, 480);
273
CHECK_OFFSET(guest_gs_base, 488);
274
CHECK_OFFSET(guest_ldtr_base, 496);
275
CHECK_OFFSET(guest_tr_base, 504);
276
CHECK_OFFSET(guest_gdtr_base, 512);
277
CHECK_OFFSET(guest_idtr_base, 520);
278
CHECK_OFFSET(guest_dr7, 528);
279
CHECK_OFFSET(guest_rsp, 536);
280
CHECK_OFFSET(guest_rip, 544);
281
CHECK_OFFSET(guest_rflags, 552);
282
CHECK_OFFSET(guest_pending_dbg_exceptions, 560);
283
CHECK_OFFSET(guest_sysenter_esp, 568);
284
CHECK_OFFSET(guest_sysenter_eip, 576);
285
CHECK_OFFSET(host_cr0, 584);
286
CHECK_OFFSET(host_cr3, 592);
287
CHECK_OFFSET(host_cr4, 600);
288
CHECK_OFFSET(host_fs_base, 608);
289
CHECK_OFFSET(host_gs_base, 616);
290
CHECK_OFFSET(host_tr_base, 624);
291
CHECK_OFFSET(host_gdtr_base, 632);
292
CHECK_OFFSET(host_idtr_base, 640);
293
CHECK_OFFSET(host_ia32_sysenter_esp, 648);
294
CHECK_OFFSET(host_ia32_sysenter_eip, 656);
295
CHECK_OFFSET(host_rsp, 664);
296
CHECK_OFFSET(host_rip, 672);
297
CHECK_OFFSET(pin_based_vm_exec_control, 744);
298
CHECK_OFFSET(cpu_based_vm_exec_control, 748);
299
CHECK_OFFSET(exception_bitmap, 752);
300
CHECK_OFFSET(page_fault_error_code_mask, 756);
301
CHECK_OFFSET(page_fault_error_code_match, 760);
302
CHECK_OFFSET(cr3_target_count, 764);
303
CHECK_OFFSET(vm_exit_controls, 768);
304
CHECK_OFFSET(vm_exit_msr_store_count, 772);
305
CHECK_OFFSET(vm_exit_msr_load_count, 776);
306
CHECK_OFFSET(vm_entry_controls, 780);
307
CHECK_OFFSET(vm_entry_msr_load_count, 784);
308
CHECK_OFFSET(vm_entry_intr_info_field, 788);
309
CHECK_OFFSET(vm_entry_exception_error_code, 792);
310
CHECK_OFFSET(vm_entry_instruction_len, 796);
311
CHECK_OFFSET(tpr_threshold, 800);
312
CHECK_OFFSET(secondary_vm_exec_control, 804);
313
CHECK_OFFSET(vm_instruction_error, 808);
314
CHECK_OFFSET(vm_exit_reason, 812);
315
CHECK_OFFSET(vm_exit_intr_info, 816);
316
CHECK_OFFSET(vm_exit_intr_error_code, 820);
317
CHECK_OFFSET(idt_vectoring_info_field, 824);
318
CHECK_OFFSET(idt_vectoring_error_code, 828);
319
CHECK_OFFSET(vm_exit_instruction_len, 832);
320
CHECK_OFFSET(vmx_instruction_info, 836);
321
CHECK_OFFSET(guest_es_limit, 840);
322
CHECK_OFFSET(guest_cs_limit, 844);
323
CHECK_OFFSET(guest_ss_limit, 848);
324
CHECK_OFFSET(guest_ds_limit, 852);
325
CHECK_OFFSET(guest_fs_limit, 856);
326
CHECK_OFFSET(guest_gs_limit, 860);
327
CHECK_OFFSET(guest_ldtr_limit, 864);
328
CHECK_OFFSET(guest_tr_limit, 868);
329
CHECK_OFFSET(guest_gdtr_limit, 872);
330
CHECK_OFFSET(guest_idtr_limit, 876);
331
CHECK_OFFSET(guest_es_ar_bytes, 880);
332
CHECK_OFFSET(guest_cs_ar_bytes, 884);
333
CHECK_OFFSET(guest_ss_ar_bytes, 888);
334
CHECK_OFFSET(guest_ds_ar_bytes, 892);
335
CHECK_OFFSET(guest_fs_ar_bytes, 896);
336
CHECK_OFFSET(guest_gs_ar_bytes, 900);
337
CHECK_OFFSET(guest_ldtr_ar_bytes, 904);
338
CHECK_OFFSET(guest_tr_ar_bytes, 908);
339
CHECK_OFFSET(guest_interruptibility_info, 912);
340
CHECK_OFFSET(guest_activity_state, 916);
341
CHECK_OFFSET(guest_sysenter_cs, 920);
342
CHECK_OFFSET(host_ia32_sysenter_cs, 924);
343
CHECK_OFFSET(vmx_preemption_timer_value, 928);
344
CHECK_OFFSET(virtual_processor_id, 960);
345
CHECK_OFFSET(posted_intr_nv, 962);
346
CHECK_OFFSET(guest_es_selector, 964);
347
CHECK_OFFSET(guest_cs_selector, 966);
348
CHECK_OFFSET(guest_ss_selector, 968);
349
CHECK_OFFSET(guest_ds_selector, 970);
350
CHECK_OFFSET(guest_fs_selector, 972);
351
CHECK_OFFSET(guest_gs_selector, 974);
352
CHECK_OFFSET(guest_ldtr_selector, 976);
353
CHECK_OFFSET(guest_tr_selector, 978);
354
CHECK_OFFSET(guest_intr_status, 980);
355
CHECK_OFFSET(host_es_selector, 982);
356
CHECK_OFFSET(host_cs_selector, 984);
357
CHECK_OFFSET(host_ss_selector, 986);
358
CHECK_OFFSET(host_ds_selector, 988);
359
CHECK_OFFSET(host_fs_selector, 990);
360
CHECK_OFFSET(host_gs_selector, 992);
361
CHECK_OFFSET(host_tr_selector, 994);
362
CHECK_OFFSET(guest_pml_index, 996);
363
}
364
365
extern const unsigned short vmcs12_field_offsets[];
366
extern const unsigned int nr_vmcs12_fields;
367
368
static inline short get_vmcs12_field_offset(unsigned long field)
369
{
370
unsigned short offset;
371
unsigned int index;
372
373
if (field >> 15)
374
return -ENOENT;
375
376
index = ROL16(field, 6);
377
if (index >= nr_vmcs12_fields)
378
return -ENOENT;
379
380
index = array_index_nospec(index, nr_vmcs12_fields);
381
offset = vmcs12_field_offsets[index];
382
if (offset == 0)
383
return -ENOENT;
384
return offset;
385
}
386
387
static inline u64 vmcs12_read_any(struct vmcs12 *vmcs12, unsigned long field,
388
u16 offset)
389
{
390
char *p = (char *)vmcs12 + offset;
391
392
switch (vmcs_field_width(field)) {
393
case VMCS_FIELD_WIDTH_NATURAL_WIDTH:
394
return *((natural_width *)p);
395
case VMCS_FIELD_WIDTH_U16:
396
return *((u16 *)p);
397
case VMCS_FIELD_WIDTH_U32:
398
return *((u32 *)p);
399
case VMCS_FIELD_WIDTH_U64:
400
return *((u64 *)p);
401
default:
402
WARN_ON_ONCE(1);
403
return -1;
404
}
405
}
406
407
static inline void vmcs12_write_any(struct vmcs12 *vmcs12, unsigned long field,
408
u16 offset, u64 field_value)
409
{
410
char *p = (char *)vmcs12 + offset;
411
412
switch (vmcs_field_width(field)) {
413
case VMCS_FIELD_WIDTH_U16:
414
*(u16 *)p = field_value;
415
break;
416
case VMCS_FIELD_WIDTH_U32:
417
*(u32 *)p = field_value;
418
break;
419
case VMCS_FIELD_WIDTH_U64:
420
*(u64 *)p = field_value;
421
break;
422
case VMCS_FIELD_WIDTH_NATURAL_WIDTH:
423
*(natural_width *)p = field_value;
424
break;
425
default:
426
WARN_ON_ONCE(1);
427
break;
428
}
429
}
430
431
#endif /* __KVM_X86_VMX_VMCS12_H */
432
433