Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/arch/x86/hyperv/hv_vtl.c
50905 views
1
// SPDX-License-Identifier: GPL-2.0
2
/*
3
* Copyright (c) 2023, Microsoft Corporation.
4
*
5
* Author:
6
* Saurabh Sengar <[email protected]>
7
*/
8
9
#include <asm/apic.h>
10
#include <asm/boot.h>
11
#include <asm/desc.h>
12
#include <asm/fpu/api.h>
13
#include <asm/fpu/types.h>
14
#include <asm/i8259.h>
15
#include <asm/mshyperv.h>
16
#include <asm/msr.h>
17
#include <asm/realmode.h>
18
#include <asm/reboot.h>
19
#include <asm/smap.h>
20
#include <linux/export.h>
21
#include <../kernel/smpboot.h>
22
#include "../../kernel/fpu/legacy.h"
23
24
extern struct boot_params boot_params;
25
static struct real_mode_header hv_vtl_real_mode_header;
26
27
static bool __init hv_vtl_msi_ext_dest_id(void)
28
{
29
return true;
30
}
31
32
/*
33
* The `native_machine_emergency_restart` function from `reboot.c` writes
34
* to the physical address 0x472 to indicate the type of reboot for the
35
* firmware. We cannot have that in VSM as the memory composition might
36
* be more generic, and such write effectively corrupts the memory thus
37
* making diagnostics harder at the very least.
38
*/
39
static void __noreturn hv_vtl_emergency_restart(void)
40
{
41
/*
42
* Cause a triple fault and the immediate reset. Here the code does not run
43
* on the top of any firmware, whereby cannot reach out to its services.
44
* The inifinite loop is for the improbable case that the triple fault does
45
* not work and have to preserve the state intact for debugging.
46
*/
47
for (;;) {
48
idt_invalidate();
49
__asm__ __volatile__("int3");
50
}
51
}
52
53
/*
54
* The only way to restart in the VTL mode is to triple fault as the kernel runs
55
* as firmware.
56
*/
57
static void __noreturn hv_vtl_restart(char __maybe_unused *cmd)
58
{
59
hv_vtl_emergency_restart();
60
}
61
62
void __init hv_vtl_init_platform(void)
63
{
64
/*
65
* This function is a no-op if the VTL mode is not enabled.
66
* If it is, this function runs if and only the kernel boots in
67
* VTL2 which the x86 hv initialization path makes sure of.
68
*/
69
pr_info("Linux runs in Hyper-V Virtual Trust Level %d\n", ms_hyperv.vtl);
70
71
x86_platform.realmode_reserve = x86_init_noop;
72
x86_platform.realmode_init = x86_init_noop;
73
x86_init.irqs.pre_vector_init = x86_init_noop;
74
x86_init.timers.timer_init = x86_init_noop;
75
x86_init.resources.probe_roms = x86_init_noop;
76
77
/* Avoid searching for BIOS MP tables */
78
x86_init.mpparse.find_mptable = x86_init_noop;
79
x86_init.mpparse.early_parse_smp_cfg = x86_init_noop;
80
81
x86_platform.get_wallclock = get_rtc_noop;
82
x86_platform.set_wallclock = set_rtc_noop;
83
x86_platform.get_nmi_reason = hv_get_nmi_reason;
84
85
x86_platform.legacy.i8042 = X86_LEGACY_I8042_PLATFORM_ABSENT;
86
x86_platform.legacy.rtc = 0;
87
x86_platform.legacy.warm_reset = 0;
88
x86_platform.legacy.reserve_bios_regions = 0;
89
x86_platform.legacy.devices.pnpbios = 0;
90
91
x86_init.hyper.msi_ext_dest_id = hv_vtl_msi_ext_dest_id;
92
}
93
94
static inline u64 hv_vtl_system_desc_base(struct ldttss_desc *desc)
95
{
96
return ((u64)desc->base3 << 32) | ((u64)desc->base2 << 24) |
97
(desc->base1 << 16) | desc->base0;
98
}
99
100
static inline u32 hv_vtl_system_desc_limit(struct ldttss_desc *desc)
101
{
102
return ((u32)desc->limit1 << 16) | (u32)desc->limit0;
103
}
104
105
typedef void (*secondary_startup_64_fn)(void*, void*);
106
static void hv_vtl_ap_entry(void)
107
{
108
((secondary_startup_64_fn)secondary_startup_64)(&boot_params, &boot_params);
109
}
110
111
static int hv_vtl_bringup_vcpu(u32 target_vp_index, int cpu, u64 eip_ignored)
112
{
113
u64 status;
114
int ret = 0;
115
struct hv_enable_vp_vtl *input;
116
unsigned long irq_flags;
117
118
struct desc_ptr gdt_ptr;
119
struct desc_ptr idt_ptr;
120
121
struct ldttss_desc *tss;
122
struct ldttss_desc *ldt;
123
struct desc_struct *gdt;
124
125
struct task_struct *idle = idle_thread_get(cpu);
126
u64 rsp = (unsigned long)idle->thread.sp;
127
128
u64 rip = (u64)&hv_vtl_ap_entry;
129
130
native_store_gdt(&gdt_ptr);
131
store_idt(&idt_ptr);
132
133
gdt = (struct desc_struct *)((void *)(gdt_ptr.address));
134
tss = (struct ldttss_desc *)(gdt + GDT_ENTRY_TSS);
135
ldt = (struct ldttss_desc *)(gdt + GDT_ENTRY_LDT);
136
137
local_irq_save(irq_flags);
138
139
input = *this_cpu_ptr(hyperv_pcpu_input_arg);
140
memset(input, 0, sizeof(*input));
141
142
input->partition_id = HV_PARTITION_ID_SELF;
143
input->vp_index = target_vp_index;
144
input->target_vtl.target_vtl = HV_VTL_MGMT;
145
146
/*
147
* The x86_64 Linux kernel follows the 16-bit -> 32-bit -> 64-bit
148
* mode transition sequence after waking up an AP with SIPI whose
149
* vector points to the 16-bit AP startup trampoline code. Here in
150
* VTL2, we can't perform that sequence as the AP has to start in
151
* the 64-bit mode.
152
*
153
* To make this happen, we tell the hypervisor to load a valid 64-bit
154
* context (most of which is just magic numbers from the CPU manual)
155
* so that AP jumps right to the 64-bit entry of the kernel, and the
156
* control registers are loaded with values that let the AP fetch the
157
* code and data and carry on with work it gets assigned.
158
*/
159
160
input->vp_context.rip = rip;
161
input->vp_context.rsp = rsp;
162
input->vp_context.rflags = 0x0000000000000002;
163
input->vp_context.efer = native_rdmsrq(MSR_EFER);
164
input->vp_context.cr0 = native_read_cr0();
165
input->vp_context.cr3 = __native_read_cr3();
166
input->vp_context.cr4 = native_read_cr4();
167
input->vp_context.msr_cr_pat = native_rdmsrq(MSR_IA32_CR_PAT);
168
input->vp_context.idtr.limit = idt_ptr.size;
169
input->vp_context.idtr.base = idt_ptr.address;
170
input->vp_context.gdtr.limit = gdt_ptr.size;
171
input->vp_context.gdtr.base = gdt_ptr.address;
172
173
/* Non-system desc (64bit), long, code, present */
174
input->vp_context.cs.selector = __KERNEL_CS;
175
input->vp_context.cs.base = 0;
176
input->vp_context.cs.limit = 0xffffffff;
177
input->vp_context.cs.attributes = 0xa09b;
178
/* Non-system desc (64bit), data, present, granularity, default */
179
input->vp_context.ss.selector = __KERNEL_DS;
180
input->vp_context.ss.base = 0;
181
input->vp_context.ss.limit = 0xffffffff;
182
input->vp_context.ss.attributes = 0xc093;
183
184
/* System desc (128bit), present, LDT */
185
input->vp_context.ldtr.selector = GDT_ENTRY_LDT * 8;
186
input->vp_context.ldtr.base = hv_vtl_system_desc_base(ldt);
187
input->vp_context.ldtr.limit = hv_vtl_system_desc_limit(ldt);
188
input->vp_context.ldtr.attributes = 0x82;
189
190
/* System desc (128bit), present, TSS, 0x8b - busy, 0x89 -- default */
191
input->vp_context.tr.selector = GDT_ENTRY_TSS * 8;
192
input->vp_context.tr.base = hv_vtl_system_desc_base(tss);
193
input->vp_context.tr.limit = hv_vtl_system_desc_limit(tss);
194
input->vp_context.tr.attributes = 0x8b;
195
196
status = hv_do_hypercall(HVCALL_ENABLE_VP_VTL, input, NULL);
197
198
if (!hv_result_success(status) &&
199
hv_result(status) != HV_STATUS_VTL_ALREADY_ENABLED) {
200
pr_err("HVCALL_ENABLE_VP_VTL failed for VP : %d ! [Err: %#llx\n]",
201
target_vp_index, status);
202
ret = -EINVAL;
203
goto free_lock;
204
}
205
206
status = hv_do_hypercall(HVCALL_START_VP, input, NULL);
207
208
if (!hv_result_success(status)) {
209
pr_err("HVCALL_START_VP failed for VP : %d ! [Err: %#llx]\n",
210
target_vp_index, status);
211
ret = -EINVAL;
212
}
213
214
free_lock:
215
local_irq_restore(irq_flags);
216
217
return ret;
218
}
219
220
static int hv_vtl_wakeup_secondary_cpu(u32 apicid, unsigned long start_eip, unsigned int cpu)
221
{
222
int vp_index;
223
224
pr_debug("Bringing up CPU with APIC ID %d in VTL2...\n", apicid);
225
vp_index = hv_apicid_to_vp_index(apicid);
226
227
if (vp_index < 0) {
228
pr_err("Couldn't find CPU with APIC ID %d\n", apicid);
229
return -EINVAL;
230
}
231
if (vp_index > ms_hyperv.max_vp_index) {
232
pr_err("Invalid CPU id %d for APIC ID %d\n", vp_index, apicid);
233
return -EINVAL;
234
}
235
236
return hv_vtl_bringup_vcpu(vp_index, cpu, start_eip);
237
}
238
239
int __init hv_vtl_early_init(void)
240
{
241
machine_ops.emergency_restart = hv_vtl_emergency_restart;
242
machine_ops.restart = hv_vtl_restart;
243
244
/*
245
* `boot_cpu_has` returns the runtime feature support,
246
* and here is the earliest it can be used.
247
*/
248
if (cpu_feature_enabled(X86_FEATURE_XSAVE))
249
panic("XSAVE has to be disabled as it is not supported by this module.\n"
250
"Please add 'noxsave' to the kernel command line.\n");
251
252
real_mode_header = &hv_vtl_real_mode_header;
253
apic_update_callback(wakeup_secondary_cpu_64, hv_vtl_wakeup_secondary_cpu);
254
255
return 0;
256
}
257
258
DEFINE_STATIC_CALL_NULL(__mshv_vtl_return_hypercall, void (*)(void));
259
260
void mshv_vtl_return_call_init(u64 vtl_return_offset)
261
{
262
static_call_update(__mshv_vtl_return_hypercall,
263
(void *)((u8 *)hv_hypercall_pg + vtl_return_offset));
264
}
265
EXPORT_SYMBOL(mshv_vtl_return_call_init);
266
267
void mshv_vtl_return_call(struct mshv_vtl_cpu_context *vtl0)
268
{
269
struct hv_vp_assist_page *hvp;
270
271
hvp = hv_vp_assist_page[smp_processor_id()];
272
hvp->vtl_ret_x64rax = vtl0->rax;
273
hvp->vtl_ret_x64rcx = vtl0->rcx;
274
275
kernel_fpu_begin_mask(0);
276
fxrstor(&vtl0->fx_state);
277
__mshv_vtl_return_call(vtl0);
278
fxsave(&vtl0->fx_state);
279
kernel_fpu_end();
280
}
281
EXPORT_SYMBOL(mshv_vtl_return_call);
282
283