2 * Copyright (c) KATO Takenori, 1997, 1998.
3 * Copyright (c) 2008 The DragonFly Project.
5 * All rights reserved. Unpublished rights reserved under the copyright
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer as
14 * the first lines of this file unmodified.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
19 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
20 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
21 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
22 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
23 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
24 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
25 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
26 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
27 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
28 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
33 #include <sys/param.h>
34 #include <sys/kernel.h>
35 #include <sys/systm.h>
36 #include <sys/sysctl.h>
38 #include <machine/cputypes.h>
39 #include <machine/md_var.h>
40 #include <machine/specialreg.h>
41 #include <machine/smp.h>
46 static int hw_instruction_sse
;
47 SYSCTL_INT(_hw
, OID_AUTO
, instruction_sse
, CTLFLAG_RD
,
48 &hw_instruction_sse
, 0, "SIMD/MMX2 instructions available in CPU");
50 int cpu
; /* Are we 386, 386sx, 486, etc? */
51 u_int cpu_feature
; /* Feature flags */
52 u_int cpu_feature2
; /* Feature flags */
53 u_int amd_feature
; /* AMD feature flags */
54 u_int amd_feature2
; /* AMD feature flags */
55 u_int via_feature_rng
; /* VIA RNG features */
56 u_int via_feature_xcrypt
; /* VIA ACE features */
57 u_int cpu_high
; /* Highest arg to CPUID */
58 u_int cpu_exthigh
; /* Highest arg to extended CPUID */
59 u_int cpu_id
; /* Stepping ID */
60 u_int cpu_procinfo
; /* HyperThreading Info / Brand Index / CLFUSH */
61 u_int cpu_procinfo2
; /* Multicore info */
62 char cpu_vendor
[20]; /* CPU Origin code */
63 u_int cpu_vendor_id
; /* CPU vendor ID */
64 u_int cpu_fxsr
; /* SSE enabled */
65 u_int cpu_xsave
; /* AVX enabled by OS*/
66 u_int cpu_mxcsr_mask
; /* Valid bits in mxcsr */
67 u_int cpu_clflush_line_size
= 32; /* Default CLFLUSH line size */
68 u_int cpu_stdext_feature
;
69 u_int cpu_thermal_feature
;
70 u_int cpu_mwait_feature
;
71 u_int cpu_mwait_extemu
;
74 * -1: automatic (enable on h/w, disable on VMs)
76 * 1: enable (where available)
78 static int hw_clflush_enable
= -1;
80 SYSCTL_INT(_hw
, OID_AUTO
, clflush_enable
, CTLFLAG_RD
, &hw_clflush_enable
, 0,
83 SYSCTL_UINT(_hw
, OID_AUTO
, via_feature_rng
, CTLFLAG_RD
,
84 &via_feature_rng
, 0, "VIA C3/C7 RNG feature available in CPU");
85 SYSCTL_UINT(_hw
, OID_AUTO
, via_feature_xcrypt
, CTLFLAG_RD
,
86 &via_feature_xcrypt
, 0, "VIA C3/C7 xcrypt feature available in CPU");
89 * Initialize special VIA C3/C7 features
97 do_cpuid(0xc0000000, regs
);
99 if (val
>= 0xc0000001) {
100 do_cpuid(0xc0000001, regs
);
105 /* Enable RNG if present and disabled */
106 if (val
& VIA_CPUID_HAS_RNG
) {
107 if (!(val
& VIA_CPUID_DO_RNG
)) {
108 msreg
= rdmsr(0x110B);
110 wrmsr(0x110B, msreg
);
112 via_feature_rng
= VIA_HAS_RNG
;
114 /* Enable AES engine if present and disabled */
115 if (val
& VIA_CPUID_HAS_ACE
) {
116 if (!(val
& VIA_CPUID_DO_ACE
)) {
117 msreg
= rdmsr(0x1107);
118 msreg
|= (0x01 << 28);
119 wrmsr(0x1107, msreg
);
121 via_feature_xcrypt
|= VIA_HAS_AES
;
123 /* Enable ACE2 engine if present and disabled */
124 if (val
& VIA_CPUID_HAS_ACE2
) {
125 if (!(val
& VIA_CPUID_DO_ACE2
)) {
126 msreg
= rdmsr(0x1107);
127 msreg
|= (0x01 << 28);
128 wrmsr(0x1107, msreg
);
130 via_feature_xcrypt
|= VIA_HAS_AESCTR
;
132 /* Enable SHA engine if present and disabled */
133 if (val
& VIA_CPUID_HAS_PHE
) {
134 if (!(val
& VIA_CPUID_DO_PHE
)) {
135 msreg
= rdmsr(0x1107);
136 msreg
|= (0x01 << 28/**/);
137 wrmsr(0x1107, msreg
);
139 via_feature_xcrypt
|= VIA_HAS_SHA
;
141 /* Enable MM engine if present and disabled */
142 if (val
& VIA_CPUID_HAS_PMM
) {
143 if (!(val
& VIA_CPUID_DO_PMM
)) {
144 msreg
= rdmsr(0x1107);
145 msreg
|= (0x01 << 28/**/);
146 wrmsr(0x1107, msreg
);
148 via_feature_xcrypt
|= VIA_HAS_MM
;
152 static enum vmm_guest_type
155 enum vmm_guest_type guest
;
158 * [RFC] CPUID usage for interaction between Hypervisors and Linux.
159 * http://lkml.org/lkml/2008/10/1/246
161 * KB1009458: Mechanisms to determine if software is running in
162 * a VMware virtual machine
163 * http://kb.vmware.com/kb/1009458
165 if (cpu_feature2
& CPUID2_VMM
) {
168 do_cpuid(0x40000000, regs
);
169 if (regs
[0] >= 0x40000000) {
170 ((u_int
*)&vmm_vendor
)[0] = regs
[1];
171 ((u_int
*)&vmm_vendor
)[1] = regs
[2];
172 ((u_int
*)&vmm_vendor
)[2] = regs
[3];
173 vmm_vendor
[12] = '\0';
174 if (strcmp(vmm_vendor
, "VMwareVMware") == 0)
175 return VMM_GUEST_VMWARE
;
176 else if (strcmp(vmm_vendor
, "Microsoft Hv") == 0)
177 return VMM_GUEST_HYPERV
;
181 guest
= detect_virtual();
182 if (guest
== VMM_GUEST_NONE
&& (cpu_feature2
& CPUID2_VMM
))
183 guest
= VMM_GUEST_UNKNOWN
;
188 * Initialize CPU control registers
191 initializecpu(int cpu
)
195 /*Check for FXSR and SSE support and enable if available.*/
196 if ((cpu_feature
& CPUID_XMM
) && (cpu_feature
& CPUID_FXSR
)) {
197 load_cr4(rcr4() | CR4_FXSR
| CR4_XMM
);
198 cpu_fxsr
= hw_instruction_sse
= 1;
202 /* Check if we are running in a hypervisor. */
203 vmm_guest
= detect_vmm();
206 #if !defined(CPU_DISABLE_AVX)
207 /*Check for XSAVE and AVX support and enable if available.*/
208 if ((cpu_feature2
& CPUID2_AVX
) && (cpu_feature2
& CPUID2_XSAVE
)
209 && (cpu_feature
& CPUID_SSE
)) {
210 load_cr4(rcr4() | CR4_XSAVE
);
212 /* Adjust size of savefpu in npx.h before adding to mask.*/
213 xsetbv(0, CPU_XFEATURE_X87
| CPU_XFEATURE_SSE
| CPU_XFEATURE_YMM
, 0);
218 if (cpu_vendor_id
== CPU_VENDOR_AMD
) {
219 switch((cpu_id
& 0xFF0000)) {
223 * Errata 721 is the cpu bug found by your's truly
224 * (Matthew Dillon). It is a bug where a sequence
225 * of 5 or more popq's + a retq, under involved
226 * deep recursion circumstances, can cause the %rsp
227 * to not be properly updated, almost always
228 * resulting in a seg-fault soon after.
230 * Do not install the workaround when we are running
231 * in a virtual machine.
236 msr
= rdmsr(MSR_AMD_DE_CFG
);
237 if ((msr
& 1) == 0) {
239 kprintf("Errata 721 workaround "
242 wrmsr(MSR_AMD_DE_CFG
, msr
);
248 if ((amd_feature
& AMDID_NX
) != 0) {
249 msr
= rdmsr(MSR_EFER
) | EFER_NXE
;
250 wrmsr(MSR_EFER
, msr
);
255 if (cpu_vendor_id
== CPU_VENDOR_CENTAUR
&&
256 CPUID_TO_FAMILY(cpu_id
) == 0x6 &&
257 CPUID_TO_MODEL(cpu_id
) >= 0xf)
260 TUNABLE_INT_FETCH("hw.clflush_enable", &hw_clflush_enable
);
261 if (cpu_feature
& CPUID_CLFSH
) {
262 cpu_clflush_line_size
= ((cpu_procinfo
>> 8) & 0xff) * 8;
264 if (hw_clflush_enable
== 0 ||
265 ((hw_clflush_enable
== -1) && vmm_guest
))
266 cpu_feature
&= ~CPUID_CLFSH
;