4 * Copyright (c) 2003 Fabrice Bellard
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
29 #include <sys/ucontext.h>
32 #if !defined(__x86_64__)
36 //#define LINUX_VM86_IOPL_FIX
37 //#define TEST_P4_FLAGS
38 #if defined(__x86_64__)
48 #if defined(__x86_64__)
49 #define FMT64X "%016lx"
50 #define FMTLX "%016lx"
51 #define X86_64_ONLY(x) x
53 #define FMT64X "%016llx"
55 #define X86_64_ONLY(x)
62 #define xglue(x, y) x ## y
63 #define glue(x, y) xglue(x, y)
64 #define stringify(s) tostring(s)
65 #define tostring(s) #s
74 #define __init_call __attribute__ ((unused,__section__ ("initcall")))
76 #define CC_MASK (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A)
78 #if defined(__x86_64__)
79 static inline long i2l(long v
)
81 return v
| ((v
^ 0xabcd) << 32);
84 static inline long i2l(long v
)
91 #include "test-i386.h"
94 #include "test-i386.h"
97 #include "test-i386.h"
100 #include "test-i386.h"
103 #include "test-i386.h"
106 #include "test-i386.h"
110 #include "test-i386.h"
114 #include "test-i386.h"
119 #include "test-i386.h"
124 #include "test-i386.h"
129 #include "test-i386.h"
134 #include "test-i386.h"
137 #define CC_MASK (CC_C | CC_P | CC_Z | CC_S | CC_O)
140 #include "test-i386-shift.h"
143 #include "test-i386-shift.h"
146 #include "test-i386-shift.h"
149 #include "test-i386-shift.h"
152 #include "test-i386-shift.h"
156 #include "test-i386-shift.h"
160 #include "test-i386-shift.h"
165 #include "test-i386-shift.h"
170 #include "test-i386-shift.h"
172 /* XXX: should be more precise ? */
174 #define CC_MASK (CC_C)
178 #include "test-i386-shift.h"
182 #include "test-i386-shift.h"
186 #include "test-i386-shift.h"
190 #include "test-i386-shift.h"
192 /* lea test (modrm support) */
193 #define TEST_LEAQ(STR)\
195 asm("lea " STR ", %0"\
197 : "a" (eax), "b" (ebx), "c" (ecx), "d" (edx), "S" (esi), "D" (edi));\
198 printf("lea %s = " FMTLX "\n", STR, res);\
201 #define TEST_LEA(STR)\
203 asm("lea " STR ", %0"\
205 : "a" (eax), "b" (ebx), "c" (ecx), "d" (edx), "S" (esi), "D" (edi));\
206 printf("lea %s = " FMTLX "\n", STR, res);\
209 #define TEST_LEA16(STR)\
211 asm(".code16 ; .byte 0x67 ; leal " STR ", %0 ; .code32"\
213 : "a" (eax), "b" (ebx), "c" (ecx), "d" (edx), "S" (esi), "D" (edi));\
214 printf("lea %s = %08lx\n", STR, res);\
220 long eax
, ebx
, ecx
, edx
, esi
, edi
, res
;
237 TEST_LEA("0x40(%%eax)");
238 TEST_LEA("0x40(%%ebx)");
239 TEST_LEA("0x40(%%ecx)");
240 TEST_LEA("0x40(%%edx)");
241 TEST_LEA("0x40(%%esi)");
242 TEST_LEA("0x40(%%edi)");
244 TEST_LEA("0x4000(%%eax)");
245 TEST_LEA("0x4000(%%ebx)");
246 TEST_LEA("0x4000(%%ecx)");
247 TEST_LEA("0x4000(%%edx)");
248 TEST_LEA("0x4000(%%esi)");
249 TEST_LEA("0x4000(%%edi)");
251 TEST_LEA("(%%eax, %%ecx)");
252 TEST_LEA("(%%ebx, %%edx)");
253 TEST_LEA("(%%ecx, %%ecx)");
254 TEST_LEA("(%%edx, %%ecx)");
255 TEST_LEA("(%%esi, %%ecx)");
256 TEST_LEA("(%%edi, %%ecx)");
258 TEST_LEA("0x40(%%eax, %%ecx)");
259 TEST_LEA("0x4000(%%ebx, %%edx)");
261 TEST_LEA("(%%ecx, %%ecx, 2)");
262 TEST_LEA("(%%edx, %%ecx, 4)");
263 TEST_LEA("(%%esi, %%ecx, 8)");
265 TEST_LEA("(,%%eax, 2)");
266 TEST_LEA("(,%%ebx, 4)");
267 TEST_LEA("(,%%ecx, 8)");
269 TEST_LEA("0x40(,%%eax, 2)");
270 TEST_LEA("0x40(,%%ebx, 4)");
271 TEST_LEA("0x40(,%%ecx, 8)");
274 TEST_LEA("-10(%%ecx, %%ecx, 2)");
275 TEST_LEA("-10(%%edx, %%ecx, 4)");
276 TEST_LEA("-10(%%esi, %%ecx, 8)");
278 TEST_LEA("0x4000(%%ecx, %%ecx, 2)");
279 TEST_LEA("0x4000(%%edx, %%ecx, 4)");
280 TEST_LEA("0x4000(%%esi, %%ecx, 8)");
282 #if defined(__x86_64__)
284 TEST_LEAQ("0x4000(%%rip)");
286 TEST_LEAQ("(%%rax)");
287 TEST_LEAQ("(%%rbx)");
288 TEST_LEAQ("(%%rcx)");
289 TEST_LEAQ("(%%rdx)");
290 TEST_LEAQ("(%%rsi)");
291 TEST_LEAQ("(%%rdi)");
293 TEST_LEAQ("0x40(%%rax)");
294 TEST_LEAQ("0x40(%%rbx)");
295 TEST_LEAQ("0x40(%%rcx)");
296 TEST_LEAQ("0x40(%%rdx)");
297 TEST_LEAQ("0x40(%%rsi)");
298 TEST_LEAQ("0x40(%%rdi)");
300 TEST_LEAQ("0x4000(%%rax)");
301 TEST_LEAQ("0x4000(%%rbx)");
302 TEST_LEAQ("0x4000(%%rcx)");
303 TEST_LEAQ("0x4000(%%rdx)");
304 TEST_LEAQ("0x4000(%%rsi)");
305 TEST_LEAQ("0x4000(%%rdi)");
307 TEST_LEAQ("(%%rax, %%rcx)");
308 TEST_LEAQ("(%%rbx, %%rdx)");
309 TEST_LEAQ("(%%rcx, %%rcx)");
310 TEST_LEAQ("(%%rdx, %%rcx)");
311 TEST_LEAQ("(%%rsi, %%rcx)");
312 TEST_LEAQ("(%%rdi, %%rcx)");
314 TEST_LEAQ("0x40(%%rax, %%rcx)");
315 TEST_LEAQ("0x4000(%%rbx, %%rdx)");
317 TEST_LEAQ("(%%rcx, %%rcx, 2)");
318 TEST_LEAQ("(%%rdx, %%rcx, 4)");
319 TEST_LEAQ("(%%rsi, %%rcx, 8)");
321 TEST_LEAQ("(,%%rax, 2)");
322 TEST_LEAQ("(,%%rbx, 4)");
323 TEST_LEAQ("(,%%rcx, 8)");
325 TEST_LEAQ("0x40(,%%rax, 2)");
326 TEST_LEAQ("0x40(,%%rbx, 4)");
327 TEST_LEAQ("0x40(,%%rcx, 8)");
330 TEST_LEAQ("-10(%%rcx, %%rcx, 2)");
331 TEST_LEAQ("-10(%%rdx, %%rcx, 4)");
332 TEST_LEAQ("-10(%%rsi, %%rcx, 8)");
334 TEST_LEAQ("0x4000(%%rcx, %%rcx, 2)");
335 TEST_LEAQ("0x4000(%%rdx, %%rcx, 4)");
336 TEST_LEAQ("0x4000(%%rsi, %%rcx, 8)");
338 /* limited 16 bit addressing test */
339 TEST_LEA16("0x4000");
340 TEST_LEA16("(%%bx)");
341 TEST_LEA16("(%%si)");
342 TEST_LEA16("(%%di)");
343 TEST_LEA16("0x40(%%bx)");
344 TEST_LEA16("0x40(%%si)");
345 TEST_LEA16("0x40(%%di)");
346 TEST_LEA16("0x4000(%%bx)");
347 TEST_LEA16("0x4000(%%si)");
348 TEST_LEA16("(%%bx,%%si)");
349 TEST_LEA16("(%%bx,%%di)");
350 TEST_LEA16("0x40(%%bx,%%si)");
351 TEST_LEA16("0x40(%%bx,%%di)");
352 TEST_LEA16("0x4000(%%bx,%%si)");
353 TEST_LEA16("0x4000(%%bx,%%di)");
357 #define TEST_JCC(JCC, v1, v2)\
360 asm("movl $1, %0\n\t"\
366 : "r" (v1), "r" (v2));\
367 printf("%-10s %d\n", "j" JCC, res);\
369 asm("movl $0, %0\n\t"\
371 "set" JCC " %b0\n\t"\
373 : "r" (v1), "r" (v2));\
374 printf("%-10s %d\n", "set" JCC, res);\
377 long res = i2l(0x12345678);\
379 asm("cmpl %2, %1\n\t"\
380 "cmov" JCC "q %3, %0\n\t"\
382 : "r" (v1), "r" (v2), "m" (val), "0" (res));\
383 printf("%-10s R=" FMTLX "\n", "cmov" JCC "q", res);)\
384 asm("cmpl %2, %1\n\t"\
385 "cmov" JCC "l %k3, %k0\n\t"\
387 : "r" (v1), "r" (v2), "m" (val), "0" (res));\
388 printf("%-10s R=" FMTLX "\n", "cmov" JCC "l", res);\
389 asm("cmpl %2, %1\n\t"\
390 "cmov" JCC "w %w3, %w0\n\t"\
392 : "r" (v1), "r" (v2), "r" (1), "0" (res));\
393 printf("%-10s R=" FMTLX "\n", "cmov" JCC "w", res);\
397 /* various jump tests */
400 TEST_JCC("ne", 1, 1);
401 TEST_JCC("ne", 1, 0);
408 TEST_JCC("l", 1, -1);
410 TEST_JCC("le", 1, 1);
411 TEST_JCC("le", 1, 0);
412 TEST_JCC("le", 1, -1);
414 TEST_JCC("ge", 1, 1);
415 TEST_JCC("ge", 1, 0);
416 TEST_JCC("ge", -1, 1);
420 TEST_JCC("g", 1, -1);
424 TEST_JCC("b", 1, -1);
426 TEST_JCC("be", 1, 1);
427 TEST_JCC("be", 1, 0);
428 TEST_JCC("be", 1, -1);
430 TEST_JCC("ae", 1, 1);
431 TEST_JCC("ae", 1, 0);
432 TEST_JCC("ae", 1, -1);
436 TEST_JCC("a", 1, -1);
442 TEST_JCC("np", 1, 1);
443 TEST_JCC("np", 1, 0);
445 TEST_JCC("o", 0x7fffffff, 0);
446 TEST_JCC("o", 0x7fffffff, -1);
448 TEST_JCC("no", 0x7fffffff, 0);
449 TEST_JCC("no", 0x7fffffff, -1);
452 TEST_JCC("s", 0, -1);
455 TEST_JCC("ns", 0, 1);
456 TEST_JCC("ns", 0, -1);
457 TEST_JCC("ns", 0, 0);
462 #define CC_MASK (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A)
464 #define CC_MASK (CC_O | CC_C)
468 #include "test-i386-muldiv.h"
471 #include "test-i386-muldiv.h"
473 void test_imulw2(long op0
, long op1
)
475 long res
, s1
, s0
, flags
;
480 asm volatile ("push %4\n\t"
485 : "=q" (res
), "=g" (flags
)
486 : "q" (s1
), "0" (res
), "1" (flags
));
487 printf("%-10s A=" FMTLX
" B=" FMTLX
" R=" FMTLX
" CC=%04lx\n",
488 "imulw", s0
, s1
, res
, flags
& CC_MASK
);
491 void test_imull2(long op0
, long op1
)
493 long res
, s1
, s0
, flags
;
498 asm volatile ("push %4\n\t"
503 : "=q" (res
), "=g" (flags
)
504 : "q" (s1
), "0" (res
), "1" (flags
));
505 printf("%-10s A=" FMTLX
" B=" FMTLX
" R=" FMTLX
" CC=%04lx\n",
506 "imull", s0
, s1
, res
, flags
& CC_MASK
);
509 #if defined(__x86_64__)
510 void test_imulq2(long op0
, long op1
)
512 long res
, s1
, s0
, flags
;
517 asm volatile ("push %4\n\t"
522 : "=q" (res
), "=g" (flags
)
523 : "q" (s1
), "0" (res
), "1" (flags
));
524 printf("%-10s A=" FMTLX
" B=" FMTLX
" R=" FMTLX
" CC=%04lx\n",
525 "imulq", s0
, s1
, res
, flags
& CC_MASK
);
529 #define TEST_IMUL_IM(size, rsize, op0, op1)\
531 long res, flags, s1;\
535 asm volatile ("push %3\n\t"\
537 "imul" size " $" #op0 ", %" rsize "2, %" rsize "0\n\t" \
540 : "=r" (res), "=g" (flags)\
541 : "r" (s1), "1" (flags), "0" (res));\
542 printf("%-10s A=" FMTLX " B=" FMTLX " R=" FMTLX " CC=%04lx\n",\
543 "imul" size " im", (long)op0, (long)op1, res, flags & CC_MASK);\
551 #include "test-i386-muldiv.h"
554 #include "test-i386-muldiv.h"
558 test_imulb(0x1234561d, 4);
560 test_imulb(0x80, 0x80);
561 test_imulb(0x10, 0x10);
563 test_imulw(0, 0x1234001d, 45);
564 test_imulw(0, 23, -45);
565 test_imulw(0, 0x8000, 0x8000);
566 test_imulw(0, 0x100, 0x100);
568 test_imull(0, 0x1234001d, 45);
569 test_imull(0, 23, -45);
570 test_imull(0, 0x80000000, 0x80000000);
571 test_imull(0, 0x10000, 0x10000);
573 test_mulb(0x1234561d, 4);
575 test_mulb(0x80, 0x80);
576 test_mulb(0x10, 0x10);
578 test_mulw(0, 0x1234001d, 45);
579 test_mulw(0, 23, -45);
580 test_mulw(0, 0x8000, 0x8000);
581 test_mulw(0, 0x100, 0x100);
583 test_mull(0, 0x1234001d, 45);
584 test_mull(0, 23, -45);
585 test_mull(0, 0x80000000, 0x80000000);
586 test_mull(0, 0x10000, 0x10000);
588 test_imulw2(0x1234001d, 45);
589 test_imulw2(23, -45);
590 test_imulw2(0x8000, 0x8000);
591 test_imulw2(0x100, 0x100);
593 test_imull2(0x1234001d, 45);
594 test_imull2(23, -45);
595 test_imull2(0x80000000, 0x80000000);
596 test_imull2(0x10000, 0x10000);
598 TEST_IMUL_IM("w", "w", 45, 0x1234);
599 TEST_IMUL_IM("w", "w", -45, 23);
600 TEST_IMUL_IM("w", "w", 0x8000, 0x80000000);
601 TEST_IMUL_IM("w", "w", 0x7fff, 0x1000);
603 TEST_IMUL_IM("l", "k", 45, 0x1234);
604 TEST_IMUL_IM("l", "k", -45, 23);
605 TEST_IMUL_IM("l", "k", 0x8000, 0x80000000);
606 TEST_IMUL_IM("l", "k", 0x7fff, 0x1000);
608 test_idivb(0x12341678, 0x127e);
609 test_idivb(0x43210123, -5);
610 test_idivb(0x12340004, -1);
612 test_idivw(0, 0x12345678, 12347);
613 test_idivw(0, -23223, -45);
614 test_idivw(0, 0x12348000, -1);
615 test_idivw(0x12343, 0x12345678, 0x81238567);
617 test_idivl(0, 0x12345678, 12347);
618 test_idivl(0, -233223, -45);
619 test_idivl(0, 0x80000000, -1);
620 test_idivl(0x12343, 0x12345678, 0x81234567);
622 test_divb(0x12341678, 0x127e);
623 test_divb(0x43210123, -5);
624 test_divb(0x12340004, -1);
626 test_divw(0, 0x12345678, 12347);
627 test_divw(0, -23223, -45);
628 test_divw(0, 0x12348000, -1);
629 test_divw(0x12343, 0x12345678, 0x81238567);
631 test_divl(0, 0x12345678, 12347);
632 test_divl(0, -233223, -45);
633 test_divl(0, 0x80000000, -1);
634 test_divl(0x12343, 0x12345678, 0x81234567);
636 #if defined(__x86_64__)
637 test_imulq(0, 0x1234001d1234001d, 45);
638 test_imulq(0, 23, -45);
639 test_imulq(0, 0x8000000000000000, 0x8000000000000000);
640 test_imulq(0, 0x100000000, 0x100000000);
642 test_mulq(0, 0x1234001d1234001d, 45);
643 test_mulq(0, 23, -45);
644 test_mulq(0, 0x8000000000000000, 0x8000000000000000);
645 test_mulq(0, 0x100000000, 0x100000000);
647 test_imulq2(0x1234001d1234001d, 45);
648 test_imulq2(23, -45);
649 test_imulq2(0x8000000000000000, 0x8000000000000000);
650 test_imulq2(0x100000000, 0x100000000);
652 TEST_IMUL_IM("q", "", 45, 0x12341234);
653 TEST_IMUL_IM("q", "", -45, 23);
654 TEST_IMUL_IM("q", "", 0x8000, 0x8000000000000000);
655 TEST_IMUL_IM("q", "", 0x7fff, 0x10000000);
657 test_idivq(0, 0x12345678abcdef, 12347);
658 test_idivq(0, -233223, -45);
659 test_idivq(0, 0x8000000000000000, -1);
660 test_idivq(0x12343, 0x12345678, 0x81234567);
662 test_divq(0, 0x12345678abcdef, 12347);
663 test_divq(0, -233223, -45);
664 test_divq(0, 0x8000000000000000, -1);
665 test_divq(0x12343, 0x12345678, 0x81234567);
669 #define TEST_BSX(op, size, op0)\
671 long res, val, resz;\
674 "mov $0x12345678, %0\n"\
675 #op " %" size "2, %" size "0 ; setz %b1" \
676 : "=r" (res), "=q" (resz)\
678 printf("%-10s A=" FMTLX " R=" FMTLX " %ld\n", #op, val, res, resz);\
683 TEST_BSX(bsrw
, "w", 0);
684 TEST_BSX(bsrw
, "w", 0x12340128);
685 TEST_BSX(bsfw
, "w", 0);
686 TEST_BSX(bsfw
, "w", 0x12340128);
687 TEST_BSX(bsrl
, "k", 0);
688 TEST_BSX(bsrl
, "k", 0x00340128);
689 TEST_BSX(bsfl
, "k", 0);
690 TEST_BSX(bsfl
, "k", 0x00340128);
691 #if defined(__x86_64__)
692 TEST_BSX(bsrq
, "", 0);
693 TEST_BSX(bsrq
, "", 0x003401281234);
694 TEST_BSX(bsfq
, "", 0);
695 TEST_BSX(bsfq
, "", 0x003401281234);
699 /**********************************************/
706 union float64u q_nan
= { .l
= 0xFFF8000000000000 };
707 union float64u s_nan
= { .l
= 0xFFF0000000000000 };
709 void test_fops(double a
, double b
)
711 printf("a=%f b=%f a+b=%f\n", a
, b
, a
+ b
);
712 printf("a=%f b=%f a-b=%f\n", a
, b
, a
- b
);
713 printf("a=%f b=%f a*b=%f\n", a
, b
, a
* b
);
714 printf("a=%f b=%f a/b=%f\n", a
, b
, a
/ b
);
715 printf("a=%f b=%f fmod(a, b)=%f\n", a
, b
, fmod(a
, b
));
716 printf("a=%f sqrt(a)=%f\n", a
, sqrt(a
));
717 printf("a=%f sin(a)=%f\n", a
, sin(a
));
718 printf("a=%f cos(a)=%f\n", a
, cos(a
));
719 printf("a=%f tan(a)=%f\n", a
, tan(a
));
720 printf("a=%f log(a)=%f\n", a
, log(a
));
721 printf("a=%f exp(a)=%f\n", a
, exp(a
));
722 printf("a=%f b=%f atan2(a, b)=%f\n", a
, b
, atan2(a
, b
));
723 /* just to test some op combining */
724 printf("a=%f asin(sin(a))=%f\n", a
, asin(sin(a
)));
725 printf("a=%f acos(cos(a))=%f\n", a
, acos(cos(a
)));
726 printf("a=%f atan(tan(a))=%f\n", a
, atan(tan(a
)));
730 void fpu_clear_exceptions(void)
732 struct __attribute__((packed
)) {
740 long double fpregs
[8];
743 asm volatile ("fnstenv %0\n" : : "m" (float_env32
));
744 float_env32
.fpus
&= ~0x7f;
745 asm volatile ("fldenv %0\n" : : "m" (float_env32
));
748 /* XXX: display exception bits when supported */
749 #define FPUS_EMASK 0x0000
750 //#define FPUS_EMASK 0x007f
752 void test_fcmp(double a
, double b
)
756 fpu_clear_exceptions();
761 printf("fcom(%f %f)=%04lx \n",
762 a
, b
, fpus
& (0x4500 | FPUS_EMASK
));
763 fpu_clear_exceptions();
768 printf("fucom(%f %f)=%04lx\n",
769 a
, b
, fpus
& (0x4500 | FPUS_EMASK
));
771 /* test f(u)comi instruction */
772 fpu_clear_exceptions();
777 : "=r" (eflags
), "=a" (fpus
)
779 printf("fcomi(%f %f)=%04lx %02lx\n",
780 a
, b
, fpus
& FPUS_EMASK
, eflags
& (CC_Z
| CC_P
| CC_C
));
781 fpu_clear_exceptions();
782 asm("fucomi %3, %2\n"
786 : "=r" (eflags
), "=a" (fpus
)
788 printf("fucomi(%f %f)=%04lx %02lx\n",
789 a
, b
, fpus
& FPUS_EMASK
, eflags
& (CC_Z
| CC_P
| CC_C
));
791 fpu_clear_exceptions();
794 void test_fcvt(double a
)
807 printf("(float)%f = %f\n", a
, fa
);
808 printf("(long double)%f = %Lf\n", a
, la
);
809 printf("a=" FMT64X
"\n", *(uint64_t *)&a
);
810 printf("la=" FMT64X
" %04x\n", *(uint64_t *)&la
,
811 *(unsigned short *)((char *)(&la
) + 8));
813 /* test all roundings */
814 asm volatile ("fstcw %0" : "=m" (fpuc
));
816 asm volatile ("fldcw %0" : : "m" ((fpuc
& ~0x0c00) | (i
<< 10)));
817 asm volatile ("fist %0" : "=m" (wa
) : "t" (a
));
818 asm volatile ("fistl %0" : "=m" (ia
) : "t" (a
));
819 asm volatile ("fistpll %0" : "=m" (lla
) : "t" (a
) : "st");
820 asm volatile ("frndint ; fstl %0" : "=m" (ra
) : "t" (a
));
821 asm volatile ("fldcw %0" : : "m" (fpuc
));
822 printf("(short)a = %d\n", wa
);
823 printf("(int)a = %d\n", ia
);
824 printf("(int64_t)a = " FMT64X
"\n", lla
);
825 printf("rint(a) = %f\n", ra
);
830 asm("fld" #N : "=t" (a)); \
831 printf("fld" #N "= %f\n", a);
833 void test_fconst(void)
845 void test_fbcd(double a
)
847 unsigned short bcd
[5];
850 asm("fbstp %0" : "=m" (bcd
[0]) : "t" (a
) : "st");
851 asm("fbld %1" : "=t" (b
) : "m" (bcd
[0]));
852 printf("a=%f bcd=%04x%04x%04x%04x%04x b=%f\n",
853 a
, bcd
[4], bcd
[3], bcd
[2], bcd
[1], bcd
[0], b
);
856 #define TEST_ENV(env, save, restore)\
858 memset((env), 0xaa, sizeof(*(env)));\
860 asm volatile ("fldl %0" : : "m" (dtab[i]));\
861 asm volatile (save " %0\n" : : "m" (*(env)));\
862 asm volatile (restore " %0\n": : "m" (*(env)));\
864 asm volatile ("fstpl %0" : "=m" (rtab[i]));\
866 printf("res[%d]=%f\n", i, rtab[i]);\
867 printf("fpuc=%04x fpus=%04x fptag=%04x\n",\
869 (env)->fpus & 0xff00,\
875 struct __attribute__((packed
)) {
883 long double fpregs
[8];
885 struct __attribute__((packed
)) {
890 long double fpregs
[8];
899 TEST_ENV(&float_env16
, "data16 fnstenv", "data16 fldenv");
900 TEST_ENV(&float_env16
, "data16 fnsave", "data16 frstor");
901 TEST_ENV(&float_env32
, "fnstenv", "fldenv");
902 TEST_ENV(&float_env32
, "fnsave", "frstor");
906 asm volatile ("fldl %0" : : "m" (dtab
[i
]));
907 asm volatile("ffree %st(2)");
908 asm volatile ("fnstenv %0\n" : : "m" (float_env32
));
909 asm volatile ("fninit");
910 printf("fptag=%04x\n", float_env32
.fptag
);
914 #define TEST_FCMOV(a, b, eflags, CC)\
919 "fcmov" CC " %2, %0\n"\
921 : "0" (a), "u" (b), "g" (eflags));\
922 printf("fcmov%s eflags=0x%04lx-> %f\n", \
923 CC, (long)eflags, res);\
926 void test_fcmov(void)
933 for(i
= 0; i
< 4; i
++) {
939 TEST_FCMOV(a
, b
, eflags
, "b");
940 TEST_FCMOV(a
, b
, eflags
, "e");
941 TEST_FCMOV(a
, b
, eflags
, "be");
942 TEST_FCMOV(a
, b
, eflags
, "nb");
943 TEST_FCMOV(a
, b
, eflags
, "ne");
944 TEST_FCMOV(a
, b
, eflags
, "nbe");
946 TEST_FCMOV(a
, b
, 0, "u");
947 TEST_FCMOV(a
, b
, CC_P
, "u");
948 TEST_FCMOV(a
, b
, 0, "nu");
949 TEST_FCMOV(a
, b
, CC_P
, "nu");
952 void test_floats(void)
959 test_fcmp(2, q_nan
.d
);
960 test_fcmp(q_nan
.d
, -1);
968 test_fbcd(1234567890123456);
969 test_fbcd(-123451234567890);
976 /**********************************************/
977 #if !defined(__x86_64__)
979 #define TEST_BCD(op, op0, cc_in, cc_mask)\
989 : "=a" (res), "=g" (flags)\
990 : "0" (res), "1" (flags));\
991 printf("%-10s A=%08x R=%08x CCIN=%04x CC=%04x\n",\
992 #op, op0, res, cc_in, flags & cc_mask);\
997 TEST_BCD(daa
, 0x12340503, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
998 TEST_BCD(daa
, 0x12340506, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
999 TEST_BCD(daa
, 0x12340507, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1000 TEST_BCD(daa
, 0x12340559, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1001 TEST_BCD(daa
, 0x12340560, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1002 TEST_BCD(daa
, 0x1234059f, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1003 TEST_BCD(daa
, 0x123405a0, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1004 TEST_BCD(daa
, 0x12340503, 0, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1005 TEST_BCD(daa
, 0x12340506, 0, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1006 TEST_BCD(daa
, 0x12340503, CC_C
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1007 TEST_BCD(daa
, 0x12340506, CC_C
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1008 TEST_BCD(daa
, 0x12340503, CC_C
| CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1009 TEST_BCD(daa
, 0x12340506, CC_C
| CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1011 TEST_BCD(das
, 0x12340503, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1012 TEST_BCD(das
, 0x12340506, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1013 TEST_BCD(das
, 0x12340507, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1014 TEST_BCD(das
, 0x12340559, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1015 TEST_BCD(das
, 0x12340560, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1016 TEST_BCD(das
, 0x1234059f, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1017 TEST_BCD(das
, 0x123405a0, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1018 TEST_BCD(das
, 0x12340503, 0, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1019 TEST_BCD(das
, 0x12340506, 0, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1020 TEST_BCD(das
, 0x12340503, CC_C
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1021 TEST_BCD(das
, 0x12340506, CC_C
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1022 TEST_BCD(das
, 0x12340503, CC_C
| CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1023 TEST_BCD(das
, 0x12340506, CC_C
| CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_A
));
1025 TEST_BCD(aaa
, 0x12340205, CC_A
, (CC_C
| CC_A
));
1026 TEST_BCD(aaa
, 0x12340306, CC_A
, (CC_C
| CC_A
));
1027 TEST_BCD(aaa
, 0x1234040a, CC_A
, (CC_C
| CC_A
));
1028 TEST_BCD(aaa
, 0x123405fa, CC_A
, (CC_C
| CC_A
));
1029 TEST_BCD(aaa
, 0x12340205, 0, (CC_C
| CC_A
));
1030 TEST_BCD(aaa
, 0x12340306, 0, (CC_C
| CC_A
));
1031 TEST_BCD(aaa
, 0x1234040a, 0, (CC_C
| CC_A
));
1032 TEST_BCD(aaa
, 0x123405fa, 0, (CC_C
| CC_A
));
1034 TEST_BCD(aas
, 0x12340205, CC_A
, (CC_C
| CC_A
));
1035 TEST_BCD(aas
, 0x12340306, CC_A
, (CC_C
| CC_A
));
1036 TEST_BCD(aas
, 0x1234040a, CC_A
, (CC_C
| CC_A
));
1037 TEST_BCD(aas
, 0x123405fa, CC_A
, (CC_C
| CC_A
));
1038 TEST_BCD(aas
, 0x12340205, 0, (CC_C
| CC_A
));
1039 TEST_BCD(aas
, 0x12340306, 0, (CC_C
| CC_A
));
1040 TEST_BCD(aas
, 0x1234040a, 0, (CC_C
| CC_A
));
1041 TEST_BCD(aas
, 0x123405fa, 0, (CC_C
| CC_A
));
1043 TEST_BCD(aam
, 0x12340547, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_O
| CC_A
));
1044 TEST_BCD(aad
, 0x12340407, CC_A
, (CC_C
| CC_P
| CC_Z
| CC_S
| CC_O
| CC_A
));
1048 #define TEST_XCHG(op, size, opconst)\
1051 op0 = i2l(0x12345678);\
1052 op1 = i2l(0xfbca7654);\
1053 asm(#op " %" size "0, %" size "1" \
1054 : "=q" (op0), opconst (op1) \
1055 : "0" (op0), "1" (op1));\
1056 printf("%-10s A=" FMTLX " B=" FMTLX "\n",\
1060 #define TEST_CMPXCHG(op, size, opconst, eax)\
1062 long op0, op1, op2;\
1063 op0 = i2l(0x12345678);\
1064 op1 = i2l(0xfbca7654);\
1066 asm(#op " %" size "0, %" size "1" \
1067 : "=q" (op0), opconst (op1) \
1068 : "0" (op0), "1" (op1), "a" (op2));\
1069 printf("%-10s EAX=" FMTLX " A=" FMTLX " C=" FMTLX "\n",\
1070 #op, op2, op0, op1);\
1073 void test_xchg(void)
1075 #if defined(__x86_64__)
1076 TEST_XCHG(xchgq
, "", "=q");
1078 TEST_XCHG(xchgl
, "k", "=q");
1079 TEST_XCHG(xchgw
, "w", "=q");
1080 TEST_XCHG(xchgb
, "b", "=q");
1082 #if defined(__x86_64__)
1083 TEST_XCHG(xchgq
, "", "=m");
1085 TEST_XCHG(xchgl
, "k", "=m");
1086 TEST_XCHG(xchgw
, "w", "=m");
1087 TEST_XCHG(xchgb
, "b", "=m");
1089 #if defined(__x86_64__)
1090 TEST_XCHG(xaddq
, "", "=q");
1092 TEST_XCHG(xaddl
, "k", "=q");
1093 TEST_XCHG(xaddw
, "w", "=q");
1094 TEST_XCHG(xaddb
, "b", "=q");
1099 asm("xaddl %1, %0" : "=r" (res
) : "0" (res
));
1100 printf("xaddl same res=%08x\n", res
);
1103 #if defined(__x86_64__)
1104 TEST_XCHG(xaddq
, "", "=m");
1106 TEST_XCHG(xaddl
, "k", "=m");
1107 TEST_XCHG(xaddw
, "w", "=m");
1108 TEST_XCHG(xaddb
, "b", "=m");
1110 #if defined(__x86_64__)
1111 TEST_CMPXCHG(cmpxchgq
, "", "=q", 0xfbca7654);
1113 TEST_CMPXCHG(cmpxchgl
, "k", "=q", 0xfbca7654);
1114 TEST_CMPXCHG(cmpxchgw
, "w", "=q", 0xfbca7654);
1115 TEST_CMPXCHG(cmpxchgb
, "b", "=q", 0xfbca7654);
1117 #if defined(__x86_64__)
1118 TEST_CMPXCHG(cmpxchgq
, "", "=q", 0xfffefdfc);
1120 TEST_CMPXCHG(cmpxchgl
, "k", "=q", 0xfffefdfc);
1121 TEST_CMPXCHG(cmpxchgw
, "w", "=q", 0xfffefdfc);
1122 TEST_CMPXCHG(cmpxchgb
, "b", "=q", 0xfffefdfc);
1124 #if defined(__x86_64__)
1125 TEST_CMPXCHG(cmpxchgq
, "", "=m", 0xfbca7654);
1127 TEST_CMPXCHG(cmpxchgl
, "k", "=m", 0xfbca7654);
1128 TEST_CMPXCHG(cmpxchgw
, "w", "=m", 0xfbca7654);
1129 TEST_CMPXCHG(cmpxchgb
, "b", "=m", 0xfbca7654);
1131 #if defined(__x86_64__)
1132 TEST_CMPXCHG(cmpxchgq
, "", "=m", 0xfffefdfc);
1134 TEST_CMPXCHG(cmpxchgl
, "k", "=m", 0xfffefdfc);
1135 TEST_CMPXCHG(cmpxchgw
, "w", "=m", 0xfffefdfc);
1136 TEST_CMPXCHG(cmpxchgb
, "b", "=m", 0xfffefdfc);
1139 uint64_t op0
, op1
, op2
;
1142 for(i
= 0; i
< 2; i
++) {
1143 op0
= 0x123456789abcd;
1145 op1
= 0xfbca765423456;
1148 op2
= 0x6532432432434;
1149 asm("cmpxchg8b %1\n"
1152 : "=A" (op0
), "=m" (op1
), "=g" (eflags
)
1153 : "0" (op0
), "m" (op1
), "b" ((int)op2
), "c" ((int)(op2
>> 32)));
1154 printf("cmpxchg8b: op0=" FMT64X
" op1=" FMT64X
" CC=%02lx\n",
1155 op0
, op1
, eflags
& CC_Z
);
1161 /**********************************************/
1162 /* segmentation tests */
1164 #include <asm/ldt.h>
1165 #include <linux/unistd.h>
1166 #include <linux/version.h>
1168 _syscall3(int, modify_ldt
, int, func
, void *, ptr
, unsigned long, bytecount
)
1170 #if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 5, 66)
1171 #define modify_ldt_ldt_s user_desc
1174 #define MK_SEL(n) (((n) << 3) | 7)
1176 uint8_t seg_data1
[4096];
1177 uint8_t seg_data2
[4096];
1179 #define TEST_LR(op, size, seg, mask)\
1183 asm (op " %" size "2, %" size "0\n" \
1188 : "=r" (res), "=r" (res2) : "m" (seg), "0" (res));\
1189 printf(op ": Z=%d %08x\n", res2, res & ~(mask));\
1192 /* NOTE: we use Linux modify_ldt syscall */
1193 void test_segs(void)
1195 struct modify_ldt_ldt_s ldt
;
1196 long long ldt_table
[3];
1202 } __attribute__((packed
)) segoff
;
1204 ldt
.entry_number
= 1;
1205 ldt
.base_addr
= (unsigned long)&seg_data1
;
1206 ldt
.limit
= (sizeof(seg_data1
) + 0xfff) >> 12;
1208 ldt
.contents
= MODIFY_LDT_CONTENTS_DATA
;
1209 ldt
.read_exec_only
= 0;
1210 ldt
.limit_in_pages
= 1;
1211 ldt
.seg_not_present
= 0;
1213 modify_ldt(1, &ldt
, sizeof(ldt
)); /* write ldt entry */
1215 ldt
.entry_number
= 2;
1216 ldt
.base_addr
= (unsigned long)&seg_data2
;
1217 ldt
.limit
= (sizeof(seg_data2
) + 0xfff) >> 12;
1219 ldt
.contents
= MODIFY_LDT_CONTENTS_DATA
;
1220 ldt
.read_exec_only
= 0;
1221 ldt
.limit_in_pages
= 1;
1222 ldt
.seg_not_present
= 0;
1224 modify_ldt(1, &ldt
, sizeof(ldt
)); /* write ldt entry */
1226 modify_ldt(0, &ldt_table
, sizeof(ldt_table
)); /* read ldt entries */
1231 printf("%d: %016Lx\n", i
, ldt_table
[i
]);
1234 /* do some tests with fs or gs */
1235 asm volatile ("movl %0, %%fs" : : "r" (MK_SEL(1)));
1237 seg_data1
[1] = 0xaa;
1238 seg_data2
[1] = 0x55;
1240 asm volatile ("fs movzbl 0x1, %0" : "=r" (res
));
1241 printf("FS[1] = %02x\n", res
);
1243 asm volatile ("pushl %%gs\n"
1245 "gs movzbl 0x1, %0\n"
1249 printf("GS[1] = %02x\n", res
);
1251 /* tests with ds/ss (implicit segment case) */
1253 asm volatile ("pushl %%ebp\n\t"
1256 "movl %3, %%ebp\n\t"
1257 "movzbl 0x1, %0\n\t"
1258 "movzbl (%%ebp), %1\n\t"
1261 : "=r" (res
), "=r" (res2
)
1262 : "r" (MK_SEL(1)), "r" (&tmp
));
1263 printf("DS[1] = %02x\n", res
);
1264 printf("SS[tmp] = %02x\n", res2
);
1266 segoff
.seg
= MK_SEL(2);
1267 segoff
.offset
= 0xabcdef12;
1268 asm volatile("lfs %2, %0\n\t"
1270 : "=r" (res
), "=g" (res2
)
1272 printf("FS:reg = %04x:%08x\n", res2
, res
);
1274 TEST_LR("larw", "w", MK_SEL(2), 0x0100);
1275 TEST_LR("larl", "", MK_SEL(2), 0x0100);
1276 TEST_LR("lslw", "w", MK_SEL(2), 0);
1277 TEST_LR("lsll", "", MK_SEL(2), 0);
1279 TEST_LR("larw", "w", 0xfff8, 0);
1280 TEST_LR("larl", "", 0xfff8, 0);
1281 TEST_LR("lslw", "w", 0xfff8, 0);
1282 TEST_LR("lsll", "", 0xfff8, 0);
1285 /* 16 bit code test */
1286 extern char code16_start
, code16_end
;
1287 extern char code16_func1
;
1288 extern char code16_func2
;
1289 extern char code16_func3
;
1291 void test_code16(void)
1293 struct modify_ldt_ldt_s ldt
;
1296 /* build a code segment */
1297 ldt
.entry_number
= 1;
1298 ldt
.base_addr
= (unsigned long)&code16_start
;
1299 ldt
.limit
= &code16_end
- &code16_start
;
1301 ldt
.contents
= MODIFY_LDT_CONTENTS_CODE
;
1302 ldt
.read_exec_only
= 0;
1303 ldt
.limit_in_pages
= 0;
1304 ldt
.seg_not_present
= 0;
1306 modify_ldt(1, &ldt
, sizeof(ldt
)); /* write ldt entry */
1308 /* call the first function */
1309 asm volatile ("lcall %1, %2"
1311 : "i" (MK_SEL(1)), "i" (&code16_func1
): "memory", "cc");
1312 printf("func1() = 0x%08x\n", res
);
1313 asm volatile ("lcall %2, %3"
1314 : "=a" (res
), "=c" (res2
)
1315 : "i" (MK_SEL(1)), "i" (&code16_func2
): "memory", "cc");
1316 printf("func2() = 0x%08x spdec=%d\n", res
, res2
);
1317 asm volatile ("lcall %1, %2"
1319 : "i" (MK_SEL(1)), "i" (&code16_func3
): "memory", "cc");
1320 printf("func3() = 0x%08x\n", res
);
1324 #if defined(__x86_64__)
1325 asm(".globl func_lret\n"
1327 "movl $0x87654641, %eax\n"
1330 asm(".globl func_lret\n"
1332 "movl $0x87654321, %eax\n"
1335 ".globl func_iret\n"
1337 "movl $0xabcd4321, %eax\n"
1341 extern char func_lret
;
1342 extern char func_iret
;
1344 void test_misc(void)
1349 for(i
=0;i
<256;i
++) table
[i
] = 256 - i
;
1351 asm ("xlat" : "=a" (res
) : "b" (table
), "0" (res
));
1352 printf("xlat: EAX=" FMTLX
"\n", res
);
1354 #if defined(__x86_64__)
1356 static struct __attribute__((packed
)) {
1362 asm volatile ("mov %%cs, %0" : "=r" (cs_sel
));
1364 asm volatile ("push %1\n"
1367 : "r" (cs_sel
) : "memory", "cc");
1368 printf("func_lret=" FMTLX
"\n", res
);
1370 /* NOTE: we assume that &func_lret < 4GB */
1371 desc
.offset
= (long)&func_lret
;
1374 asm volatile ("xor %%rax, %%rax\n"
1379 printf("func_lret2=" FMTLX
"\n", res
);
1381 asm volatile ("push %2\n"
1387 : "m" (desc
), "b" (cs_sel
)
1389 printf("func_lret3=" FMTLX
"\n", res
);
1392 asm volatile ("push %%cs ; call %1"
1394 : "m" (func_lret
): "memory", "cc");
1395 printf("func_lret=" FMTLX
"\n", res
);
1397 asm volatile ("pushf ; push %%cs ; call %1"
1399 : "m" (func_iret
): "memory", "cc");
1400 printf("func_iret=" FMTLX
"\n", res
);
1403 #if defined(__x86_64__)
1404 /* specific popl test */
1405 asm volatile ("push $12345432 ; push $0x9abcdef ; pop (%%rsp) ; pop %0"
1407 printf("popl esp=" FMTLX
"\n", res
);
1409 /* specific popl test */
1410 asm volatile ("pushl $12345432 ; pushl $0x9abcdef ; popl (%%esp) ; popl %0"
1412 printf("popl esp=" FMTLX
"\n", res
);
1414 /* specific popw test */
1415 asm volatile ("pushl $12345432 ; pushl $0x9abcdef ; popw (%%esp) ; addl $2, %%esp ; popl %0"
1417 printf("popw esp=" FMTLX
"\n", res
);
1421 uint8_t str_buffer
[4096];
1423 #define TEST_STRING1(OP, size, DF, REP)\
1425 long esi, edi, eax, ecx, eflags;\
1427 esi = (long)(str_buffer + sizeof(str_buffer) / 2);\
1428 edi = (long)(str_buffer + sizeof(str_buffer) / 2) + 16;\
1429 eax = i2l(0x12345678);\
1432 asm volatile ("push $0\n\t"\
1435 REP #OP size "\n\t"\
1439 : "=S" (esi), "=D" (edi), "=a" (eax), "=c" (ecx), "=g" (eflags)\
1440 : "0" (esi), "1" (edi), "2" (eax), "3" (ecx));\
1441 printf("%-10s ESI=" FMTLX " EDI=" FMTLX " EAX=" FMTLX " ECX=" FMTLX " EFL=%04x\n",\
1442 REP #OP size, esi, edi, eax, ecx,\
1443 (int)(eflags & (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A)));\
1446 #define TEST_STRING(OP, REP)\
1447 TEST_STRING1(OP, "b", "", REP);\
1448 TEST_STRING1(OP, "w", "", REP);\
1449 TEST_STRING1(OP, "l", "", REP);\
1450 X86_64_ONLY(TEST_STRING1(OP, "q", "", REP));\
1451 TEST_STRING1(OP, "b", "std", REP);\
1452 TEST_STRING1(OP, "w", "std", REP);\
1453 TEST_STRING1(OP, "l", "std", REP);\
1454 X86_64_ONLY(TEST_STRING1(OP, "q", "std", REP))
1456 void test_string(void)
1459 for(i
= 0;i
< sizeof(str_buffer
); i
++)
1460 str_buffer
[i
] = i
+ 0x56;
1461 TEST_STRING(stos
, "");
1462 TEST_STRING(stos
, "rep ");
1463 TEST_STRING(lods
, ""); /* to verify stos */
1464 TEST_STRING(lods
, "rep ");
1465 TEST_STRING(movs
, "");
1466 TEST_STRING(movs
, "rep ");
1467 TEST_STRING(lods
, ""); /* to verify stos */
1469 /* XXX: better tests */
1470 TEST_STRING(scas
, "");
1471 TEST_STRING(scas
, "repz ");
1472 TEST_STRING(scas
, "repnz ");
1473 TEST_STRING(cmps
, "");
1474 TEST_STRING(cmps
, "repz ");
1475 TEST_STRING(cmps
, "repnz ");
1481 static inline void set_bit(uint8_t *a
, unsigned int bit
)
1483 a
[bit
/ 8] |= (1 << (bit
% 8));
1486 static inline uint8_t *seg_to_linear(unsigned int seg
, unsigned int reg
)
1488 return (uint8_t *)((seg
<< 4) + (reg
& 0xffff));
1491 static inline void pushw(struct vm86_regs
*r
, int val
)
1493 r
->esp
= (r
->esp
& ~0xffff) | ((r
->esp
- 2) & 0xffff);
1494 *(uint16_t *)seg_to_linear(r
->ss
, r
->esp
) = val
;
1497 #undef __syscall_return
1498 #define __syscall_return(type, res) \
1500 return (type) (res); \
1503 _syscall2(int, vm86
, int, func
, struct vm86plus_struct
*, v86
)
1505 extern char vm86_code_start
;
1506 extern char vm86_code_end
;
1508 #define VM86_CODE_CS 0x100
1509 #define VM86_CODE_IP 0x100
1511 void test_vm86(void)
1513 struct vm86plus_struct ctx
;
1514 struct vm86_regs
*r
;
1518 vm86_mem
= mmap((void *)0x00000000, 0x110000,
1519 PROT_WRITE
| PROT_READ
| PROT_EXEC
,
1520 MAP_FIXED
| MAP_ANON
| MAP_PRIVATE
, -1, 0);
1521 if (vm86_mem
== MAP_FAILED
) {
1522 printf("ERROR: could not map vm86 memory");
1525 memset(&ctx
, 0, sizeof(ctx
));
1527 /* init basic registers */
1529 r
->eip
= VM86_CODE_IP
;
1538 r
->eflags
= VIF_MASK
;
1540 /* move code to proper address. We use the same layout as a .com
1542 memcpy(vm86_mem
+ (VM86_CODE_CS
<< 4) + VM86_CODE_IP
,
1543 &vm86_code_start
, &vm86_code_end
- &vm86_code_start
);
1545 /* mark int 0x21 as being emulated */
1546 set_bit((uint8_t *)&ctx
.int_revectored
, 0x21);
1549 ret
= vm86(VM86_ENTER
, &ctx
);
1550 switch(VM86_TYPE(ret
)) {
1555 int_num
= VM86_ARG(ret
);
1556 if (int_num
!= 0x21)
1558 ah
= (r
->eax
>> 8) & 0xff;
1560 case 0x00: /* exit */
1562 case 0x02: /* write char */
1568 case 0x09: /* write string */
1571 ptr
= seg_to_linear(r
->ds
, r
->edx
);
1578 r
->eax
= (r
->eax
& ~0xff) | '$';
1581 case 0xff: /* extension: write eflags number in edx */
1583 #ifndef LINUX_VM86_IOPL_FIX
1586 printf("%08x\n", v
);
1590 printf("unsupported int 0x%02x\n", int_num
);
1596 /* a signal came, we just ignore that */
1601 printf("ERROR: unhandled vm86 return code (0x%x)\n", ret
);
1606 printf("VM86 end\n");
1607 munmap(vm86_mem
, 0x110000);
1611 /* exception tests */
1612 #if defined(__i386__) && !defined(REG_EAX)
1623 #define REG_TRAPNO TRAPNO
1627 #if defined(__x86_64__)
1628 #define REG_EIP REG_RIP
1635 void sig_handler(int sig
, siginfo_t
*info
, void *puc
)
1637 struct ucontext
*uc
= puc
;
1639 printf("si_signo=%d si_errno=%d si_code=%d",
1640 info
->si_signo
, info
->si_errno
, info
->si_code
);
1641 printf(" si_addr=0x%08lx",
1642 (unsigned long)info
->si_addr
);
1645 printf("trapno=" FMTLX
" err=" FMTLX
,
1646 (long)uc
->uc_mcontext
.gregs
[REG_TRAPNO
],
1647 (long)uc
->uc_mcontext
.gregs
[REG_ERR
]);
1648 printf(" EIP=" FMTLX
, (long)uc
->uc_mcontext
.gregs
[REG_EIP
]);
1650 longjmp(jmp_env
, 1);
1653 void test_exceptions(void)
1655 struct sigaction act
;
1658 act
.sa_sigaction
= sig_handler
;
1659 sigemptyset(&act
.sa_mask
);
1660 act
.sa_flags
= SA_SIGINFO
| SA_NODEFER
;
1661 sigaction(SIGFPE
, &act
, NULL
);
1662 sigaction(SIGILL
, &act
, NULL
);
1663 sigaction(SIGSEGV
, &act
, NULL
);
1664 sigaction(SIGBUS
, &act
, NULL
);
1665 sigaction(SIGTRAP
, &act
, NULL
);
1667 /* test division by zero reporting */
1668 printf("DIVZ exception:\n");
1669 if (setjmp(jmp_env
) == 0) {
1670 /* now divide by zero */
1675 #if !defined(__x86_64__)
1676 printf("BOUND exception:\n");
1677 if (setjmp(jmp_env
) == 0) {
1678 /* bound exception */
1681 asm volatile ("bound %0, %1" : : "r" (11), "m" (tab
[0]));
1686 printf("segment exceptions:\n");
1687 if (setjmp(jmp_env
) == 0) {
1688 /* load an invalid segment */
1689 asm volatile ("movl %0, %%fs" : : "r" ((0x1234 << 3) | 1));
1691 if (setjmp(jmp_env
) == 0) {
1692 /* null data segment is valid */
1693 asm volatile ("movl %0, %%fs" : : "r" (3));
1694 /* null stack segment */
1695 asm volatile ("movl %0, %%ss" : : "r" (3));
1699 struct modify_ldt_ldt_s ldt
;
1700 ldt
.entry_number
= 1;
1701 ldt
.base_addr
= (unsigned long)&seg_data1
;
1702 ldt
.limit
= (sizeof(seg_data1
) + 0xfff) >> 12;
1704 ldt
.contents
= MODIFY_LDT_CONTENTS_DATA
;
1705 ldt
.read_exec_only
= 0;
1706 ldt
.limit_in_pages
= 1;
1707 ldt
.seg_not_present
= 1;
1709 modify_ldt(1, &ldt
, sizeof(ldt
)); /* write ldt entry */
1711 if (setjmp(jmp_env
) == 0) {
1712 /* segment not present */
1713 asm volatile ("movl %0, %%fs" : : "r" (MK_SEL(1)));
1718 /* test SEGV reporting */
1719 printf("PF exception:\n");
1720 if (setjmp(jmp_env
) == 0) {
1722 /* we add a nop to test a weird PC retrieval case */
1723 asm volatile ("nop");
1724 /* now store in an invalid address */
1725 *(char *)0x1234 = 1;
1728 /* test SEGV reporting */
1729 printf("PF exception:\n");
1730 if (setjmp(jmp_env
) == 0) {
1732 /* read from an invalid address */
1733 v1
= *(char *)0x1234;
1736 /* test illegal instruction reporting */
1737 printf("UD2 exception:\n");
1738 if (setjmp(jmp_env
) == 0) {
1739 /* now execute an invalid instruction */
1740 asm volatile("ud2");
1742 printf("lock nop exception:\n");
1743 if (setjmp(jmp_env
) == 0) {
1744 /* now execute an invalid instruction */
1745 asm volatile("lock nop");
1748 printf("INT exception:\n");
1749 if (setjmp(jmp_env
) == 0) {
1750 asm volatile ("int $0xfd");
1752 if (setjmp(jmp_env
) == 0) {
1753 asm volatile ("int $0x01");
1755 if (setjmp(jmp_env
) == 0) {
1756 asm volatile (".byte 0xcd, 0x03");
1758 if (setjmp(jmp_env
) == 0) {
1759 asm volatile ("int $0x04");
1761 if (setjmp(jmp_env
) == 0) {
1762 asm volatile ("int $0x05");
1765 printf("INT3 exception:\n");
1766 if (setjmp(jmp_env
) == 0) {
1767 asm volatile ("int3");
1770 printf("CLI exception:\n");
1771 if (setjmp(jmp_env
) == 0) {
1772 asm volatile ("cli");
1775 printf("STI exception:\n");
1776 if (setjmp(jmp_env
) == 0) {
1777 asm volatile ("cli");
1780 #if !defined(__x86_64__)
1781 printf("INTO exception:\n");
1782 if (setjmp(jmp_env
) == 0) {
1783 /* overflow exception */
1784 asm volatile ("addl $1, %0 ; into" : : "r" (0x7fffffff));
1788 printf("OUTB exception:\n");
1789 if (setjmp(jmp_env
) == 0) {
1790 asm volatile ("outb %%al, %%dx" : : "d" (0x4321), "a" (0));
1793 printf("INB exception:\n");
1794 if (setjmp(jmp_env
) == 0) {
1795 asm volatile ("inb %%dx, %%al" : "=a" (val
) : "d" (0x4321));
1798 printf("REP OUTSB exception:\n");
1799 if (setjmp(jmp_env
) == 0) {
1800 asm volatile ("rep outsb" : : "d" (0x4321), "S" (tab
), "c" (1));
1803 printf("REP INSB exception:\n");
1804 if (setjmp(jmp_env
) == 0) {
1805 asm volatile ("rep insb" : : "d" (0x4321), "D" (tab
), "c" (1));
1808 printf("HLT exception:\n");
1809 if (setjmp(jmp_env
) == 0) {
1810 asm volatile ("hlt");
1813 printf("single step exception:\n");
1815 if (setjmp(jmp_env
) == 0) {
1816 asm volatile ("pushf\n"
1817 "orl $0x00100, (%%esp)\n"
1819 "movl $0xabcd, %0\n"
1820 "movl $0x0, %0\n" : "=m" (val
) : : "cc", "memory");
1822 printf("val=0x%x\n", val
);
1825 #if !defined(__x86_64__)
1826 /* specific precise single step test */
1827 void sig_trap_handler(int sig
, siginfo_t
*info
, void *puc
)
1829 struct ucontext
*uc
= puc
;
1830 printf("EIP=" FMTLX
"\n", (long)uc
->uc_mcontext
.gregs
[REG_EIP
]);
1833 const uint8_t sstep_buf1
[4] = { 1, 2, 3, 4};
1834 uint8_t sstep_buf2
[4];
1836 void test_single_step(void)
1838 struct sigaction act
;
1843 act
.sa_sigaction
= sig_trap_handler
;
1844 sigemptyset(&act
.sa_mask
);
1845 act
.sa_flags
= SA_SIGINFO
;
1846 sigaction(SIGTRAP
, &act
, NULL
);
1847 asm volatile ("pushf\n"
1848 "orl $0x00100, (%%esp)\n"
1850 "movl $0xabcd, %0\n"
1859 /* movsb: the single step should stop at each movsb iteration */
1860 "movl $sstep_buf1, %%esi\n"
1861 "movl $sstep_buf2, %%edi\n"
1869 /* cmpsb: the single step should stop at each cmpsb iteration */
1870 "movl $sstep_buf1, %%esi\n"
1871 "movl $sstep_buf2, %%edi\n"
1877 /* getpid() syscall: single step should skip one
1883 /* when modifying SS, trace is not done on the next
1885 "movl %%ss, %%ecx\n"
1886 "movl %%ecx, %%ss\n"
1889 "movl %%ecx, %%ss\n"
1900 "andl $~0x00100, (%%esp)\n"
1904 : "cc", "memory", "eax", "ecx", "esi", "edi");
1905 printf("val=%d\n", val
);
1906 for(i
= 0; i
< 4; i
++)
1907 printf("sstep_buf2[%d] = %d\n", i
, sstep_buf2
[i
]);
1910 /* self modifying code test */
1912 0xb8, 0x1, 0x00, 0x00, 0x00, /* movl $1, %eax */
1917 "movl 4(%esp), %eax\n"
1918 "movl %eax, smc_patch_addr2 + 1\n"
1927 "smc_patch_addr2:\n"
1931 typedef int FuncType(void);
1932 extern int smc_code2(int);
1933 void test_self_modifying_code(void)
1937 printf("self modifying code:\n");
1938 printf("func1 = 0x%x\n", ((FuncType
*)code
)());
1939 for(i
= 2; i
<= 4; i
++) {
1941 printf("func%d = 0x%x\n", i
, ((FuncType
*)code
)());
1944 /* more difficult test : the modified code is just after the
1945 modifying instruction. It is forbidden in Intel specs, but it
1946 is used by old DOS programs */
1947 for(i
= 2; i
<= 4; i
++) {
1948 printf("smc_code2(%d) = %d\n", i
, smc_code2(i
));
1953 long enter_stack
[4096];
1955 #if defined(__x86_64__)
1963 #define TEST_ENTER(size, stack_type, level)\
1965 long esp_save, esp_val, ebp_val, ebp_save, i;\
1966 stack_type *ptr, *stack_end, *stack_ptr;\
1967 memset(enter_stack, 0, sizeof(enter_stack));\
1968 stack_end = stack_ptr = (stack_type *)(enter_stack + 4096);\
1969 ebp_val = (long)stack_ptr;\
1972 esp_val = (long)stack_ptr;\
1973 asm("mov " RSP ", %[esp_save]\n"\
1974 "mov " RBP ", %[ebp_save]\n"\
1975 "mov %[esp_val], " RSP "\n"\
1976 "mov %[ebp_val], " RBP "\n"\
1977 "enter" size " $8, $" #level "\n"\
1978 "mov " RSP ", %[esp_val]\n"\
1979 "mov " RBP ", %[ebp_val]\n"\
1980 "mov %[esp_save], " RSP "\n"\
1981 "mov %[ebp_save], " RBP "\n"\
1982 : [esp_save] "=r" (esp_save),\
1983 [ebp_save] "=r" (ebp_save),\
1984 [esp_val] "=r" (esp_val),\
1985 [ebp_val] "=r" (ebp_val)\
1986 : "[esp_val]" (esp_val),\
1987 "[ebp_val]" (ebp_val));\
1988 printf("level=%d:\n", level);\
1989 printf("esp_val=" FMTLX "\n", esp_val - (long)stack_end);\
1990 printf("ebp_val=" FMTLX "\n", ebp_val - (long)stack_end);\
1991 for(ptr = (stack_type *)esp_val; ptr < stack_end; ptr++)\
1992 printf(FMTLX "\n", (long)ptr[0]);\
1995 static void test_enter(void)
1997 #if defined(__x86_64__)
1998 TEST_ENTER("q", uint64_t, 0);
1999 TEST_ENTER("q", uint64_t, 1);
2000 TEST_ENTER("q", uint64_t, 2);
2001 TEST_ENTER("q", uint64_t, 31);
2003 TEST_ENTER("l", uint32_t, 0);
2004 TEST_ENTER("l", uint32_t, 1);
2005 TEST_ENTER("l", uint32_t, 2);
2006 TEST_ENTER("l", uint32_t, 31);
2009 TEST_ENTER("w", uint16_t, 0);
2010 TEST_ENTER("w", uint16_t, 1);
2011 TEST_ENTER("w", uint16_t, 2);
2012 TEST_ENTER("w", uint16_t, 31);
2017 typedef int __m64
__attribute__ ((__mode__ (__V2SI__
)));
2018 typedef int __m128
__attribute__ ((__mode__(__V4SF__
)));
2028 static uint64_t __attribute__((aligned(16))) test_values
[4][2] = {
2029 { 0x456723c698694873, 0xdc515cff944a58ec },
2030 { 0x1f297ccd58bad7ab, 0x41f21efba9e3e146 },
2031 { 0x007c62c2085427f8, 0x231be9e8cde7438d },
2032 { 0x0f76255a085427f8, 0xc233e9e8c4c9439a },
2037 asm volatile (#op " %2, %0" : "=x" (r.dq) : "0" (a.dq), "x" (b.dq));\
2038 printf("%-9s: a=" FMT64X "" FMT64X " b=" FMT64X "" FMT64X " r=" FMT64X "" FMT64X "\n",\
2045 #define SSE_OP2(op)\
2049 a.q[0] = test_values[2*i][0];\
2050 a.q[1] = test_values[2*i][1];\
2051 b.q[0] = test_values[2*i+1][0];\
2052 b.q[1] = test_values[2*i+1][1];\
2057 #define MMX_OP2(op)\
2061 a.q[0] = test_values[2*i][0];\
2062 b.q[0] = test_values[2*i+1][0];\
2063 asm volatile (#op " %2, %0" : "=y" (r.q[0]) : "0" (a.q[0]), "y" (b.q[0]));\
2064 printf("%-9s: a=" FMT64X " b=" FMT64X " r=" FMT64X "\n",\
2073 #define SHUF_OP(op, ib)\
2075 a.q[0] = test_values[0][0];\
2076 a.q[1] = test_values[0][1];\
2077 b.q[0] = test_values[1][0];\
2078 b.q[1] = test_values[1][1];\
2079 asm volatile (#op " $" #ib ", %2, %0" : "=x" (r.dq) : "0" (a.dq), "x" (b.dq));\
2080 printf("%-9s: a=" FMT64X "" FMT64X " b=" FMT64X "" FMT64X " ib=%02x r=" FMT64X "" FMT64X "\n",\
2088 #define PSHUF_OP(op, ib)\
2092 a.q[0] = test_values[2*i][0];\
2093 a.q[1] = test_values[2*i][1];\
2094 asm volatile (#op " $" #ib ", %1, %0" : "=x" (r.dq) : "x" (a.dq));\
2095 printf("%-9s: a=" FMT64X "" FMT64X " ib=%02x r=" FMT64X "" FMT64X "\n",\
2103 #define SHIFT_IM(op, ib)\
2107 a.q[0] = test_values[2*i][0];\
2108 a.q[1] = test_values[2*i][1];\
2109 asm volatile (#op " $" #ib ", %0" : "=x" (r.dq) : "0" (a.dq));\
2110 printf("%-9s: a=" FMT64X "" FMT64X " ib=%02x r=" FMT64X "" FMT64X "\n",\
2118 #define SHIFT_OP(op, ib)\
2123 a.q[0] = test_values[2*i][0];\
2124 a.q[1] = test_values[2*i][1];\
2127 asm volatile (#op " %2, %0" : "=x" (r.dq) : "0" (a.dq), "x" (b.dq));\
2128 printf("%-9s: a=" FMT64X "" FMT64X " b=" FMT64X "" FMT64X " r=" FMT64X "" FMT64X "\n",\
2140 a.q[0] = test_values[2*i][0];\
2141 a.q[1] = test_values[2*i][1];\
2142 asm volatile (#op " %1, %0" : "=r" (reg) : "x" (a.dq));\
2143 printf("%-9s: a=" FMT64X "" FMT64X " r=%08x\n",\
2150 #define SSE_OPS(a) \
2154 #define SSE_OPD(a) \
2158 #define SSE_COMI(op, field)\
2160 unsigned int eflags;\
2164 asm volatile (#op " %2, %1\n"\
2168 : "x" (a.dq), "x" (b.dq));\
2169 printf("%-9s: a=%f b=%f cc=%04x\n",\
2171 eflags & (CC_C | CC_P | CC_Z | CC_S | CC_O | CC_A));\
2174 void test_sse_comi(double a1
, double b1
)
2176 SSE_COMI(ucomiss
, s
);
2177 SSE_COMI(ucomisd
, d
);
2178 SSE_COMI(comiss
, s
);
2179 SSE_COMI(comisd
, d
);
2182 #define CVT_OP_XMM(op)\
2184 asm volatile (#op " %1, %0" : "=x" (r.dq) : "x" (a.dq));\
2185 printf("%-9s: a=" FMT64X "" FMT64X " r=" FMT64X "" FMT64X "\n",\
2191 #define CVT_OP_XMM2MMX(op)\
2193 asm volatile (#op " %1, %0" : "=y" (r.q[0]) : "x" (a.dq));\
2194 printf("%-9s: a=" FMT64X "" FMT64X " r=" FMT64X "\n",\
2200 #define CVT_OP_MMX2XMM(op)\
2202 asm volatile (#op " %1, %0" : "=x" (r.dq) : "y" (a.q[0]));\
2203 printf("%-9s: a=" FMT64X " r=" FMT64X "" FMT64X "\n",\
2209 #define CVT_OP_REG2XMM(op)\
2211 asm volatile (#op " %1, %0" : "=x" (r.dq) : "r" (a.l[0]));\
2212 printf("%-9s: a=%08x r=" FMT64X "" FMT64X "\n",\
2218 #define CVT_OP_XMM2REG(op)\
2220 asm volatile (#op " %1, %0" : "=r" (r.l[0]) : "x" (a.dq));\
2221 printf("%-9s: a=" FMT64X "" FMT64X " r=%08x\n",\
2239 uint32_t mxcsr_mask
;
2240 uint8_t fpregs1
[8 * 16];
2241 uint8_t xmm_regs
[8 * 16];
2242 uint8_t dummy2
[224];
2245 static struct fpxstate fpx_state
__attribute__((aligned(16)));
2246 static struct fpxstate fpx_state2
__attribute__((aligned(16)));
2248 void test_fxsave(void)
2250 struct fpxstate
*fp
= &fpx_state
;
2251 struct fpxstate
*fp2
= &fpx_state2
;
2254 a
.q
[0] = test_values
[0][0];
2255 a
.q
[1] = test_values
[0][1];
2256 b
.q
[0] = test_values
[1][0];
2257 b
.q
[1] = test_values
[1][1];
2259 asm("movdqa %2, %%xmm0\n"
2260 "movdqa %3, %%xmm7\n"
2261 #if defined(__x86_64__)
2262 "movdqa %2, %%xmm15\n"
2271 : "=m" (*(uint32_t *)fp2
), "=m" (*(uint32_t *)fp
)
2272 : "m" (a
), "m" (b
));
2273 printf("fpuc=%04x\n", fp
->fpuc
);
2274 printf("fpus=%04x\n", fp
->fpus
);
2275 printf("fptag=%04x\n", fp
->fptag
);
2276 for(i
= 0; i
< 3; i
++) {
2277 printf("ST%d: " FMT64X
" %04x\n",
2279 *(uint64_t *)&fp
->fpregs1
[i
* 16],
2280 *(uint16_t *)&fp
->fpregs1
[i
* 16 + 8]);
2282 printf("mxcsr=%08x\n", fp
->mxcsr
& 0x1f80);
2283 #if defined(__x86_64__)
2288 for(i
= 0; i
< nb_xmm
; i
++) {
2289 printf("xmm%d: " FMT64X
"" FMT64X
"\n",
2291 *(uint64_t *)&fp
->xmm_regs
[i
* 16],
2292 *(uint64_t *)&fp
->xmm_regs
[i
* 16 + 8]);
2353 asm volatile ("pinsrw $1, %1, %0" : "=y" (r
.q
[0]) : "r" (0x12345678));
2354 printf("%-9s: r=" FMT64X
"\n", "pinsrw", r
.q
[0]);
2356 asm volatile ("pinsrw $5, %1, %0" : "=x" (r
.dq
) : "r" (0x12345678));
2357 printf("%-9s: r=" FMT64X
"" FMT64X
"\n", "pinsrw", r
.q
[1], r
.q
[0]);
2359 a
.q
[0] = test_values
[0][0];
2360 a
.q
[1] = test_values
[0][1];
2361 asm volatile ("pextrw $1, %1, %0" : "=r" (r
.l
[0]) : "y" (a
.q
[0]));
2362 printf("%-9s: r=%08x\n", "pextrw", r
.l
[0]);
2364 asm volatile ("pextrw $5, %1, %0" : "=r" (r
.l
[0]) : "x" (a
.dq
));
2365 printf("%-9s: r=%08x\n", "pextrw", r
.l
[0]);
2367 asm volatile ("pmovmskb %1, %0" : "=r" (r
.l
[0]) : "y" (a
.q
[0]));
2368 printf("%-9s: r=%08x\n", "pmovmskb", r
.l
[0]);
2370 asm volatile ("pmovmskb %1, %0" : "=r" (r
.l
[0]) : "x" (a
.dq
));
2371 printf("%-9s: r=%08x\n", "pmovmskb", r
.l
[0]);
2377 a
.q
[0] = test_values
[0][0];
2378 a
.q
[1] = test_values
[0][1];
2379 b
.q
[0] = test_values
[1][0];
2380 b
.q
[1] = test_values
[1][1];
2381 asm volatile("maskmovq %1, %0" :
2382 : "y" (a
.q
[0]), "y" (b
.q
[0]), "D" (&r
)
2384 printf("%-9s: r=" FMT64X
" a=" FMT64X
" b=" FMT64X
"\n",
2389 asm volatile("maskmovdqu %1, %0" :
2390 : "x" (a
.dq
), "x" (b
.dq
), "D" (&r
)
2392 printf("%-9s: r=" FMT64X
"" FMT64X
" a=" FMT64X
"" FMT64X
" b=" FMT64X
"" FMT64X
"\n",
2399 asm volatile ("emms");
2401 SSE_OP2(punpcklqdq
);
2402 SSE_OP2(punpckhqdq
);
2417 SHUF_OP(shufps
, 0x78);
2418 SHUF_OP(shufpd
, 0x02);
2420 PSHUF_OP(pshufd
, 0x78);
2421 PSHUF_OP(pshuflw
, 0x78);
2422 PSHUF_OP(pshufhw
, 0x78);
2425 SHIFT_OP(psrlw
, 16);
2427 SHIFT_OP(psraw
, 16);
2429 SHIFT_OP(psllw
, 16);
2432 SHIFT_OP(psrld
, 32);
2434 SHIFT_OP(psrad
, 32);
2436 SHIFT_OP(pslld
, 32);
2439 SHIFT_OP(psrlq
, 32);
2441 SHIFT_OP(psllq
, 32);
2443 SHIFT_IM(psrldq
, 16);
2444 SHIFT_IM(psrldq
, 7);
2445 SHIFT_IM(pslldq
, 16);
2446 SHIFT_IM(pslldq
, 7);
2451 /* FPU specific ops */
2455 asm volatile("stmxcsr %0" : "=m" (mxcsr
));
2456 printf("mxcsr=%08x\n", mxcsr
& 0x1f80);
2457 asm volatile("ldmxcsr %0" : : "m" (mxcsr
));
2460 test_sse_comi(2, -1);
2461 test_sse_comi(2, 2);
2462 test_sse_comi(2, 3);
2463 test_sse_comi(2, q_nan
.d
);
2464 test_sse_comi(q_nan
.d
, -1);
2466 for(i
= 0; i
< 2; i
++) {
2522 /* float to float/int */
2527 CVT_OP_XMM(cvtps2pd
);
2528 CVT_OP_XMM(cvtss2sd
);
2529 CVT_OP_XMM2MMX(cvtps2pi
);
2530 CVT_OP_XMM2MMX(cvttps2pi
);
2531 CVT_OP_XMM2REG(cvtss2si
);
2532 CVT_OP_XMM2REG(cvttss2si
);
2533 CVT_OP_XMM(cvtps2dq
);
2534 CVT_OP_XMM(cvttps2dq
);
2538 CVT_OP_XMM(cvtpd2ps
);
2539 CVT_OP_XMM(cvtsd2ss
);
2540 CVT_OP_XMM2MMX(cvtpd2pi
);
2541 CVT_OP_XMM2MMX(cvttpd2pi
);
2542 CVT_OP_XMM2REG(cvtsd2si
);
2543 CVT_OP_XMM2REG(cvttsd2si
);
2544 CVT_OP_XMM(cvtpd2dq
);
2545 CVT_OP_XMM(cvttpd2dq
);
2552 CVT_OP_MMX2XMM(cvtpi2ps
);
2553 CVT_OP_MMX2XMM(cvtpi2pd
);
2554 CVT_OP_REG2XMM(cvtsi2ss
);
2555 CVT_OP_REG2XMM(cvtsi2sd
);
2556 CVT_OP_XMM(cvtdq2ps
);
2557 CVT_OP_XMM(cvtdq2pd
);
2559 /* XXX: test PNI insns */
2563 asm volatile ("emms");
2568 extern void *__start_initcall
;
2569 extern void *__stop_initcall
;
2572 int main(int argc
, char **argv
)
2577 ptr
= &__start_initcall
;
2578 while (ptr
!= &__stop_initcall
) {
2586 #if !defined(__x86_64__)
2601 #if !defined(__x86_64__)
2602 test_self_modifying_code();