2 * MMX/3DNow!/SSE/SSE2/SSE3/SSSE3/SSE4/PNI support
4 * Copyright (c) 2005 Fabrice Bellard
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.1 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 #include "crypto/aes.h"
35 #define XMM_ONLY(...) __VA_ARGS__
43 #define LANE_WIDTH (SHIFT ? 16 : 8)
44 #define PACK_WIDTH (LANE_WIDTH / 2)
47 * Copy the relevant parts of a Reg value around. In the case where
48 * sizeof(Reg) > SIZE, these helpers operate only on the lower bytes of
49 * a 64 byte ZMMReg, so we must copy only those and keep the top bytes
50 * untouched in the guest-visible destination destination register.
51 * Note that the "lower bytes" are placed last in memory on big-endian
52 * hosts, which store the vector backwards in memory. In that case the
53 * copy *starts* at B(SIZE - 1) and ends at B(0), the opposite of
54 * the little-endian case.
57 #define MOVE(d, r) memcpy(&((d).B(SIZE - 1)), &(r).B(SIZE - 1), SIZE)
59 #define MOVE(d, r) memcpy(&(d).B(0), &(r).B(0), SIZE)
63 #define FPSRL(x, c) ((x) >> shift)
64 #define FPSRAW(x, c) ((int16_t)(x) >> shift)
65 #define FPSRAL(x, c) ((int32_t)(x) >> shift)
66 #define FPSLL(x, c) ((x) << shift)
69 void glue(helper_psrlw, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
74 for (int i = 0; i < 1 << SHIFT; i++) {
79 for (int i = 0; i < 4 << SHIFT; i++) {
80 d->W(i) = FPSRL(s->W(i), shift);
85 void glue(helper_psllw, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
90 for (int i = 0; i < 1 << SHIFT; i++) {
95 for (int i = 0; i < 4 << SHIFT; i++) {
96 d->W(i) = FPSLL(s->W(i), shift);
101 void glue(helper_psraw, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
110 for (int i = 0; i < 4 << SHIFT; i++) {
111 d->W(i) = FPSRAW(s->W(i), shift);
115 void glue(helper_psrld, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
120 for (int i = 0; i < 1 << SHIFT; i++) {
125 for (int i = 0; i < 2 << SHIFT; i++) {
126 d->L(i) = FPSRL(s->L(i), shift);
131 void glue(helper_pslld, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
136 for (int i = 0; i < 1 << SHIFT; i++) {
141 for (int i = 0; i < 2 << SHIFT; i++) {
142 d->L(i) = FPSLL(s->L(i), shift);
147 void glue(helper_psrad, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
156 for (int i = 0; i < 2 << SHIFT; i++) {
157 d->L(i) = FPSRAL(s->L(i), shift);
161 void glue(helper_psrlq, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
166 for (int i = 0; i < 1 << SHIFT; i++) {
171 for (int i = 0; i < 1 << SHIFT; i++) {
172 d->Q(i) = FPSRL(s->Q(i), shift);
177 void glue(helper_psllq, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
182 for (int i = 0; i < 1 << SHIFT; i++) {
187 for (int i = 0; i < 1 << SHIFT; i++) {
188 d->Q(i) = FPSLL(s->Q(i), shift);
194 void glue(helper_psrldq, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
203 for (j = 0; j < 8 << SHIFT; j += LANE_WIDTH) {
204 for (i = 0; i < 16 - shift; i++) {
205 d->B(j + i) = s->B(j + i + shift);
207 for (i = 16 - shift; i < 16; i++) {
213 void glue(helper_pslldq, SUFFIX)(CPUX86State *env, Reg *d, Reg *c)
222 for (j = 0; j < 8 << SHIFT; j += LANE_WIDTH) {
223 for (i = 15; i >= shift; i--) {
224 d->B(j + i) = s->B(j + i - shift);
226 for (i = 0; i < shift; i++) {
233 #define SSE_HELPER_1(name, elem, num, F) \
234 void glue(name, SUFFIX)(CPUX86State *env, Reg *d, Reg *s) \
237 for (int i = 0; i < n; i++) { \
238 d->elem(i) = F(s->elem(i)); \
242 #define SSE_HELPER_2(name, elem, num, F) \
243 void glue(name, SUFFIX)(CPUX86State *env, Reg *d, Reg *s) \
247 for (int i = 0; i < n; i++) { \
248 d->elem(i) = F(v->elem(i), s->elem(i)); \
252 #define SSE_HELPER_B(name, F) \
253 SSE_HELPER_2(name, B, 8 << SHIFT, F)
255 #define SSE_HELPER_W(name, F) \
256 SSE_HELPER_2(name, W, 4 << SHIFT, F)
258 #define SSE_HELPER_L(name, F) \
259 SSE_HELPER_2(name, L, 2 << SHIFT, F)
261 #define SSE_HELPER_Q(name, F) \
262 SSE_HELPER_2(name, Q, 1 << SHIFT, F)
265 static inline int satub(int x)
269 } else if (x > 255) {
276 static inline int satuw(int x)
280 } else if (x > 65535) {
287 static inline int satsb(int x)
291 } else if (x > 127) {
298 static inline int satsw(int x)
302 } else if (x > 32767) {
309 #define FADD(a, b) ((a) + (b))
310 #define FADDUB(a, b) satub((a) + (b))
311 #define FADDUW(a, b) satuw((a) + (b))
312 #define FADDSB(a, b) satsb((int8_t)(a) + (int8_t)(b))
313 #define FADDSW(a, b) satsw((int16_t)(a) + (int16_t)(b))
315 #define FSUB(a, b) ((a) - (b))
316 #define FSUBUB(a, b) satub((a) - (b))
317 #define FSUBUW(a, b) satuw((a) - (b))
318 #define FSUBSB(a, b) satsb((int8_t)(a) - (int8_t)(b))
319 #define FSUBSW(a, b) satsw((int16_t)(a) - (int16_t)(b))
320 #define FMINUB(a, b) ((a) < (b)) ? (a) : (b)
321 #define FMINSW(a, b) ((int16_t)(a) < (int16_t)(b)) ? (a) : (b)
322 #define FMAXUB(a, b) ((a) > (b)) ? (a) : (b)
323 #define FMAXSW(a, b) ((int16_t)(a) > (int16_t)(b)) ? (a) : (b)
325 #define FAND(a, b) ((a) & (b))
326 #define FANDN(a, b) ((~(a)) & (b))
327 #define FOR(a, b) ((a) | (b))
328 #define FXOR(a, b) ((a) ^ (b))
330 #define FCMPGTB(a, b) ((int8_t)(a) > (int8_t)(b) ? -1 : 0)
331 #define FCMPGTW(a, b) ((int16_t)(a) > (int16_t)(b) ? -1 : 0)
332 #define FCMPGTL(a, b) ((int32_t)(a) > (int32_t)(b) ? -1 : 0)
333 #define FCMPEQ(a, b) ((a) == (b) ? -1 : 0)
335 #define FMULLW(a, b) ((a) * (b))
336 #define FMULHRW(a, b) (((int16_t)(a) * (int16_t)(b) + 0x8000) >> 16)
337 #define FMULHUW(a, b) ((a) * (b) >> 16)
338 #define FMULHW(a, b) ((int16_t)(a) * (int16_t)(b) >> 16)
340 #define FAVG(a, b) (((a) + (b) + 1) >> 1)
343 SSE_HELPER_B(helper_paddb, FADD)
344 SSE_HELPER_W(helper_paddw, FADD)
345 SSE_HELPER_L(helper_paddl, FADD)
346 SSE_HELPER_Q(helper_paddq, FADD)
348 SSE_HELPER_B(helper_psubb, FSUB)
349 SSE_HELPER_W(helper_psubw, FSUB)
350 SSE_HELPER_L(helper_psubl, FSUB)
351 SSE_HELPER_Q(helper_psubq, FSUB)
353 SSE_HELPER_B(helper_paddusb, FADDUB)
354 SSE_HELPER_B(helper_paddsb, FADDSB)
355 SSE_HELPER_B(helper_psubusb, FSUBUB)
356 SSE_HELPER_B(helper_psubsb, FSUBSB)
358 SSE_HELPER_W(helper_paddusw, FADDUW)
359 SSE_HELPER_W(helper_paddsw, FADDSW)
360 SSE_HELPER_W(helper_psubusw, FSUBUW)
361 SSE_HELPER_W(helper_psubsw, FSUBSW)
363 SSE_HELPER_B(helper_pminub, FMINUB)
364 SSE_HELPER_B(helper_pmaxub, FMAXUB)
366 SSE_HELPER_W(helper_pminsw, FMINSW)
367 SSE_HELPER_W(helper_pmaxsw, FMAXSW)
369 SSE_HELPER_Q(helper_pand, FAND)
370 SSE_HELPER_Q(helper_pandn, FANDN)
371 SSE_HELPER_Q(helper_por, FOR)
372 SSE_HELPER_Q(helper_pxor, FXOR)
374 SSE_HELPER_B(helper_pcmpgtb, FCMPGTB)
375 SSE_HELPER_W(helper_pcmpgtw, FCMPGTW)
376 SSE_HELPER_L(helper_pcmpgtl, FCMPGTL)
378 SSE_HELPER_B(helper_pcmpeqb, FCMPEQ)
379 SSE_HELPER_W(helper_pcmpeqw, FCMPEQ)
380 SSE_HELPER_L(helper_pcmpeql, FCMPEQ)
382 SSE_HELPER_W(helper_pmullw, FMULLW)
384 SSE_HELPER_W(helper_pmulhrw, FMULHRW)
386 SSE_HELPER_W(helper_pmulhuw, FMULHUW)
387 SSE_HELPER_W(helper_pmulhw, FMULHW)
389 SSE_HELPER_B(helper_pavgb, FAVG)
390 SSE_HELPER_W(helper_pavgw, FAVG)
392 void glue(helper_pmuludq, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
397 for (i = 0; i < (1 << SHIFT); i++) {
398 d->Q(i) = (uint64_t)s->L(i * 2) * (uint64_t)v->L(i * 2);
402 void glue(helper_pmaddwd, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
407 for (i = 0; i < (2 << SHIFT); i++) {
408 d->L(i) = (int16_t)s->W(2 * i) * (int16_t)v->W(2 * i) +
409 (int16_t)s->W(2 * i + 1) * (int16_t)v->W(2 * i + 1);
414 static inline int abs1(int a)
424 void glue(helper_psadbw, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
429 for (i = 0; i < (1 << SHIFT); i++) {
430 unsigned int val = 0;
431 val += abs1(v->B(8 * i + 0) - s->B(8 * i + 0));
432 val += abs1(v->B(8 * i + 1) - s->B(8 * i + 1));
433 val += abs1(v->B(8 * i + 2) - s->B(8 * i + 2));
434 val += abs1(v->B(8 * i + 3) - s->B(8 * i + 3));
435 val += abs1(v->B(8 * i + 4) - s->B(8 * i + 4));
436 val += abs1(v->B(8 * i + 5) - s->B(8 * i + 5));
437 val += abs1(v->B(8 * i + 6) - s->B(8 * i + 6));
438 val += abs1(v->B(8 * i + 7) - s->B(8 * i + 7));
443 void glue(helper_maskmov, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
448 for (i = 0; i < (8 << SHIFT); i++) {
449 if (s->B(i) & 0x80) {
450 cpu_stb_data_ra(env, a0 + i, d->B(i), GETPC());
455 void glue(helper_movl_mm_T0, SUFFIX)(Reg *d, uint32_t val)
461 for (i = 1; i < (1 << SHIFT); i++) {
467 void glue(helper_movq_mm_T0, SUFFIX)(Reg *d, uint64_t val)
472 for (i = 1; i < (1 << SHIFT); i++) {
478 #define SHUFFLE4(F, a, b, offset) do { \
479 r0 = a->F((order & 3) + offset); \
480 r1 = a->F(((order >> 2) & 3) + offset); \
481 r2 = b->F(((order >> 4) & 3) + offset); \
482 r3 = b->F(((order >> 6) & 3) + offset); \
484 d->F(offset + 1) = r1; \
485 d->F(offset + 2) = r2; \
486 d->F(offset + 3) = r3; \
490 void glue(helper_pshufw, SUFFIX)(Reg *d, Reg *s, int order)
492 uint16_t r0, r1, r2, r3;
494 SHUFFLE4(W, s, s, 0);
497 void glue(helper_shufps, SUFFIX)(Reg *d, Reg *s, int order)
500 uint32_t r0, r1, r2, r3;
503 for (i = 0; i < 2 << SHIFT; i += 4) {
504 SHUFFLE4(L, v, s, i);
508 void glue(helper_shufpd, SUFFIX)(Reg *d, Reg *s, int order)
514 for (i = 0; i < 1 << SHIFT; i += 2) {
515 r0 = v->Q(((order & 1) & 1) + i);
516 r1 = s->Q(((order >> 1) & 1) + i);
523 void glue(helper_pshufd, SUFFIX)(Reg *d, Reg *s, int order)
525 uint32_t r0, r1, r2, r3;
528 for (i = 0; i < 2 << SHIFT; i += 4) {
529 SHUFFLE4(L, s, s, i);
533 void glue(helper_pshuflw, SUFFIX)(Reg *d, Reg *s, int order)
535 uint16_t r0, r1, r2, r3;
538 for (i = 0, j = 1; j < 1 << SHIFT; i += 8, j += 2) {
539 SHUFFLE4(W, s, s, i);
544 void glue(helper_pshufhw, SUFFIX)(Reg *d, Reg *s, int order)
546 uint16_t r0, r1, r2, r3;
549 for (i = 4, j = 0; j < 1 << SHIFT; i += 8, j += 2) {
551 SHUFFLE4(W, s, s, i);
558 /* XXX: not accurate */
560 #define SSE_HELPER_P(name, F) \
561 void glue(helper_ ## name ## ps, SUFFIX)(CPUX86State *env, \
566 for (i = 0; i < 2 << SHIFT; i++) { \
567 d->ZMM_S(i) = F(32, v->ZMM_S(i), s->ZMM_S(i)); \
571 void glue(helper_ ## name ## pd, SUFFIX)(CPUX86State *env, \
576 for (i = 0; i < 1 << SHIFT; i++) { \
577 d->ZMM_D(i) = F(64, v->ZMM_D(i), s->ZMM_D(i)); \
583 #define SSE_HELPER_S(name, F) \
584 SSE_HELPER_P(name, F) \
586 void helper_ ## name ## ss(CPUX86State *env, Reg *d, Reg *s)\
589 d->ZMM_S(0) = F(32, v->ZMM_S(0), s->ZMM_S(0)); \
592 void helper_ ## name ## sd(CPUX86State *env, Reg *d, Reg *s)\
595 d->ZMM_D(0) = F(64, v->ZMM_D(0), s->ZMM_D(0)); \
600 #define SSE_HELPER_S(name, F) SSE_HELPER_P(name, F)
604 #define FPU_ADD(size, a, b) float ## size ## _add(a, b, &env->sse_status)
605 #define FPU_SUB(size, a, b) float ## size ## _sub(a, b, &env->sse_status)
606 #define FPU_MUL(size, a, b) float ## size ## _mul(a, b, &env->sse_status)
607 #define FPU_DIV(size, a, b) float ## size ## _div(a, b, &env->sse_status)
609 /* Note that the choice of comparison op here is important to get the
610 * special cases right: for min and max Intel specifies that (-0,0),
611 * (NaN, anything) and (anything, NaN) return the second argument.
613 #define FPU_MIN(size, a, b) \
614 (float ## size ## _lt(a, b, &env->sse_status) ? (a) : (b))
615 #define FPU_MAX(size, a, b) \
616 (float ## size ## _lt(b, a, &env->sse_status) ? (a) : (b))
618 SSE_HELPER_S(add, FPU_ADD)
619 SSE_HELPER_S(sub, FPU_SUB)
620 SSE_HELPER_S(mul, FPU_MUL)
621 SSE_HELPER_S(div, FPU_DIV)
622 SSE_HELPER_S(min, FPU_MIN)
623 SSE_HELPER_S(max, FPU_MAX)
625 void glue(helper_sqrtps, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
628 for (i = 0; i < 2 << SHIFT; i++) {
629 d->ZMM_S(i) = float32_sqrt(s->ZMM_S(i), &env->sse_status);
633 void glue(helper_sqrtpd, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
636 for (i = 0; i < 1 << SHIFT; i++) {
637 d->ZMM_D(i) = float64_sqrt(s->ZMM_D(i), &env->sse_status);
642 void helper_sqrtss(CPUX86State *env, Reg *d, Reg *s)
644 d->ZMM_S(0) = float32_sqrt(s->ZMM_S(0), &env->sse_status);
647 void helper_sqrtsd(CPUX86State *env, Reg *d, Reg *s)
649 d->ZMM_D(0) = float64_sqrt(s->ZMM_D(0), &env->sse_status);
653 /* float to float conversions */
654 void glue(helper_cvtps2pd, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
660 d->ZMM_D(0) = float32_to_float64(s0, &env->sse_status);
661 d->ZMM_D(1) = float32_to_float64(s1, &env->sse_status);
664 void glue(helper_cvtpd2ps, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
666 d->ZMM_S(0) = float64_to_float32(s->ZMM_D(0), &env->sse_status);
667 d->ZMM_S(1) = float64_to_float32(s->ZMM_D(1), &env->sse_status);
671 void helper_cvtss2sd(CPUX86State *env, Reg *d, Reg *s)
673 d->ZMM_D(0) = float32_to_float64(s->ZMM_S(0), &env->sse_status);
676 void helper_cvtsd2ss(CPUX86State *env, Reg *d, Reg *s)
678 d->ZMM_S(0) = float64_to_float32(s->ZMM_D(0), &env->sse_status);
681 /* integer to float */
682 void glue(helper_cvtdq2ps, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
684 d->ZMM_S(0) = int32_to_float32(s->ZMM_L(0), &env->sse_status);
685 d->ZMM_S(1) = int32_to_float32(s->ZMM_L(1), &env->sse_status);
686 d->ZMM_S(2) = int32_to_float32(s->ZMM_L(2), &env->sse_status);
687 d->ZMM_S(3) = int32_to_float32(s->ZMM_L(3), &env->sse_status);
690 void glue(helper_cvtdq2pd, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
694 l0 = (int32_t)s->ZMM_L(0);
695 l1 = (int32_t)s->ZMM_L(1);
696 d->ZMM_D(0) = int32_to_float64(l0, &env->sse_status);
697 d->ZMM_D(1) = int32_to_float64(l1, &env->sse_status);
700 void helper_cvtpi2ps(CPUX86State *env, ZMMReg *d, MMXReg *s)
702 d->ZMM_S(0) = int32_to_float32(s->MMX_L(0), &env->sse_status);
703 d->ZMM_S(1) = int32_to_float32(s->MMX_L(1), &env->sse_status);
706 void helper_cvtpi2pd(CPUX86State *env, ZMMReg *d, MMXReg *s)
708 d->ZMM_D(0) = int32_to_float64(s->MMX_L(0), &env->sse_status);
709 d->ZMM_D(1) = int32_to_float64(s->MMX_L(1), &env->sse_status);
712 void helper_cvtsi2ss(CPUX86State *env, ZMMReg *d, uint32_t val)
714 d->ZMM_S(0) = int32_to_float32(val, &env->sse_status);
717 void helper_cvtsi2sd(CPUX86State *env, ZMMReg *d, uint32_t val)
719 d->ZMM_D(0) = int32_to_float64(val, &env->sse_status);
723 void helper_cvtsq2ss(CPUX86State *env, ZMMReg *d, uint64_t val)
725 d->ZMM_S(0) = int64_to_float32(val, &env->sse_status);
728 void helper_cvtsq2sd(CPUX86State *env, ZMMReg *d, uint64_t val)
730 d->ZMM_D(0) = int64_to_float64(val, &env->sse_status);
734 /* float to integer */
737 * x86 mandates that we return the indefinite integer value for the result
738 * of any float-to-integer conversion that raises the 'invalid' exception.
739 * Wrap the softfloat functions to get this behaviour.
741 #define WRAP_FLOATCONV(RETTYPE, FN, FLOATTYPE, INDEFVALUE) \
742 static inline RETTYPE x86_##FN(FLOATTYPE a, float_status *s) \
744 int oldflags, newflags; \
747 oldflags = get_float_exception_flags(s); \
748 set_float_exception_flags(0, s); \
750 newflags = get_float_exception_flags(s); \
751 if (newflags & float_flag_invalid) { \
754 set_float_exception_flags(newflags | oldflags, s); \
758 WRAP_FLOATCONV(int32_t, float32_to_int32, float32, INT32_MIN)
759 WRAP_FLOATCONV(int32_t, float32_to_int32_round_to_zero, float32, INT32_MIN)
760 WRAP_FLOATCONV(int32_t, float64_to_int32, float64, INT32_MIN)
761 WRAP_FLOATCONV(int32_t, float64_to_int32_round_to_zero, float64, INT32_MIN)
762 WRAP_FLOATCONV(int64_t, float32_to_int64, float32, INT64_MIN)
763 WRAP_FLOATCONV(int64_t, float32_to_int64_round_to_zero, float32, INT64_MIN)
764 WRAP_FLOATCONV(int64_t, float64_to_int64, float64, INT64_MIN)
765 WRAP_FLOATCONV(int64_t, float64_to_int64_round_to_zero, float64, INT64_MIN)
767 void glue(helper_cvtps2dq, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
769 d->ZMM_L(0) = x86_float32_to_int32(s->ZMM_S(0), &env->sse_status);
770 d->ZMM_L(1) = x86_float32_to_int32(s->ZMM_S(1), &env->sse_status);
771 d->ZMM_L(2) = x86_float32_to_int32(s->ZMM_S(2), &env->sse_status);
772 d->ZMM_L(3) = x86_float32_to_int32(s->ZMM_S(3), &env->sse_status);
775 void glue(helper_cvtpd2dq, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
777 d->ZMM_L(0) = x86_float64_to_int32(s->ZMM_D(0), &env->sse_status);
778 d->ZMM_L(1) = x86_float64_to_int32(s->ZMM_D(1), &env->sse_status);
782 void helper_cvtps2pi(CPUX86State *env, MMXReg *d, ZMMReg *s)
784 d->MMX_L(0) = x86_float32_to_int32(s->ZMM_S(0), &env->sse_status);
785 d->MMX_L(1) = x86_float32_to_int32(s->ZMM_S(1), &env->sse_status);
788 void helper_cvtpd2pi(CPUX86State *env, MMXReg *d, ZMMReg *s)
790 d->MMX_L(0) = x86_float64_to_int32(s->ZMM_D(0), &env->sse_status);
791 d->MMX_L(1) = x86_float64_to_int32(s->ZMM_D(1), &env->sse_status);
794 int32_t helper_cvtss2si(CPUX86State *env, ZMMReg *s)
796 return x86_float32_to_int32(s->ZMM_S(0), &env->sse_status);
799 int32_t helper_cvtsd2si(CPUX86State *env, ZMMReg *s)
801 return x86_float64_to_int32(s->ZMM_D(0), &env->sse_status);
805 int64_t helper_cvtss2sq(CPUX86State *env, ZMMReg *s)
807 return x86_float32_to_int64(s->ZMM_S(0), &env->sse_status);
810 int64_t helper_cvtsd2sq(CPUX86State *env, ZMMReg *s)
812 return x86_float64_to_int64(s->ZMM_D(0), &env->sse_status);
816 /* float to integer truncated */
817 void glue(helper_cvttps2dq, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
819 d->ZMM_L(0) = x86_float32_to_int32_round_to_zero(s->ZMM_S(0), &env->sse_status);
820 d->ZMM_L(1) = x86_float32_to_int32_round_to_zero(s->ZMM_S(1), &env->sse_status);
821 d->ZMM_L(2) = x86_float32_to_int32_round_to_zero(s->ZMM_S(2), &env->sse_status);
822 d->ZMM_L(3) = x86_float32_to_int32_round_to_zero(s->ZMM_S(3), &env->sse_status);
825 void glue(helper_cvttpd2dq, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
827 d->ZMM_L(0) = x86_float64_to_int32_round_to_zero(s->ZMM_D(0), &env->sse_status);
828 d->ZMM_L(1) = x86_float64_to_int32_round_to_zero(s->ZMM_D(1), &env->sse_status);
832 void helper_cvttps2pi(CPUX86State *env, MMXReg *d, ZMMReg *s)
834 d->MMX_L(0) = x86_float32_to_int32_round_to_zero(s->ZMM_S(0), &env->sse_status);
835 d->MMX_L(1) = x86_float32_to_int32_round_to_zero(s->ZMM_S(1), &env->sse_status);
838 void helper_cvttpd2pi(CPUX86State *env, MMXReg *d, ZMMReg *s)
840 d->MMX_L(0) = x86_float64_to_int32_round_to_zero(s->ZMM_D(0), &env->sse_status);
841 d->MMX_L(1) = x86_float64_to_int32_round_to_zero(s->ZMM_D(1), &env->sse_status);
844 int32_t helper_cvttss2si(CPUX86State *env, ZMMReg *s)
846 return x86_float32_to_int32_round_to_zero(s->ZMM_S(0), &env->sse_status);
849 int32_t helper_cvttsd2si(CPUX86State *env, ZMMReg *s)
851 return x86_float64_to_int32_round_to_zero(s->ZMM_D(0), &env->sse_status);
855 int64_t helper_cvttss2sq(CPUX86State *env, ZMMReg *s)
857 return x86_float32_to_int64_round_to_zero(s->ZMM_S(0), &env->sse_status);
860 int64_t helper_cvttsd2sq(CPUX86State *env, ZMMReg *s)
862 return x86_float64_to_int64_round_to_zero(s->ZMM_D(0), &env->sse_status);
866 void glue(helper_rsqrtps, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
868 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
870 for (i = 0; i < 2 << SHIFT; i++) {
871 d->ZMM_S(i) = float32_div(float32_one,
872 float32_sqrt(s->ZMM_S(i), &env->sse_status),
875 set_float_exception_flags(old_flags, &env->sse_status);
878 void helper_rsqrtss(CPUX86State *env, ZMMReg *d, ZMMReg *s)
880 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
881 d->ZMM_S(0) = float32_div(float32_one,
882 float32_sqrt(s->ZMM_S(0), &env->sse_status),
884 set_float_exception_flags(old_flags, &env->sse_status);
887 void glue(helper_rcpps, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
889 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
891 for (i = 0; i < 2 << SHIFT; i++) {
892 d->ZMM_S(i) = float32_div(float32_one, s->ZMM_S(i), &env->sse_status);
894 set_float_exception_flags(old_flags, &env->sse_status);
897 void helper_rcpss(CPUX86State *env, ZMMReg *d, ZMMReg *s)
899 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
900 d->ZMM_S(0) = float32_div(float32_one, s->ZMM_S(0), &env->sse_status);
901 set_float_exception_flags(old_flags, &env->sse_status);
904 static inline uint64_t helper_extrq(uint64_t src, int shift, int len)
911 mask = (1ULL << len) - 1;
913 return (src >> shift) & mask;
916 void helper_extrq_r(CPUX86State *env, ZMMReg *d, ZMMReg *s)
918 d->ZMM_Q(0) = helper_extrq(d->ZMM_Q(0), s->ZMM_B(1), s->ZMM_B(0));
921 void helper_extrq_i(CPUX86State *env, ZMMReg *d, int index, int length)
923 d->ZMM_Q(0) = helper_extrq(d->ZMM_Q(0), index, length);
926 static inline uint64_t helper_insertq(uint64_t src, int shift, int len)
933 mask = (1ULL << len) - 1;
935 return (src & ~(mask << shift)) | ((src & mask) << shift);
938 void helper_insertq_r(CPUX86State *env, ZMMReg *d, ZMMReg *s)
940 d->ZMM_Q(0) = helper_insertq(s->ZMM_Q(0), s->ZMM_B(9), s->ZMM_B(8));
943 void helper_insertq_i(CPUX86State *env, ZMMReg *d, int index, int length)
945 d->ZMM_Q(0) = helper_insertq(d->ZMM_Q(0), index, length);
948 void glue(helper_haddps, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
952 r.ZMM_S(0) = float32_add(d->ZMM_S(0), d->ZMM_S(1), &env->sse_status);
953 r.ZMM_S(1) = float32_add(d->ZMM_S(2), d->ZMM_S(3), &env->sse_status);
954 r.ZMM_S(2) = float32_add(s->ZMM_S(0), s->ZMM_S(1), &env->sse_status);
955 r.ZMM_S(3) = float32_add(s->ZMM_S(2), s->ZMM_S(3), &env->sse_status);
959 void glue(helper_haddpd, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
963 r.ZMM_D(0) = float64_add(d->ZMM_D(0), d->ZMM_D(1), &env->sse_status);
964 r.ZMM_D(1) = float64_add(s->ZMM_D(0), s->ZMM_D(1), &env->sse_status);
968 void glue(helper_hsubps, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
972 r.ZMM_S(0) = float32_sub(d->ZMM_S(0), d->ZMM_S(1), &env->sse_status);
973 r.ZMM_S(1) = float32_sub(d->ZMM_S(2), d->ZMM_S(3), &env->sse_status);
974 r.ZMM_S(2) = float32_sub(s->ZMM_S(0), s->ZMM_S(1), &env->sse_status);
975 r.ZMM_S(3) = float32_sub(s->ZMM_S(2), s->ZMM_S(3), &env->sse_status);
979 void glue(helper_hsubpd, SUFFIX)(CPUX86State *env, ZMMReg *d, ZMMReg *s)
983 r.ZMM_D(0) = float64_sub(d->ZMM_D(0), d->ZMM_D(1), &env->sse_status);
984 r.ZMM_D(1) = float64_sub(s->ZMM_D(0), s->ZMM_D(1), &env->sse_status);
988 void glue(helper_addsubps, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
992 for (i = 0; i < 2 << SHIFT; i += 2) {
993 d->ZMM_S(i) = float32_sub(v->ZMM_S(i), s->ZMM_S(i), &env->sse_status);
994 d->ZMM_S(i+1) = float32_add(v->ZMM_S(i+1), s->ZMM_S(i+1), &env->sse_status);
998 void glue(helper_addsubpd, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1002 for (i = 0; i < 1 << SHIFT; i += 2) {
1003 d->ZMM_D(i) = float64_sub(v->ZMM_D(i), s->ZMM_D(i), &env->sse_status);
1004 d->ZMM_D(i+1) = float64_add(v->ZMM_D(i+1), s->ZMM_D(i+1), &env->sse_status);
1008 #define SSE_HELPER_CMP_P(name, F, C) \
1009 void glue(helper_ ## name ## ps, SUFFIX)(CPUX86State *env, \
1014 for (i = 0; i < 2 << SHIFT; i++) { \
1015 d->ZMM_L(i) = C(F(32, v->ZMM_S(i), s->ZMM_S(i))) ? -1 : 0; \
1019 void glue(helper_ ## name ## pd, SUFFIX)(CPUX86State *env, \
1024 for (i = 0; i < 1 << SHIFT; i++) { \
1025 d->ZMM_Q(i) = C(F(64, v->ZMM_D(i), s->ZMM_D(i))) ? -1 : 0; \
1030 #define SSE_HELPER_CMP(name, F, C) \
1031 SSE_HELPER_CMP_P(name, F, C) \
1032 void helper_ ## name ## ss(CPUX86State *env, Reg *d, Reg *s) \
1035 d->ZMM_L(0) = C(F(32, v->ZMM_S(0), s->ZMM_S(0))) ? -1 : 0; \
1038 void helper_ ## name ## sd(CPUX86State *env, Reg *d, Reg *s) \
1041 d->ZMM_Q(0) = C(F(64, v->ZMM_D(0), s->ZMM_D(0))) ? -1 : 0; \
1044 #define FPU_EQ(x) (x == float_relation_equal)
1045 #define FPU_LT(x) (x == float_relation_less)
1046 #define FPU_LE(x) (x <= float_relation_equal)
1047 #define FPU_UNORD(x) (x == float_relation_unordered)
1049 #define FPU_CMPQ(size, a, b) \
1050 float ## size ## _compare_quiet(a, b, &env->sse_status)
1051 #define FPU_CMPS(size, a, b) \
1052 float ## size ## _compare(a, b, &env->sse_status)
1055 #define SSE_HELPER_CMP(name, F, C) SSE_HELPER_CMP_P(name, F, C)
1058 SSE_HELPER_CMP(cmpeq, FPU_CMPQ, FPU_EQ)
1059 SSE_HELPER_CMP(cmplt, FPU_CMPS, FPU_LT)
1060 SSE_HELPER_CMP(cmple, FPU_CMPS, FPU_LE)
1061 SSE_HELPER_CMP(cmpunord, FPU_CMPQ, FPU_UNORD)
1062 SSE_HELPER_CMP(cmpneq, FPU_CMPQ, !FPU_EQ)
1063 SSE_HELPER_CMP(cmpnlt, FPU_CMPS, !FPU_LT)
1064 SSE_HELPER_CMP(cmpnle, FPU_CMPS, !FPU_LE)
1065 SSE_HELPER_CMP(cmpord, FPU_CMPQ, !FPU_UNORD)
1067 #undef SSE_HELPER_CMP
1069 static const int comis_eflags[4] = {CC_C, CC_Z, 0, CC_Z | CC_P | CC_C};
1071 void helper_ucomiss(CPUX86State *env, Reg *d, Reg *s)
1078 ret = float32_compare_quiet(s0, s1, &env->sse_status);
1079 CC_SRC = comis_eflags[ret + 1];
1082 void helper_comiss(CPUX86State *env, Reg *d, Reg *s)
1089 ret = float32_compare(s0, s1, &env->sse_status);
1090 CC_SRC = comis_eflags[ret + 1];
1093 void helper_ucomisd(CPUX86State *env, Reg *d, Reg *s)
1100 ret = float64_compare_quiet(d0, d1, &env->sse_status);
1101 CC_SRC = comis_eflags[ret + 1];
1104 void helper_comisd(CPUX86State *env, Reg *d, Reg *s)
1111 ret = float64_compare(d0, d1, &env->sse_status);
1112 CC_SRC = comis_eflags[ret + 1];
1115 uint32_t glue(helper_movmskps, SUFFIX)(CPUX86State *env, Reg *s)
1119 b0 = s->ZMM_L(0) >> 31;
1120 b1 = s->ZMM_L(1) >> 31;
1121 b2 = s->ZMM_L(2) >> 31;
1122 b3 = s->ZMM_L(3) >> 31;
1123 return b0 | (b1 << 1) | (b2 << 2) | (b3 << 3);
1126 uint32_t glue(helper_movmskpd, SUFFIX)(CPUX86State *env, Reg *s)
1130 b0 = s->ZMM_L(1) >> 31;
1131 b1 = s->ZMM_L(3) >> 31;
1132 return b0 | (b1 << 1);
1137 uint32_t glue(helper_pmovmskb, SUFFIX)(CPUX86State *env, Reg *s)
1143 for (i = 0; i < (1 << SHIFT); i++) {
1145 byte |= (s->B(8 * i + 0) >> 7);
1146 byte |= (s->B(8 * i + 1) >> 6) & 0x02;
1147 byte |= (s->B(8 * i + 2) >> 5) & 0x04;
1148 byte |= (s->B(8 * i + 3) >> 4) & 0x08;
1149 byte |= (s->B(8 * i + 4) >> 3) & 0x10;
1150 byte |= (s->B(8 * i + 5) >> 2) & 0x20;
1151 byte |= (s->B(8 * i + 6) >> 1) & 0x40;
1152 byte |= (s->B(8 * i + 7)) & 0x80;
1153 val |= byte << (8 * i);
1158 #define PACK_HELPER_B(name, F) \
1159 void glue(helper_pack ## name, SUFFIX)(CPUX86State *env, \
1163 uint8_t r[PACK_WIDTH * 2]; \
1165 for (j = 0; j < 4 << SHIFT; j += PACK_WIDTH) { \
1166 for (k = 0; k < PACK_WIDTH; k++) { \
1167 r[k] = F((int16_t)v->W(j + k)); \
1169 for (k = 0; k < PACK_WIDTH; k++) { \
1170 r[PACK_WIDTH + k] = F((int16_t)s->W(j + k)); \
1172 for (k = 0; k < PACK_WIDTH * 2; k++) { \
1173 d->B(2 * j + k) = r[k]; \
1178 PACK_HELPER_B(sswb, satsb)
1179 PACK_HELPER_B(uswb, satub)
1181 void glue(helper_packssdw, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1184 uint16_t r[PACK_WIDTH];
1187 for (j = 0; j < 2 << SHIFT; j += PACK_WIDTH / 2) {
1188 for (k = 0; k < PACK_WIDTH / 2; k++) {
1189 r[k] = satsw(v->L(j + k));
1191 for (k = 0; k < PACK_WIDTH / 2; k++) {
1192 r[PACK_WIDTH / 2 + k] = satsw(s->L(j + k));
1194 for (k = 0; k < PACK_WIDTH; k++) {
1195 d->W(2 * j + k) = r[k];
1200 #define UNPCK_OP(base_name, base) \
1202 void glue(helper_punpck ## base_name ## bw, SUFFIX)(CPUX86State *env,\
1206 uint8_t r[PACK_WIDTH * 2]; \
1209 for (j = 0; j < 8 << SHIFT; ) { \
1210 int k = j + base * PACK_WIDTH; \
1211 for (i = 0; i < PACK_WIDTH; i++) { \
1212 r[2 * i] = v->B(k + i); \
1213 r[2 * i + 1] = s->B(k + i); \
1215 for (i = 0; i < PACK_WIDTH * 2; i++, j++) { \
1221 void glue(helper_punpck ## base_name ## wd, SUFFIX)(CPUX86State *env,\
1225 uint16_t r[PACK_WIDTH]; \
1228 for (j = 0; j < 4 << SHIFT; ) { \
1229 int k = j + base * PACK_WIDTH / 2; \
1230 for (i = 0; i < PACK_WIDTH / 2; i++) { \
1231 r[2 * i] = v->W(k + i); \
1232 r[2 * i + 1] = s->W(k + i); \
1234 for (i = 0; i < PACK_WIDTH; i++, j++) { \
1240 void glue(helper_punpck ## base_name ## dq, SUFFIX)(CPUX86State *env,\
1244 uint32_t r[PACK_WIDTH / 2]; \
1247 for (j = 0; j < 2 << SHIFT; ) { \
1248 int k = j + base * PACK_WIDTH / 4; \
1249 for (i = 0; i < PACK_WIDTH / 4; i++) { \
1250 r[2 * i] = v->L(k + i); \
1251 r[2 * i + 1] = s->L(k + i); \
1253 for (i = 0; i < PACK_WIDTH / 2; i++, j++) { \
1260 void glue(helper_punpck ## base_name ## qdq, SUFFIX)( \
1261 CPUX86State *env, Reg *d, Reg *s) \
1267 for (i = 0; i < 1 << SHIFT; i += 2) { \
1268 r[0] = v->Q(base + i); \
1269 r[1] = s->Q(base + i); \
1271 d->Q(i + 1) = r[1]; \
1280 #undef PACK_HELPER_B
1284 /* 3DNow! float ops */
1286 void helper_pi2fd(CPUX86State *env, MMXReg *d, MMXReg *s)
1288 d->MMX_S(0) = int32_to_float32(s->MMX_L(0), &env->mmx_status);
1289 d->MMX_S(1) = int32_to_float32(s->MMX_L(1), &env->mmx_status);
1292 void helper_pi2fw(CPUX86State *env, MMXReg *d, MMXReg *s)
1294 d->MMX_S(0) = int32_to_float32((int16_t)s->MMX_W(0), &env->mmx_status);
1295 d->MMX_S(1) = int32_to_float32((int16_t)s->MMX_W(2), &env->mmx_status);
1298 void helper_pf2id(CPUX86State *env, MMXReg *d, MMXReg *s)
1300 d->MMX_L(0) = float32_to_int32_round_to_zero(s->MMX_S(0), &env->mmx_status);
1301 d->MMX_L(1) = float32_to_int32_round_to_zero(s->MMX_S(1), &env->mmx_status);
1304 void helper_pf2iw(CPUX86State *env, MMXReg *d, MMXReg *s)
1306 d->MMX_L(0) = satsw(float32_to_int32_round_to_zero(s->MMX_S(0),
1308 d->MMX_L(1) = satsw(float32_to_int32_round_to_zero(s->MMX_S(1),
1312 void helper_pfacc(CPUX86State *env, MMXReg *d, MMXReg *s)
1316 r = float32_add(d->MMX_S(0), d->MMX_S(1), &env->mmx_status);
1317 d->MMX_S(1) = float32_add(s->MMX_S(0), s->MMX_S(1), &env->mmx_status);
1321 void helper_pfadd(CPUX86State *env, MMXReg *d, MMXReg *s)
1323 d->MMX_S(0) = float32_add(d->MMX_S(0), s->MMX_S(0), &env->mmx_status);
1324 d->MMX_S(1) = float32_add(d->MMX_S(1), s->MMX_S(1), &env->mmx_status);
1327 void helper_pfcmpeq(CPUX86State *env, MMXReg *d, MMXReg *s)
1329 d->MMX_L(0) = float32_eq_quiet(d->MMX_S(0), s->MMX_S(0),
1330 &env->mmx_status) ? -1 : 0;
1331 d->MMX_L(1) = float32_eq_quiet(d->MMX_S(1), s->MMX_S(1),
1332 &env->mmx_status) ? -1 : 0;
1335 void helper_pfcmpge(CPUX86State *env, MMXReg *d, MMXReg *s)
1337 d->MMX_L(0) = float32_le(s->MMX_S(0), d->MMX_S(0),
1338 &env->mmx_status) ? -1 : 0;
1339 d->MMX_L(1) = float32_le(s->MMX_S(1), d->MMX_S(1),
1340 &env->mmx_status) ? -1 : 0;
1343 void helper_pfcmpgt(CPUX86State *env, MMXReg *d, MMXReg *s)
1345 d->MMX_L(0) = float32_lt(s->MMX_S(0), d->MMX_S(0),
1346 &env->mmx_status) ? -1 : 0;
1347 d->MMX_L(1) = float32_lt(s->MMX_S(1), d->MMX_S(1),
1348 &env->mmx_status) ? -1 : 0;
1351 void helper_pfmax(CPUX86State *env, MMXReg *d, MMXReg *s)
1353 if (float32_lt(d->MMX_S(0), s->MMX_S(0), &env->mmx_status)) {
1354 d->MMX_S(0) = s->MMX_S(0);
1356 if (float32_lt(d->MMX_S(1), s->MMX_S(1), &env->mmx_status)) {
1357 d->MMX_S(1) = s->MMX_S(1);
1361 void helper_pfmin(CPUX86State *env, MMXReg *d, MMXReg *s)
1363 if (float32_lt(s->MMX_S(0), d->MMX_S(0), &env->mmx_status)) {
1364 d->MMX_S(0) = s->MMX_S(0);
1366 if (float32_lt(s->MMX_S(1), d->MMX_S(1), &env->mmx_status)) {
1367 d->MMX_S(1) = s->MMX_S(1);
1371 void helper_pfmul(CPUX86State *env, MMXReg *d, MMXReg *s)
1373 d->MMX_S(0) = float32_mul(d->MMX_S(0), s->MMX_S(0), &env->mmx_status);
1374 d->MMX_S(1) = float32_mul(d->MMX_S(1), s->MMX_S(1), &env->mmx_status);
1377 void helper_pfnacc(CPUX86State *env, MMXReg *d, MMXReg *s)
1381 r = float32_sub(d->MMX_S(0), d->MMX_S(1), &env->mmx_status);
1382 d->MMX_S(1) = float32_sub(s->MMX_S(0), s->MMX_S(1), &env->mmx_status);
1386 void helper_pfpnacc(CPUX86State *env, MMXReg *d, MMXReg *s)
1390 r = float32_sub(d->MMX_S(0), d->MMX_S(1), &env->mmx_status);
1391 d->MMX_S(1) = float32_add(s->MMX_S(0), s->MMX_S(1), &env->mmx_status);
1395 void helper_pfrcp(CPUX86State *env, MMXReg *d, MMXReg *s)
1397 d->MMX_S(0) = float32_div(float32_one, s->MMX_S(0), &env->mmx_status);
1398 d->MMX_S(1) = d->MMX_S(0);
1401 void helper_pfrsqrt(CPUX86State *env, MMXReg *d, MMXReg *s)
1403 d->MMX_L(1) = s->MMX_L(0) & 0x7fffffff;
1404 d->MMX_S(1) = float32_div(float32_one,
1405 float32_sqrt(d->MMX_S(1), &env->mmx_status),
1407 d->MMX_L(1) |= s->MMX_L(0) & 0x80000000;
1408 d->MMX_L(0) = d->MMX_L(1);
1411 void helper_pfsub(CPUX86State *env, MMXReg *d, MMXReg *s)
1413 d->MMX_S(0) = float32_sub(d->MMX_S(0), s->MMX_S(0), &env->mmx_status);
1414 d->MMX_S(1) = float32_sub(d->MMX_S(1), s->MMX_S(1), &env->mmx_status);
1417 void helper_pfsubr(CPUX86State *env, MMXReg *d, MMXReg *s)
1419 d->MMX_S(0) = float32_sub(s->MMX_S(0), d->MMX_S(0), &env->mmx_status);
1420 d->MMX_S(1) = float32_sub(s->MMX_S(1), d->MMX_S(1), &env->mmx_status);
1423 void helper_pswapd(CPUX86State *env, MMXReg *d, MMXReg *s)
1428 d->MMX_L(0) = s->MMX_L(1);
1433 /* SSSE3 op helpers */
1434 void glue(helper_pshufb, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1441 for (i = 0; i < 8; i++) {
1442 r[i] = (s->B(i) & 0x80) ? 0 : (v->B(s->B(i) & 7));
1444 for (i = 0; i < 8; i++) {
1448 uint8_t r[8 << SHIFT];
1450 for (i = 0; i < 8 << SHIFT; i++) {
1452 r[i] = (s->B(i) & 0x80) ? 0 : v->B(j | (s->B(i) & 0xf));
1454 for (i = 0; i < 8 << SHIFT; i++) {
1460 #define SSE_HELPER_HW(name, F) \
1461 void glue(helper_ ## name, SUFFIX)(CPUX86State *env, Reg *d, Reg *s) \
1464 uint16_t r[4 << SHIFT]; \
1466 for (k = 0; k < 4 << SHIFT; k += LANE_WIDTH / 2) { \
1467 for (i = j = 0; j < LANE_WIDTH / 2; i++, j += 2) { \
1468 r[i + k] = F(v->W(j + k), v->W(j + k + 1)); \
1470 for (j = 0; j < LANE_WIDTH / 2; i++, j += 2) { \
1471 r[i + k] = F(s->W(j + k), s->W(j + k + 1)); \
1474 for (i = 0; i < 4 << SHIFT; i++) { \
1479 #define SSE_HELPER_HL(name, F) \
1480 void glue(helper_ ## name, SUFFIX)(CPUX86State *env, Reg *d, Reg *s) \
1483 uint32_t r[2 << SHIFT]; \
1485 for (k = 0; k < 2 << SHIFT; k += LANE_WIDTH / 4) { \
1486 for (i = j = 0; j < LANE_WIDTH / 4; i++, j += 2) { \
1487 r[i + k] = F(v->L(j + k), v->L(j + k + 1)); \
1489 for (j = 0; j < LANE_WIDTH / 4; i++, j += 2) { \
1490 r[i + k] = F(s->L(j + k), s->L(j + k + 1)); \
1493 for (i = 0; i < 2 << SHIFT; i++) { \
1498 SSE_HELPER_HW(phaddw, FADD)
1499 SSE_HELPER_HW(phsubw, FSUB)
1500 SSE_HELPER_HW(phaddsw, FADDSW)
1501 SSE_HELPER_HW(phsubsw, FSUBSW)
1502 SSE_HELPER_HL(phaddd, FADD)
1503 SSE_HELPER_HL(phsubd, FSUB)
1505 #undef SSE_HELPER_HW
1506 #undef SSE_HELPER_HL
1508 void glue(helper_pmaddubsw, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1512 for (i = 0; i < 4 << SHIFT; i++) {
1513 d->W(i) = satsw((int8_t)s->B(i * 2) * (uint8_t)v->B(i * 2) +
1514 (int8_t)s->B(i * 2 + 1) * (uint8_t)v->B(i * 2 + 1));
1518 #define FABSB(x) (x > INT8_MAX ? -(int8_t)x : x)
1519 #define FABSW(x) (x > INT16_MAX ? -(int16_t)x : x)
1520 #define FABSL(x) (x > INT32_MAX ? -(int32_t)x : x)
1521 SSE_HELPER_1(helper_pabsb, B, 8 << SHIFT, FABSB)
1522 SSE_HELPER_1(helper_pabsw, W, 4 << SHIFT, FABSW)
1523 SSE_HELPER_1(helper_pabsd, L, 2 << SHIFT, FABSL)
1525 #define FMULHRSW(d, s) (((int16_t) d * (int16_t)s + 0x4000) >> 15)
1526 SSE_HELPER_W(helper_pmulhrsw, FMULHRSW)
1528 #define FSIGNB(d, s) (s <= INT8_MAX ? s ? d : 0 : -(int8_t)d)
1529 #define FSIGNW(d, s) (s <= INT16_MAX ? s ? d : 0 : -(int16_t)d)
1530 #define FSIGNL(d, s) (s <= INT32_MAX ? s ? d : 0 : -(int32_t)d)
1531 SSE_HELPER_B(helper_psignb, FSIGNB)
1532 SSE_HELPER_W(helper_psignw, FSIGNW)
1533 SSE_HELPER_L(helper_psignd, FSIGNL)
1535 void glue(helper_palignr, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
1541 /* XXX could be checked during translation */
1542 if (shift >= (SHIFT ? 32 : 16)) {
1543 for (i = 0; i < (1 << SHIFT); i++) {
1548 #define SHR(v, i) (i < 64 && i > -64 ? i > 0 ? v >> (i) : (v << -(i)) : 0)
1550 d->Q(0) = SHR(s->Q(0), shift - 0) |
1551 SHR(v->Q(0), shift - 64);
1553 for (i = 0; i < (1 << SHIFT); i += 2) {
1556 r0 = SHR(s->Q(i), shift - 0) |
1557 SHR(s->Q(i + 1), shift - 64) |
1558 SHR(v->Q(i), shift - 128) |
1559 SHR(v->Q(i + 1), shift - 192);
1560 r1 = SHR(s->Q(i), shift + 64) |
1561 SHR(s->Q(i + 1), shift - 0) |
1562 SHR(v->Q(i), shift - 64) |
1563 SHR(v->Q(i + 1), shift - 128);
1572 #define XMM0 (env->xmm_regs[0])
1575 #define SSE_HELPER_V(name, elem, num, F) \
1576 void glue(name, SUFFIX)(CPUX86State *env, Reg *d, Reg *s) \
1578 d->elem(0) = F(d->elem(0), s->elem(0), XMM0.elem(0)); \
1579 d->elem(1) = F(d->elem(1), s->elem(1), XMM0.elem(1)); \
1581 d->elem(2) = F(d->elem(2), s->elem(2), XMM0.elem(2)); \
1582 d->elem(3) = F(d->elem(3), s->elem(3), XMM0.elem(3)); \
1584 d->elem(4) = F(d->elem(4), s->elem(4), XMM0.elem(4)); \
1585 d->elem(5) = F(d->elem(5), s->elem(5), XMM0.elem(5)); \
1586 d->elem(6) = F(d->elem(6), s->elem(6), XMM0.elem(6)); \
1587 d->elem(7) = F(d->elem(7), s->elem(7), XMM0.elem(7)); \
1589 d->elem(8) = F(d->elem(8), s->elem(8), XMM0.elem(8)); \
1590 d->elem(9) = F(d->elem(9), s->elem(9), XMM0.elem(9)); \
1591 d->elem(10) = F(d->elem(10), s->elem(10), XMM0.elem(10)); \
1592 d->elem(11) = F(d->elem(11), s->elem(11), XMM0.elem(11)); \
1593 d->elem(12) = F(d->elem(12), s->elem(12), XMM0.elem(12)); \
1594 d->elem(13) = F(d->elem(13), s->elem(13), XMM0.elem(13)); \
1595 d->elem(14) = F(d->elem(14), s->elem(14), XMM0.elem(14)); \
1596 d->elem(15) = F(d->elem(15), s->elem(15), XMM0.elem(15)); \
1602 #define SSE_HELPER_I(name, elem, num, F) \
1603 void glue(name, SUFFIX)(CPUX86State *env, Reg *d, Reg *s, uint32_t imm) \
1605 d->elem(0) = F(d->elem(0), s->elem(0), ((imm >> 0) & 1)); \
1606 d->elem(1) = F(d->elem(1), s->elem(1), ((imm >> 1) & 1)); \
1608 d->elem(2) = F(d->elem(2), s->elem(2), ((imm >> 2) & 1)); \
1609 d->elem(3) = F(d->elem(3), s->elem(3), ((imm >> 3) & 1)); \
1611 d->elem(4) = F(d->elem(4), s->elem(4), ((imm >> 4) & 1)); \
1612 d->elem(5) = F(d->elem(5), s->elem(5), ((imm >> 5) & 1)); \
1613 d->elem(6) = F(d->elem(6), s->elem(6), ((imm >> 6) & 1)); \
1614 d->elem(7) = F(d->elem(7), s->elem(7), ((imm >> 7) & 1)); \
1616 d->elem(8) = F(d->elem(8), s->elem(8), ((imm >> 8) & 1)); \
1617 d->elem(9) = F(d->elem(9), s->elem(9), ((imm >> 9) & 1)); \
1618 d->elem(10) = F(d->elem(10), s->elem(10), \
1619 ((imm >> 10) & 1)); \
1620 d->elem(11) = F(d->elem(11), s->elem(11), \
1621 ((imm >> 11) & 1)); \
1622 d->elem(12) = F(d->elem(12), s->elem(12), \
1623 ((imm >> 12) & 1)); \
1624 d->elem(13) = F(d->elem(13), s->elem(13), \
1625 ((imm >> 13) & 1)); \
1626 d->elem(14) = F(d->elem(14), s->elem(14), \
1627 ((imm >> 14) & 1)); \
1628 d->elem(15) = F(d->elem(15), s->elem(15), \
1629 ((imm >> 15) & 1)); \
1635 /* SSE4.1 op helpers */
1636 #define FBLENDVB(d, s, m) ((m & 0x80) ? s : d)
1637 #define FBLENDVPS(d, s, m) ((m & 0x80000000) ? s : d)
1638 #define FBLENDVPD(d, s, m) ((m & 0x8000000000000000LL) ? s : d)
1639 SSE_HELPER_V(helper_pblendvb, B, 16, FBLENDVB)
1640 SSE_HELPER_V(helper_blendvps, L, 4, FBLENDVPS)
1641 SSE_HELPER_V(helper_blendvpd, Q, 2, FBLENDVPD)
1643 void glue(helper_ptest, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1645 uint64_t zf = 0, cf = 0;
1648 for (i = 0; i < 1 << SHIFT; i++) {
1649 zf |= (s->Q(i) & d->Q(i));
1650 cf |= (s->Q(i) & ~d->Q(i));
1652 CC_SRC = (zf ? 0 : CC_Z) | (cf ? 0 : CC_C);
1655 #define SSE_HELPER_F(name, elem, num, F) \
1656 void glue(name, SUFFIX)(CPUX86State *env, Reg *d, Reg *s) \
1659 for (int i = n; --i >= 0; ) { \
1660 d->elem(i) = F(i); \
1665 SSE_HELPER_F(helper_pmovsxbw, W, 4 << SHIFT, (int8_t) s->B)
1666 SSE_HELPER_F(helper_pmovsxbd, L, 2 << SHIFT, (int8_t) s->B)
1667 SSE_HELPER_F(helper_pmovsxbq, Q, 1 << SHIFT, (int8_t) s->B)
1668 SSE_HELPER_F(helper_pmovsxwd, L, 2 << SHIFT, (int16_t) s->W)
1669 SSE_HELPER_F(helper_pmovsxwq, Q, 1 << SHIFT, (int16_t) s->W)
1670 SSE_HELPER_F(helper_pmovsxdq, Q, 1 << SHIFT, (int32_t) s->L)
1671 SSE_HELPER_F(helper_pmovzxbw, W, 4 << SHIFT, s->B)
1672 SSE_HELPER_F(helper_pmovzxbd, L, 2 << SHIFT, s->B)
1673 SSE_HELPER_F(helper_pmovzxbq, Q, 1 << SHIFT, s->B)
1674 SSE_HELPER_F(helper_pmovzxwd, L, 2 << SHIFT, s->W)
1675 SSE_HELPER_F(helper_pmovzxwq, Q, 1 << SHIFT, s->W)
1676 SSE_HELPER_F(helper_pmovzxdq, Q, 1 << SHIFT, s->L)
1679 void glue(helper_pmuldq, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1684 for (i = 0; i < 1 << SHIFT; i++) {
1685 d->Q(i) = (int64_t)(int32_t) v->L(2 * i) * (int32_t) s->L(2 * i);
1689 #define FCMPEQQ(d, s) (d == s ? -1 : 0)
1690 SSE_HELPER_Q(helper_pcmpeqq, FCMPEQQ)
1692 void glue(helper_packusdw, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1698 for (i = 0, j = 0; i <= 2 << SHIFT; i += 8, j += 4) {
1699 r[0] = satuw(v->L(j));
1700 r[1] = satuw(v->L(j + 1));
1701 r[2] = satuw(v->L(j + 2));
1702 r[3] = satuw(v->L(j + 3));
1703 r[4] = satuw(s->L(j));
1704 r[5] = satuw(s->L(j + 1));
1705 r[6] = satuw(s->L(j + 2));
1706 r[7] = satuw(s->L(j + 3));
1707 for (k = 0; k < 8; k++) {
1713 #define FMINSB(d, s) MIN((int8_t)d, (int8_t)s)
1714 #define FMINSD(d, s) MIN((int32_t)d, (int32_t)s)
1715 #define FMAXSB(d, s) MAX((int8_t)d, (int8_t)s)
1716 #define FMAXSD(d, s) MAX((int32_t)d, (int32_t)s)
1717 SSE_HELPER_B(helper_pminsb, FMINSB)
1718 SSE_HELPER_L(helper_pminsd, FMINSD)
1719 SSE_HELPER_W(helper_pminuw, MIN)
1720 SSE_HELPER_L(helper_pminud, MIN)
1721 SSE_HELPER_B(helper_pmaxsb, FMAXSB)
1722 SSE_HELPER_L(helper_pmaxsd, FMAXSD)
1723 SSE_HELPER_W(helper_pmaxuw, MAX)
1724 SSE_HELPER_L(helper_pmaxud, MAX)
1726 #define FMULLD(d, s) ((int32_t)d * (int32_t)s)
1727 SSE_HELPER_L(helper_pmulld, FMULLD)
1729 void glue(helper_phminposuw, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
1733 if (s->W(1) < s->W(idx)) {
1736 if (s->W(2) < s->W(idx)) {
1739 if (s->W(3) < s->W(idx)) {
1742 if (s->W(4) < s->W(idx)) {
1745 if (s->W(5) < s->W(idx)) {
1748 if (s->W(6) < s->W(idx)) {
1751 if (s->W(7) < s->W(idx)) {
1755 d->W(0) = s->W(idx);
1761 void glue(helper_roundps, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
1764 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
1765 signed char prev_rounding_mode;
1767 prev_rounding_mode = env->sse_status.float_rounding_mode;
1768 if (!(mode & (1 << 2))) {
1771 set_float_rounding_mode(float_round_nearest_even, &env->sse_status);
1774 set_float_rounding_mode(float_round_down, &env->sse_status);
1777 set_float_rounding_mode(float_round_up, &env->sse_status);
1780 set_float_rounding_mode(float_round_to_zero, &env->sse_status);
1785 d->ZMM_S(0) = float32_round_to_int(s->ZMM_S(0), &env->sse_status);
1786 d->ZMM_S(1) = float32_round_to_int(s->ZMM_S(1), &env->sse_status);
1787 d->ZMM_S(2) = float32_round_to_int(s->ZMM_S(2), &env->sse_status);
1788 d->ZMM_S(3) = float32_round_to_int(s->ZMM_S(3), &env->sse_status);
1790 if (mode & (1 << 3) && !(old_flags & float_flag_inexact)) {
1791 set_float_exception_flags(get_float_exception_flags(&env->sse_status) &
1792 ~float_flag_inexact,
1795 env->sse_status.float_rounding_mode = prev_rounding_mode;
1798 void glue(helper_roundpd, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
1801 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
1802 signed char prev_rounding_mode;
1804 prev_rounding_mode = env->sse_status.float_rounding_mode;
1805 if (!(mode & (1 << 2))) {
1808 set_float_rounding_mode(float_round_nearest_even, &env->sse_status);
1811 set_float_rounding_mode(float_round_down, &env->sse_status);
1814 set_float_rounding_mode(float_round_up, &env->sse_status);
1817 set_float_rounding_mode(float_round_to_zero, &env->sse_status);
1822 d->ZMM_D(0) = float64_round_to_int(s->ZMM_D(0), &env->sse_status);
1823 d->ZMM_D(1) = float64_round_to_int(s->ZMM_D(1), &env->sse_status);
1825 if (mode & (1 << 3) && !(old_flags & float_flag_inexact)) {
1826 set_float_exception_flags(get_float_exception_flags(&env->sse_status) &
1827 ~float_flag_inexact,
1830 env->sse_status.float_rounding_mode = prev_rounding_mode;
1833 void glue(helper_roundss, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
1836 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
1837 signed char prev_rounding_mode;
1839 prev_rounding_mode = env->sse_status.float_rounding_mode;
1840 if (!(mode & (1 << 2))) {
1843 set_float_rounding_mode(float_round_nearest_even, &env->sse_status);
1846 set_float_rounding_mode(float_round_down, &env->sse_status);
1849 set_float_rounding_mode(float_round_up, &env->sse_status);
1852 set_float_rounding_mode(float_round_to_zero, &env->sse_status);
1857 d->ZMM_S(0) = float32_round_to_int(s->ZMM_S(0), &env->sse_status);
1859 if (mode & (1 << 3) && !(old_flags & float_flag_inexact)) {
1860 set_float_exception_flags(get_float_exception_flags(&env->sse_status) &
1861 ~float_flag_inexact,
1864 env->sse_status.float_rounding_mode = prev_rounding_mode;
1867 void glue(helper_roundsd, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
1870 uint8_t old_flags = get_float_exception_flags(&env->sse_status);
1871 signed char prev_rounding_mode;
1873 prev_rounding_mode = env->sse_status.float_rounding_mode;
1874 if (!(mode & (1 << 2))) {
1877 set_float_rounding_mode(float_round_nearest_even, &env->sse_status);
1880 set_float_rounding_mode(float_round_down, &env->sse_status);
1883 set_float_rounding_mode(float_round_up, &env->sse_status);
1886 set_float_rounding_mode(float_round_to_zero, &env->sse_status);
1891 d->ZMM_D(0) = float64_round_to_int(s->ZMM_D(0), &env->sse_status);
1893 if (mode & (1 << 3) && !(old_flags & float_flag_inexact)) {
1894 set_float_exception_flags(get_float_exception_flags(&env->sse_status) &
1895 ~float_flag_inexact,
1898 env->sse_status.float_rounding_mode = prev_rounding_mode;
1901 #define FBLENDP(d, s, m) (m ? s : d)
1902 SSE_HELPER_I(helper_blendps, L, 4, FBLENDP)
1903 SSE_HELPER_I(helper_blendpd, Q, 2, FBLENDP)
1904 SSE_HELPER_I(helper_pblendw, W, 8, FBLENDP)
1906 void glue(helper_dpps, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
1910 float32 prod1, prod2, temp2, temp3, temp4;
1913 for (i = 0; i < 2 << SHIFT; i += 4) {
1915 * We must evaluate (A+B)+(C+D), not ((A+B)+C)+D
1916 * to correctly round the intermediate results
1918 if (mask & (1 << 4)) {
1919 prod1 = float32_mul(v->ZMM_S(i), s->ZMM_S(i), &env->sse_status);
1921 prod1 = float32_zero;
1923 if (mask & (1 << 5)) {
1924 prod2 = float32_mul(v->ZMM_S(i+1), s->ZMM_S(i+1), &env->sse_status);
1926 prod2 = float32_zero;
1928 temp2 = float32_add(prod1, prod2, &env->sse_status);
1929 if (mask & (1 << 6)) {
1930 prod1 = float32_mul(v->ZMM_S(i+2), s->ZMM_S(i+2), &env->sse_status);
1932 prod1 = float32_zero;
1934 if (mask & (1 << 7)) {
1935 prod2 = float32_mul(v->ZMM_S(i+3), s->ZMM_S(i+3), &env->sse_status);
1937 prod2 = float32_zero;
1939 temp3 = float32_add(prod1, prod2, &env->sse_status);
1940 temp4 = float32_add(temp2, temp3, &env->sse_status);
1942 d->ZMM_S(i) = (mask & (1 << 0)) ? temp4 : float32_zero;
1943 d->ZMM_S(i+1) = (mask & (1 << 1)) ? temp4 : float32_zero;
1944 d->ZMM_S(i+2) = (mask & (1 << 2)) ? temp4 : float32_zero;
1945 d->ZMM_S(i+3) = (mask & (1 << 3)) ? temp4 : float32_zero;
1950 /* Oddly, there is no ymm version of dppd */
1951 void glue(helper_dppd, SUFFIX)(CPUX86State *env,
1952 Reg *d, Reg *s, uint32_t mask)
1955 float64 prod1, prod2, temp2;
1957 if (mask & (1 << 4)) {
1958 prod1 = float64_mul(v->ZMM_D(0), s->ZMM_D(0), &env->sse_status);
1960 prod1 = float64_zero;
1962 if (mask & (1 << 5)) {
1963 prod2 = float64_mul(v->ZMM_D(1), s->ZMM_D(1), &env->sse_status);
1965 prod2 = float64_zero;
1967 temp2 = float64_add(prod1, prod2, &env->sse_status);
1968 d->ZMM_D(0) = (mask & (1 << 0)) ? temp2 : float64_zero;
1969 d->ZMM_D(1) = (mask & (1 << 1)) ? temp2 : float64_zero;
1973 void glue(helper_mpsadbw, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
1980 for (j = 0; j < 4 << SHIFT; ) {
1981 int s0 = (j * 2) + ((offset & 3) << 2);
1982 int d0 = (j * 2) + ((offset & 4) << 0);
1983 for (i = 0; i < LANE_WIDTH / 2; i++, d0++) {
1985 r[i] += abs1(v->B(d0 + 0) - s->B(s0 + 0));
1986 r[i] += abs1(v->B(d0 + 1) - s->B(s0 + 1));
1987 r[i] += abs1(v->B(d0 + 2) - s->B(s0 + 2));
1988 r[i] += abs1(v->B(d0 + 3) - s->B(s0 + 3));
1990 for (i = 0; i < LANE_WIDTH / 2; i++, j++) {
1997 /* SSE4.2 op helpers */
1998 #define FCMPGTQ(d, s) ((int64_t)d > (int64_t)s ? -1 : 0)
1999 SSE_HELPER_Q(helper_pcmpgtq, FCMPGTQ)
2001 static inline int pcmp_elen(CPUX86State *env, int reg, uint32_t ctrl)
2003 target_long val, limit;
2005 /* Presence of REX.W is indicated by a bit higher than 7 set */
2007 val = (target_long)env->regs[reg];
2009 val = (int32_t)env->regs[reg];
2016 if ((val > limit) || (val < -limit)) {
2022 static inline int pcmp_ilen(Reg *r, uint8_t ctrl)
2027 while (val < 8 && r->W(val)) {
2031 while (val < 16 && r->B(val)) {
2039 static inline int pcmp_val(Reg *r, uint8_t ctrl, int i)
2041 switch ((ctrl >> 0) & 3) {
2047 return (int8_t)r->B(i);
2050 return (int16_t)r->W(i);
2054 static inline unsigned pcmpxstrx(CPUX86State *env, Reg *d, Reg *s,
2055 int8_t ctrl, int valids, int validd)
2057 unsigned int res = 0;
2060 int upper = (ctrl & 1) ? 7 : 15;
2065 CC_SRC = (valids < upper ? CC_Z : 0) | (validd < upper ? CC_S : 0);
2067 switch ((ctrl >> 2) & 3) {
2069 for (j = valids; j >= 0; j--) {
2071 v = pcmp_val(s, ctrl, j);
2072 for (i = validd; i >= 0; i--) {
2073 res |= (v == pcmp_val(d, ctrl, i));
2078 for (j = valids; j >= 0; j--) {
2080 v = pcmp_val(s, ctrl, j);
2081 for (i = ((validd - 1) | 1); i >= 0; i -= 2) {
2082 res |= (pcmp_val(d, ctrl, i - 0) >= v &&
2083 pcmp_val(d, ctrl, i - 1) <= v);
2088 res = (1 << (upper - MAX(valids, validd))) - 1;
2089 res <<= MAX(valids, validd) - MIN(valids, validd);
2090 for (i = MIN(valids, validd); i >= 0; i--) {
2092 v = pcmp_val(s, ctrl, i);
2093 res |= (v == pcmp_val(d, ctrl, i));
2098 res = (2 << upper) - 1;
2101 for (j = valids == upper ? valids : valids - validd; j >= 0; j--) {
2104 for (i = MIN(valids - j, validd); i >= 0; i--) {
2105 v &= (pcmp_val(s, ctrl, i + j) == pcmp_val(d, ctrl, i));
2112 switch ((ctrl >> 4) & 3) {
2114 res ^= (2 << upper) - 1;
2117 res ^= (1 << (valids + 1)) - 1;
2131 void glue(helper_pcmpestri, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
2134 unsigned int res = pcmpxstrx(env, d, s, ctrl,
2135 pcmp_elen(env, R_EDX, ctrl),
2136 pcmp_elen(env, R_EAX, ctrl));
2139 env->regs[R_ECX] = (ctrl & (1 << 6)) ? 31 - clz32(res) : ctz32(res);
2141 env->regs[R_ECX] = 16 >> (ctrl & (1 << 0));
2145 void glue(helper_pcmpestrm, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
2149 unsigned int res = pcmpxstrx(env, d, s, ctrl,
2150 pcmp_elen(env, R_EDX, ctrl),
2151 pcmp_elen(env, R_EAX, ctrl));
2153 if ((ctrl >> 6) & 1) {
2155 for (i = 0; i < 8; i++, res >>= 1) {
2156 env->xmm_regs[0].W(i) = (res & 1) ? ~0 : 0;
2159 for (i = 0; i < 16; i++, res >>= 1) {
2160 env->xmm_regs[0].B(i) = (res & 1) ? ~0 : 0;
2164 env->xmm_regs[0].Q(1) = 0;
2165 env->xmm_regs[0].Q(0) = res;
2169 void glue(helper_pcmpistri, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
2172 unsigned int res = pcmpxstrx(env, d, s, ctrl,
2174 pcmp_ilen(d, ctrl));
2177 env->regs[R_ECX] = (ctrl & (1 << 6)) ? 31 - clz32(res) : ctz32(res);
2179 env->regs[R_ECX] = 16 >> (ctrl & (1 << 0));
2183 void glue(helper_pcmpistrm, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
2187 unsigned int res = pcmpxstrx(env, d, s, ctrl,
2189 pcmp_ilen(d, ctrl));
2191 if ((ctrl >> 6) & 1) {
2193 for (i = 0; i < 8; i++, res >>= 1) {
2194 env->xmm_regs[0].W(i) = (res & 1) ? ~0 : 0;
2197 for (i = 0; i < 16; i++, res >>= 1) {
2198 env->xmm_regs[0].B(i) = (res & 1) ? ~0 : 0;
2202 env->xmm_regs[0].Q(1) = 0;
2203 env->xmm_regs[0].Q(0) = res;
2207 #define CRCPOLY 0x1edc6f41
2208 #define CRCPOLY_BITREV 0x82f63b78
2209 target_ulong helper_crc32(uint32_t crc1, target_ulong msg, uint32_t len)
2211 target_ulong crc = (msg & ((target_ulong) -1 >>
2212 (TARGET_LONG_BITS - len))) ^ crc1;
2215 crc = (crc >> 1) ^ ((crc & 1) ? CRCPOLY_BITREV : 0);
2221 void glue(helper_pclmulqdq, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
2224 uint64_t ah, al, b, resh, resl;
2227 al = d->Q((ctrl & 1) != 0);
2228 b = s->Q((ctrl & 16) != 0);
2236 ah = (ah << 1) | (al >> 63);
2245 void glue(helper_aesdec, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
2251 for (i = 0 ; i < 4 ; i++) {
2252 d->L(i) = rk.L(i) ^ bswap32(AES_Td0[st.B(AES_ishifts[4*i+0])] ^
2253 AES_Td1[st.B(AES_ishifts[4*i+1])] ^
2254 AES_Td2[st.B(AES_ishifts[4*i+2])] ^
2255 AES_Td3[st.B(AES_ishifts[4*i+3])]);
2259 void glue(helper_aesdeclast, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
2265 for (i = 0; i < 16; i++) {
2266 d->B(i) = rk.B(i) ^ (AES_isbox[st.B(AES_ishifts[i])]);
2270 void glue(helper_aesenc, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
2276 for (i = 0 ; i < 4 ; i++) {
2277 d->L(i) = rk.L(i) ^ bswap32(AES_Te0[st.B(AES_shifts[4*i+0])] ^
2278 AES_Te1[st.B(AES_shifts[4*i+1])] ^
2279 AES_Te2[st.B(AES_shifts[4*i+2])] ^
2280 AES_Te3[st.B(AES_shifts[4*i+3])]);
2284 void glue(helper_aesenclast, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
2290 for (i = 0; i < 16; i++) {
2291 d->B(i) = rk.B(i) ^ (AES_sbox[st.B(AES_shifts[i])]);
2296 void glue(helper_aesimc, SUFFIX)(CPUX86State *env, Reg *d, Reg *s)
2301 for (i = 0 ; i < 4 ; i++) {
2302 d->L(i) = bswap32(AES_imc[tmp.B(4*i+0)][0] ^
2303 AES_imc[tmp.B(4*i+1)][1] ^
2304 AES_imc[tmp.B(4*i+2)][2] ^
2305 AES_imc[tmp.B(4*i+3)][3]);
2309 void glue(helper_aeskeygenassist, SUFFIX)(CPUX86State *env, Reg *d, Reg *s,
2315 for (i = 0 ; i < 4 ; i++) {
2316 d->B(i) = AES_sbox[tmp.B(i + 4)];
2317 d->B(i + 8) = AES_sbox[tmp.B(i + 12)];
2319 d->L(1) = (d->L(0) << 24 | d->L(0) >> 8) ^ ctrl;
2320 d->L(3) = (d->L(2) << 24 | d->L(2) >> 8) ^ ctrl;