diff mbox series

[v11,17/20] target/arm: Use vector infrastructure for aa64 multiplies

Message ID 20180126045742.5487-18-richard.henderson@linaro.org
State Superseded
Headers show
Series tcg: generic vector operations | expand

Commit Message

Richard Henderson Jan. 26, 2018, 4:57 a.m. UTC
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>

Signed-off-by: Richard Henderson <richard.henderson@linaro.org>

---
 target/arm/translate-a64.c | 154 +++++++++++++++++++++++++++++++++++++--------
 1 file changed, 129 insertions(+), 25 deletions(-)

-- 
2.14.3

Comments

Alex Bennée Feb. 6, 2018, 11:11 a.m. UTC | #1
Richard Henderson <richard.henderson@linaro.org> writes:

> Reviewed-by: Peter Maydell <peter.maydell@linaro.org>

> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>


Reviewed-by: Alex Bennée <alex.bennee@linaro.org>


> ---

>  target/arm/translate-a64.c | 154 +++++++++++++++++++++++++++++++++++++--------

>  1 file changed, 129 insertions(+), 25 deletions(-)

>

> diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c

> index c928c4787c..64a2c2df59 100644

> --- a/target/arm/translate-a64.c

> +++ b/target/arm/translate-a64.c

> @@ -9753,6 +9753,66 @@ static void disas_simd_3same_float(DisasContext *s, uint32_t insn)

>      }

>  }

>

> +static void gen_mla8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)

> +{

> +    gen_helper_neon_mul_u8(a, a, b);

> +    gen_helper_neon_add_u8(d, d, a);

> +}

> +

> +static void gen_mla16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)

> +{

> +    gen_helper_neon_mul_u16(a, a, b);

> +    gen_helper_neon_add_u16(d, d, a);

> +}

> +

> +static void gen_mla32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)

> +{

> +    tcg_gen_mul_i32(a, a, b);

> +    tcg_gen_add_i32(d, d, a);

> +}

> +

> +static void gen_mla64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)

> +{

> +    tcg_gen_mul_i64(a, a, b);

> +    tcg_gen_add_i64(d, d, a);

> +}

> +

> +static void gen_mla_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)

> +{

> +    tcg_gen_mul_vec(vece, a, a, b);

> +    tcg_gen_add_vec(vece, d, d, a);

> +}

> +

> +static void gen_mls8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)

> +{

> +    gen_helper_neon_mul_u8(a, a, b);

> +    gen_helper_neon_sub_u8(d, d, a);

> +}

> +

> +static void gen_mls16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)

> +{

> +    gen_helper_neon_mul_u16(a, a, b);

> +    gen_helper_neon_sub_u16(d, d, a);

> +}

> +

> +static void gen_mls32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)

> +{

> +    tcg_gen_mul_i32(a, a, b);

> +    tcg_gen_sub_i32(d, d, a);

> +}

> +

> +static void gen_mls64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)

> +{

> +    tcg_gen_mul_i64(a, a, b);

> +    tcg_gen_sub_i64(d, d, a);

> +}

> +

> +static void gen_mls_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)

> +{

> +    tcg_gen_mul_vec(vece, a, a, b);

> +    tcg_gen_sub_vec(vece, d, d, a);

> +}

> +

>  /* Integer op subgroup of C3.6.16. */

>  static void disas_simd_3same_int(DisasContext *s, uint32_t insn)

>  {

> @@ -9771,6 +9831,52 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn)

>            .prefer_i64 = TCG_TARGET_REG_BITS == 64,

>            .vece = MO_64 },

>      };

> +    static const GVecGen3 mla_op[4] = {

> +        { .fni4 = gen_mla8_i32,

> +          .fniv = gen_mla_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .load_dest = true,

> +          .vece = MO_8 },

> +        { .fni4 = gen_mla16_i32,

> +          .fniv = gen_mla_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .load_dest = true,

> +          .vece = MO_16 },

> +        { .fni4 = gen_mla32_i32,

> +          .fniv = gen_mla_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .load_dest = true,

> +          .vece = MO_32 },

> +        { .fni8 = gen_mla64_i64,

> +          .fniv = gen_mla_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .prefer_i64 = TCG_TARGET_REG_BITS == 64,

> +          .load_dest = true,

> +          .vece = MO_64 },

> +    };

> +    static const GVecGen3 mls_op[4] = {

> +        { .fni4 = gen_mls8_i32,

> +          .fniv = gen_mls_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .load_dest = true,

> +          .vece = MO_8 },

> +        { .fni4 = gen_mls16_i32,

> +          .fniv = gen_mls_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .load_dest = true,

> +          .vece = MO_16 },

> +        { .fni4 = gen_mls32_i32,

> +          .fniv = gen_mls_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .load_dest = true,

> +          .vece = MO_32 },

> +        { .fni8 = gen_mls64_i64,

> +          .fniv = gen_mls_vec,

> +          .opc = INDEX_op_mul_vec,

> +          .prefer_i64 = TCG_TARGET_REG_BITS == 64,

> +          .load_dest = true,

> +          .vece = MO_64 },

> +    };

>

>      int is_q = extract32(insn, 30, 1);

>      int u = extract32(insn, 29, 1);

> @@ -9828,6 +9934,19 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn)

>              gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_add, size);

>          }

>          return;

> +    case 0x13: /* MUL, PMUL */

> +        if (!u) { /* MUL */

> +            gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_mul, size);

> +            return;

> +        }

> +        break;

> +    case 0x12: /* MLA, MLS */

> +        if (u) {

> +            gen_gvec_op3(s, is_q, rd, rn, rm, &mls_op[size]);

> +        } else {

> +            gen_gvec_op3(s, is_q, rd, rn, rm, &mla_op[size]);

> +        }

> +        return;

>      case 0x11:

>          if (!u) { /* CMTST */

>              gen_gvec_op3(s, is_q, rd, rn, rm, &cmtst_op[size]);

> @@ -10002,23 +10121,10 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn)

>                  break;

>              }

>              case 0x13: /* MUL, PMUL */

> -                if (u) {

> -                    /* PMUL */

> -                    assert(size == 0);

> -                    genfn = gen_helper_neon_mul_p8;

> -                    break;

> -                }

> -                /* fall through : MUL */

> -            case 0x12: /* MLA, MLS */

> -            {

> -                static NeonGenTwoOpFn * const fns[3] = {

> -                    gen_helper_neon_mul_u8,

> -                    gen_helper_neon_mul_u16,

> -                    tcg_gen_mul_i32,

> -                };

> -                genfn = fns[size];

> +                assert(u); /* PMUL */

> +                assert(size == 0);

> +                genfn = gen_helper_neon_mul_p8;

>                  break;

> -            }

>              case 0x16: /* SQDMULH, SQRDMULH */

>              {

>                  static NeonGenTwoOpEnvFn * const fns[2][2] = {

> @@ -10039,18 +10145,16 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn)

>                  genfn(tcg_res, tcg_op1, tcg_op2);

>              }

>

> -            if (opcode == 0xf || opcode == 0x12) {

> -                /* SABA, UABA, MLA, MLS: accumulating ops */

> -                static NeonGenTwoOpFn * const fns[3][2] = {

> -                    { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 },

> -                    { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 },

> -                    { tcg_gen_add_i32, tcg_gen_sub_i32 },

> +            if (opcode == 0xf) {

> +                /* SABA, UABA: accumulating ops */

> +                static NeonGenTwoOpFn * const fns[3] = {

> +                    gen_helper_neon_add_u8,

> +                    gen_helper_neon_add_u16,

> +                    tcg_gen_add_i32,

>                  };

> -                bool is_sub = (opcode == 0x12 && u); /* MLS */

>

> -                genfn = fns[size][is_sub];

>                  read_vec_element_i32(s, tcg_op1, rd, pass, MO_32);

> -                genfn(tcg_res, tcg_op1, tcg_res);

> +                fns[size](tcg_res, tcg_op1, tcg_res);

>              }

>

>              write_vec_element_i32(s, tcg_res, rd, pass, MO_32);



--
Alex Bennée
diff mbox series

Patch

diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
index c928c4787c..64a2c2df59 100644
--- a/target/arm/translate-a64.c
+++ b/target/arm/translate-a64.c
@@ -9753,6 +9753,66 @@  static void disas_simd_3same_float(DisasContext *s, uint32_t insn)
     }
 }
 
+static void gen_mla8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
+{
+    gen_helper_neon_mul_u8(a, a, b);
+    gen_helper_neon_add_u8(d, d, a);
+}
+
+static void gen_mla16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
+{
+    gen_helper_neon_mul_u16(a, a, b);
+    gen_helper_neon_add_u16(d, d, a);
+}
+
+static void gen_mla32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
+{
+    tcg_gen_mul_i32(a, a, b);
+    tcg_gen_add_i32(d, d, a);
+}
+
+static void gen_mla64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
+{
+    tcg_gen_mul_i64(a, a, b);
+    tcg_gen_add_i64(d, d, a);
+}
+
+static void gen_mla_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
+{
+    tcg_gen_mul_vec(vece, a, a, b);
+    tcg_gen_add_vec(vece, d, d, a);
+}
+
+static void gen_mls8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
+{
+    gen_helper_neon_mul_u8(a, a, b);
+    gen_helper_neon_sub_u8(d, d, a);
+}
+
+static void gen_mls16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
+{
+    gen_helper_neon_mul_u16(a, a, b);
+    gen_helper_neon_sub_u16(d, d, a);
+}
+
+static void gen_mls32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b)
+{
+    tcg_gen_mul_i32(a, a, b);
+    tcg_gen_sub_i32(d, d, a);
+}
+
+static void gen_mls64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b)
+{
+    tcg_gen_mul_i64(a, a, b);
+    tcg_gen_sub_i64(d, d, a);
+}
+
+static void gen_mls_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b)
+{
+    tcg_gen_mul_vec(vece, a, a, b);
+    tcg_gen_sub_vec(vece, d, d, a);
+}
+
 /* Integer op subgroup of C3.6.16. */
 static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
 {
@@ -9771,6 +9831,52 @@  static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
           .prefer_i64 = TCG_TARGET_REG_BITS == 64,
           .vece = MO_64 },
     };
+    static const GVecGen3 mla_op[4] = {
+        { .fni4 = gen_mla8_i32,
+          .fniv = gen_mla_vec,
+          .opc = INDEX_op_mul_vec,
+          .load_dest = true,
+          .vece = MO_8 },
+        { .fni4 = gen_mla16_i32,
+          .fniv = gen_mla_vec,
+          .opc = INDEX_op_mul_vec,
+          .load_dest = true,
+          .vece = MO_16 },
+        { .fni4 = gen_mla32_i32,
+          .fniv = gen_mla_vec,
+          .opc = INDEX_op_mul_vec,
+          .load_dest = true,
+          .vece = MO_32 },
+        { .fni8 = gen_mla64_i64,
+          .fniv = gen_mla_vec,
+          .opc = INDEX_op_mul_vec,
+          .prefer_i64 = TCG_TARGET_REG_BITS == 64,
+          .load_dest = true,
+          .vece = MO_64 },
+    };
+    static const GVecGen3 mls_op[4] = {
+        { .fni4 = gen_mls8_i32,
+          .fniv = gen_mls_vec,
+          .opc = INDEX_op_mul_vec,
+          .load_dest = true,
+          .vece = MO_8 },
+        { .fni4 = gen_mls16_i32,
+          .fniv = gen_mls_vec,
+          .opc = INDEX_op_mul_vec,
+          .load_dest = true,
+          .vece = MO_16 },
+        { .fni4 = gen_mls32_i32,
+          .fniv = gen_mls_vec,
+          .opc = INDEX_op_mul_vec,
+          .load_dest = true,
+          .vece = MO_32 },
+        { .fni8 = gen_mls64_i64,
+          .fniv = gen_mls_vec,
+          .opc = INDEX_op_mul_vec,
+          .prefer_i64 = TCG_TARGET_REG_BITS == 64,
+          .load_dest = true,
+          .vece = MO_64 },
+    };
 
     int is_q = extract32(insn, 30, 1);
     int u = extract32(insn, 29, 1);
@@ -9828,6 +9934,19 @@  static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
             gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_add, size);
         }
         return;
+    case 0x13: /* MUL, PMUL */
+        if (!u) { /* MUL */
+            gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_mul, size);
+            return;
+        }
+        break;
+    case 0x12: /* MLA, MLS */
+        if (u) {
+            gen_gvec_op3(s, is_q, rd, rn, rm, &mls_op[size]);
+        } else {
+            gen_gvec_op3(s, is_q, rd, rn, rm, &mla_op[size]);
+        }
+        return;
     case 0x11:
         if (!u) { /* CMTST */
             gen_gvec_op3(s, is_q, rd, rn, rm, &cmtst_op[size]);
@@ -10002,23 +10121,10 @@  static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
                 break;
             }
             case 0x13: /* MUL, PMUL */
-                if (u) {
-                    /* PMUL */
-                    assert(size == 0);
-                    genfn = gen_helper_neon_mul_p8;
-                    break;
-                }
-                /* fall through : MUL */
-            case 0x12: /* MLA, MLS */
-            {
-                static NeonGenTwoOpFn * const fns[3] = {
-                    gen_helper_neon_mul_u8,
-                    gen_helper_neon_mul_u16,
-                    tcg_gen_mul_i32,
-                };
-                genfn = fns[size];
+                assert(u); /* PMUL */
+                assert(size == 0);
+                genfn = gen_helper_neon_mul_p8;
                 break;
-            }
             case 0x16: /* SQDMULH, SQRDMULH */
             {
                 static NeonGenTwoOpEnvFn * const fns[2][2] = {
@@ -10039,18 +10145,16 @@  static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
                 genfn(tcg_res, tcg_op1, tcg_op2);
             }
 
-            if (opcode == 0xf || opcode == 0x12) {
-                /* SABA, UABA, MLA, MLS: accumulating ops */
-                static NeonGenTwoOpFn * const fns[3][2] = {
-                    { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 },
-                    { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 },
-                    { tcg_gen_add_i32, tcg_gen_sub_i32 },
+            if (opcode == 0xf) {
+                /* SABA, UABA: accumulating ops */
+                static NeonGenTwoOpFn * const fns[3] = {
+                    gen_helper_neon_add_u8,
+                    gen_helper_neon_add_u16,
+                    tcg_gen_add_i32,
                 };
-                bool is_sub = (opcode == 0x12 && u); /* MLS */
 
-                genfn = fns[size][is_sub];
                 read_vec_element_i32(s, tcg_op1, rd, pass, MO_32);
-                genfn(tcg_res, tcg_op1, tcg_res);
+                fns[size](tcg_res, tcg_op1, tcg_res);
             }
 
             write_vec_element_i32(s, tcg_res, rd, pass, MO_32);