mirror of
https://github.com/xemu-project/xemu.git
synced 2024-11-23 19:49:43 +00:00
target/arm: implement SM3 instructions
This implements emulation of the new SM3 instructions that have been added as an optional extension to the ARMv8 Crypto Extensions in ARM v8.2. Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> Message-id: 20180207111729.15737-4-ard.biesheuvel@linaro.org Reviewed-by: Peter Maydell <peter.maydell@linaro.org> Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
This commit is contained in:
parent
cd270ade74
commit
80d6f4c6bb
@ -1343,6 +1343,7 @@ enum arm_features {
|
|||||||
ARM_FEATURE_SVE, /* has Scalable Vector Extension */
|
ARM_FEATURE_SVE, /* has Scalable Vector Extension */
|
||||||
ARM_FEATURE_V8_SHA512, /* implements SHA512 part of v8 Crypto Extensions */
|
ARM_FEATURE_V8_SHA512, /* implements SHA512 part of v8 Crypto Extensions */
|
||||||
ARM_FEATURE_V8_SHA3, /* implements SHA3 part of v8 Crypto Extensions */
|
ARM_FEATURE_V8_SHA3, /* implements SHA3 part of v8 Crypto Extensions */
|
||||||
|
ARM_FEATURE_V8_SM3, /* implements SM3 part of v8 Crypto Extensions */
|
||||||
};
|
};
|
||||||
|
|
||||||
static inline int arm_feature(CPUARMState *env, int feature)
|
static inline int arm_feature(CPUARMState *env, int feature)
|
||||||
|
@ -507,3 +507,99 @@ void HELPER(crypto_sha512su1)(void *vd, void *vn, void *vm)
|
|||||||
rd[0] += s1_512(rn[0]) + rm[0];
|
rd[0] += s1_512(rn[0]) + rm[0];
|
||||||
rd[1] += s1_512(rn[1]) + rm[1];
|
rd[1] += s1_512(rn[1]) + rm[1];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void HELPER(crypto_sm3partw1)(void *vd, void *vn, void *vm)
|
||||||
|
{
|
||||||
|
uint64_t *rd = vd;
|
||||||
|
uint64_t *rn = vn;
|
||||||
|
uint64_t *rm = vm;
|
||||||
|
union CRYPTO_STATE d = { .l = { rd[0], rd[1] } };
|
||||||
|
union CRYPTO_STATE n = { .l = { rn[0], rn[1] } };
|
||||||
|
union CRYPTO_STATE m = { .l = { rm[0], rm[1] } };
|
||||||
|
uint32_t t;
|
||||||
|
|
||||||
|
t = CR_ST_WORD(d, 0) ^ CR_ST_WORD(n, 0) ^ ror32(CR_ST_WORD(m, 1), 17);
|
||||||
|
CR_ST_WORD(d, 0) = t ^ ror32(t, 17) ^ ror32(t, 9);
|
||||||
|
|
||||||
|
t = CR_ST_WORD(d, 1) ^ CR_ST_WORD(n, 1) ^ ror32(CR_ST_WORD(m, 2), 17);
|
||||||
|
CR_ST_WORD(d, 1) = t ^ ror32(t, 17) ^ ror32(t, 9);
|
||||||
|
|
||||||
|
t = CR_ST_WORD(d, 2) ^ CR_ST_WORD(n, 2) ^ ror32(CR_ST_WORD(m, 3), 17);
|
||||||
|
CR_ST_WORD(d, 2) = t ^ ror32(t, 17) ^ ror32(t, 9);
|
||||||
|
|
||||||
|
t = CR_ST_WORD(d, 3) ^ CR_ST_WORD(n, 3) ^ ror32(CR_ST_WORD(d, 0), 17);
|
||||||
|
CR_ST_WORD(d, 3) = t ^ ror32(t, 17) ^ ror32(t, 9);
|
||||||
|
|
||||||
|
rd[0] = d.l[0];
|
||||||
|
rd[1] = d.l[1];
|
||||||
|
}
|
||||||
|
|
||||||
|
void HELPER(crypto_sm3partw2)(void *vd, void *vn, void *vm)
|
||||||
|
{
|
||||||
|
uint64_t *rd = vd;
|
||||||
|
uint64_t *rn = vn;
|
||||||
|
uint64_t *rm = vm;
|
||||||
|
union CRYPTO_STATE d = { .l = { rd[0], rd[1] } };
|
||||||
|
union CRYPTO_STATE n = { .l = { rn[0], rn[1] } };
|
||||||
|
union CRYPTO_STATE m = { .l = { rm[0], rm[1] } };
|
||||||
|
uint32_t t = CR_ST_WORD(n, 0) ^ ror32(CR_ST_WORD(m, 0), 25);
|
||||||
|
|
||||||
|
CR_ST_WORD(d, 0) ^= t;
|
||||||
|
CR_ST_WORD(d, 1) ^= CR_ST_WORD(n, 1) ^ ror32(CR_ST_WORD(m, 1), 25);
|
||||||
|
CR_ST_WORD(d, 2) ^= CR_ST_WORD(n, 2) ^ ror32(CR_ST_WORD(m, 2), 25);
|
||||||
|
CR_ST_WORD(d, 3) ^= CR_ST_WORD(n, 3) ^ ror32(CR_ST_WORD(m, 3), 25) ^
|
||||||
|
ror32(t, 17) ^ ror32(t, 2) ^ ror32(t, 26);
|
||||||
|
|
||||||
|
rd[0] = d.l[0];
|
||||||
|
rd[1] = d.l[1];
|
||||||
|
}
|
||||||
|
|
||||||
|
void HELPER(crypto_sm3tt)(void *vd, void *vn, void *vm, uint32_t imm2,
|
||||||
|
uint32_t opcode)
|
||||||
|
{
|
||||||
|
uint64_t *rd = vd;
|
||||||
|
uint64_t *rn = vn;
|
||||||
|
uint64_t *rm = vm;
|
||||||
|
union CRYPTO_STATE d = { .l = { rd[0], rd[1] } };
|
||||||
|
union CRYPTO_STATE n = { .l = { rn[0], rn[1] } };
|
||||||
|
union CRYPTO_STATE m = { .l = { rm[0], rm[1] } };
|
||||||
|
uint32_t t;
|
||||||
|
|
||||||
|
assert(imm2 < 4);
|
||||||
|
|
||||||
|
if (opcode == 0 || opcode == 2) {
|
||||||
|
/* SM3TT1A, SM3TT2A */
|
||||||
|
t = par(CR_ST_WORD(d, 3), CR_ST_WORD(d, 2), CR_ST_WORD(d, 1));
|
||||||
|
} else if (opcode == 1) {
|
||||||
|
/* SM3TT1B */
|
||||||
|
t = maj(CR_ST_WORD(d, 3), CR_ST_WORD(d, 2), CR_ST_WORD(d, 1));
|
||||||
|
} else if (opcode == 3) {
|
||||||
|
/* SM3TT2B */
|
||||||
|
t = cho(CR_ST_WORD(d, 3), CR_ST_WORD(d, 2), CR_ST_WORD(d, 1));
|
||||||
|
} else {
|
||||||
|
g_assert_not_reached();
|
||||||
|
}
|
||||||
|
|
||||||
|
t += CR_ST_WORD(d, 0) + CR_ST_WORD(m, imm2);
|
||||||
|
|
||||||
|
CR_ST_WORD(d, 0) = CR_ST_WORD(d, 1);
|
||||||
|
|
||||||
|
if (opcode < 2) {
|
||||||
|
/* SM3TT1A, SM3TT1B */
|
||||||
|
t += CR_ST_WORD(n, 3) ^ ror32(CR_ST_WORD(d, 3), 20);
|
||||||
|
|
||||||
|
CR_ST_WORD(d, 1) = ror32(CR_ST_WORD(d, 2), 23);
|
||||||
|
} else {
|
||||||
|
/* SM3TT2A, SM3TT2B */
|
||||||
|
t += CR_ST_WORD(n, 3);
|
||||||
|
t ^= rol32(t, 9) ^ rol32(t, 17);
|
||||||
|
|
||||||
|
CR_ST_WORD(d, 1) = ror32(CR_ST_WORD(d, 2), 13);
|
||||||
|
}
|
||||||
|
|
||||||
|
CR_ST_WORD(d, 2) = CR_ST_WORD(d, 3);
|
||||||
|
CR_ST_WORD(d, 3) = t;
|
||||||
|
|
||||||
|
rd[0] = d.l[0];
|
||||||
|
rd[1] = d.l[1];
|
||||||
|
}
|
||||||
|
@ -539,6 +539,10 @@ DEF_HELPER_FLAGS_3(crypto_sha512h2, TCG_CALL_NO_RWG, void, ptr, ptr, ptr)
|
|||||||
DEF_HELPER_FLAGS_2(crypto_sha512su0, TCG_CALL_NO_RWG, void, ptr, ptr)
|
DEF_HELPER_FLAGS_2(crypto_sha512su0, TCG_CALL_NO_RWG, void, ptr, ptr)
|
||||||
DEF_HELPER_FLAGS_3(crypto_sha512su1, TCG_CALL_NO_RWG, void, ptr, ptr, ptr)
|
DEF_HELPER_FLAGS_3(crypto_sha512su1, TCG_CALL_NO_RWG, void, ptr, ptr, ptr)
|
||||||
|
|
||||||
|
DEF_HELPER_FLAGS_5(crypto_sm3tt, TCG_CALL_NO_RWG, void, ptr, ptr, ptr, i32, i32)
|
||||||
|
DEF_HELPER_FLAGS_3(crypto_sm3partw1, TCG_CALL_NO_RWG, void, ptr, ptr, ptr)
|
||||||
|
DEF_HELPER_FLAGS_3(crypto_sm3partw2, TCG_CALL_NO_RWG, void, ptr, ptr, ptr)
|
||||||
|
|
||||||
DEF_HELPER_FLAGS_3(crc32, TCG_CALL_NO_RWG_SE, i32, i32, i32, i32)
|
DEF_HELPER_FLAGS_3(crc32, TCG_CALL_NO_RWG_SE, i32, i32, i32, i32)
|
||||||
DEF_HELPER_FLAGS_3(crc32c, TCG_CALL_NO_RWG_SE, i32, i32, i32, i32)
|
DEF_HELPER_FLAGS_3(crc32c, TCG_CALL_NO_RWG_SE, i32, i32, i32, i32)
|
||||||
DEF_HELPER_2(dc_zva, void, env, i64)
|
DEF_HELPER_2(dc_zva, void, env, i64)
|
||||||
|
@ -11623,9 +11623,20 @@ static void disas_crypto_three_reg_sha512(DisasContext *s, uint32_t insn)
|
|||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
switch (opcode) {
|
||||||
|
case 0: /* SM3PARTW1 */
|
||||||
|
feature = ARM_FEATURE_V8_SM3;
|
||||||
|
genfn = gen_helper_crypto_sm3partw1;
|
||||||
|
break;
|
||||||
|
case 1: /* SM3PARTW2 */
|
||||||
|
feature = ARM_FEATURE_V8_SM3;
|
||||||
|
genfn = gen_helper_crypto_sm3partw2;
|
||||||
|
break;
|
||||||
|
default:
|
||||||
unallocated_encoding(s);
|
unallocated_encoding(s);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if (!arm_dc_feature(s, feature)) {
|
if (!arm_dc_feature(s, feature)) {
|
||||||
unallocated_encoding(s);
|
unallocated_encoding(s);
|
||||||
@ -11737,6 +11748,9 @@ static void disas_crypto_four_reg(DisasContext *s, uint32_t insn)
|
|||||||
case 1: /* BCAX */
|
case 1: /* BCAX */
|
||||||
feature = ARM_FEATURE_V8_SHA3;
|
feature = ARM_FEATURE_V8_SHA3;
|
||||||
break;
|
break;
|
||||||
|
case 2: /* SM3SS1 */
|
||||||
|
feature = ARM_FEATURE_V8_SM3;
|
||||||
|
break;
|
||||||
default:
|
default:
|
||||||
unallocated_encoding(s);
|
unallocated_encoding(s);
|
||||||
return;
|
return;
|
||||||
@ -11784,7 +11798,33 @@ static void disas_crypto_four_reg(DisasContext *s, uint32_t insn)
|
|||||||
tcg_temp_free_i64(tcg_res[0]);
|
tcg_temp_free_i64(tcg_res[0]);
|
||||||
tcg_temp_free_i64(tcg_res[1]);
|
tcg_temp_free_i64(tcg_res[1]);
|
||||||
} else {
|
} else {
|
||||||
g_assert_not_reached();
|
TCGv_i32 tcg_op1, tcg_op2, tcg_op3, tcg_res, tcg_zero;
|
||||||
|
|
||||||
|
tcg_op1 = tcg_temp_new_i32();
|
||||||
|
tcg_op2 = tcg_temp_new_i32();
|
||||||
|
tcg_op3 = tcg_temp_new_i32();
|
||||||
|
tcg_res = tcg_temp_new_i32();
|
||||||
|
tcg_zero = tcg_const_i32(0);
|
||||||
|
|
||||||
|
read_vec_element_i32(s, tcg_op1, rn, 3, MO_32);
|
||||||
|
read_vec_element_i32(s, tcg_op2, rm, 3, MO_32);
|
||||||
|
read_vec_element_i32(s, tcg_op3, ra, 3, MO_32);
|
||||||
|
|
||||||
|
tcg_gen_rotri_i32(tcg_res, tcg_op1, 20);
|
||||||
|
tcg_gen_add_i32(tcg_res, tcg_res, tcg_op2);
|
||||||
|
tcg_gen_add_i32(tcg_res, tcg_res, tcg_op3);
|
||||||
|
tcg_gen_rotri_i32(tcg_res, tcg_res, 25);
|
||||||
|
|
||||||
|
write_vec_element_i32(s, tcg_zero, rd, 0, MO_32);
|
||||||
|
write_vec_element_i32(s, tcg_zero, rd, 1, MO_32);
|
||||||
|
write_vec_element_i32(s, tcg_zero, rd, 2, MO_32);
|
||||||
|
write_vec_element_i32(s, tcg_res, rd, 3, MO_32);
|
||||||
|
|
||||||
|
tcg_temp_free_i32(tcg_op1);
|
||||||
|
tcg_temp_free_i32(tcg_op2);
|
||||||
|
tcg_temp_free_i32(tcg_op3);
|
||||||
|
tcg_temp_free_i32(tcg_res);
|
||||||
|
tcg_temp_free_i32(tcg_zero);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -11833,6 +11873,47 @@ static void disas_crypto_xar(DisasContext *s, uint32_t insn)
|
|||||||
tcg_temp_free_i64(tcg_res[1]);
|
tcg_temp_free_i64(tcg_res[1]);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Crypto three-reg imm2
|
||||||
|
* 31 21 20 16 15 14 13 12 11 10 9 5 4 0
|
||||||
|
* +-----------------------+------+-----+------+--------+------+------+
|
||||||
|
* | 1 1 0 0 1 1 1 0 0 1 0 | Rm | 1 0 | imm2 | opcode | Rn | Rd |
|
||||||
|
* +-----------------------+------+-----+------+--------+------+------+
|
||||||
|
*/
|
||||||
|
static void disas_crypto_three_reg_imm2(DisasContext *s, uint32_t insn)
|
||||||
|
{
|
||||||
|
int opcode = extract32(insn, 10, 2);
|
||||||
|
int imm2 = extract32(insn, 12, 2);
|
||||||
|
int rm = extract32(insn, 16, 5);
|
||||||
|
int rn = extract32(insn, 5, 5);
|
||||||
|
int rd = extract32(insn, 0, 5);
|
||||||
|
TCGv_ptr tcg_rd_ptr, tcg_rn_ptr, tcg_rm_ptr;
|
||||||
|
TCGv_i32 tcg_imm2, tcg_opcode;
|
||||||
|
|
||||||
|
if (!arm_dc_feature(s, ARM_FEATURE_V8_SM3)) {
|
||||||
|
unallocated_encoding(s);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!fp_access_check(s)) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
tcg_rd_ptr = vec_full_reg_ptr(s, rd);
|
||||||
|
tcg_rn_ptr = vec_full_reg_ptr(s, rn);
|
||||||
|
tcg_rm_ptr = vec_full_reg_ptr(s, rm);
|
||||||
|
tcg_imm2 = tcg_const_i32(imm2);
|
||||||
|
tcg_opcode = tcg_const_i32(opcode);
|
||||||
|
|
||||||
|
gen_helper_crypto_sm3tt(tcg_rd_ptr, tcg_rn_ptr, tcg_rm_ptr, tcg_imm2,
|
||||||
|
tcg_opcode);
|
||||||
|
|
||||||
|
tcg_temp_free_ptr(tcg_rd_ptr);
|
||||||
|
tcg_temp_free_ptr(tcg_rn_ptr);
|
||||||
|
tcg_temp_free_ptr(tcg_rm_ptr);
|
||||||
|
tcg_temp_free_i32(tcg_imm2);
|
||||||
|
tcg_temp_free_i32(tcg_opcode);
|
||||||
|
}
|
||||||
|
|
||||||
/* C3.6 Data processing - SIMD, inc Crypto
|
/* C3.6 Data processing - SIMD, inc Crypto
|
||||||
*
|
*
|
||||||
* As the decode gets a little complex we are using a table based
|
* As the decode gets a little complex we are using a table based
|
||||||
@ -11866,6 +11947,7 @@ static const AArch64DecodeTable data_proc_simd[] = {
|
|||||||
{ 0xcec08000, 0xfffff000, disas_crypto_two_reg_sha512 },
|
{ 0xcec08000, 0xfffff000, disas_crypto_two_reg_sha512 },
|
||||||
{ 0xce000000, 0xff808000, disas_crypto_four_reg },
|
{ 0xce000000, 0xff808000, disas_crypto_four_reg },
|
||||||
{ 0xce800000, 0xffe00000, disas_crypto_xar },
|
{ 0xce800000, 0xffe00000, disas_crypto_xar },
|
||||||
|
{ 0xce408000, 0xffe0c000, disas_crypto_three_reg_imm2 },
|
||||||
{ 0x00000000, 0x00000000, NULL }
|
{ 0x00000000, 0x00000000, NULL }
|
||||||
};
|
};
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user