qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-devel] [PATCH v2] target-arm: Implement ARMv8 VSEL instruction


From: Will Newton
Subject: Re: [Qemu-devel] [PATCH v2] target-arm: Implement ARMv8 VSEL instruction.
Date: Thu, 3 Oct 2013 15:31:46 +0100

On 3 October 2013 13:59, Peter Maydell <address@hidden> wrote:
> On 3 October 2013 21:51, Will Newton <address@hidden> wrote:
>>
>> This adds support for the VSEL floating point selection instruction
>> which was added in ARMv8. It is based on the previous patch[1] from
>> Mans Rullgard, but attempts to address the feedback given on that patch.
>>
>> [1] http://lists.nongnu.org/archive/html/qemu-devel/2013-06/msg03117.html
>
> This sort of commentary about previous patch versions should go below
> the '---', not in the commit message.
>
>>
>> Signed-off-by: Will Newton <address@hidden>
>> ---
>>  target-arm/translate.c | 105 
>> +++++++++++++++++++++++++++++++++++++++++++++++++
>>  1 file changed, 105 insertions(+)
>>
>> Changes in v2:
>>  - Integrate vsel decoding into disas_vfp_insn
>>
>> diff --git a/target-arm/translate.c b/target-arm/translate.c
>> index 998bde2..5e49334 100644
>> --- a/target-arm/translate.c
>> +++ b/target-arm/translate.c
>> @@ -2880,6 +2880,98 @@ static int disas_vfp_insn(CPUARMState * env, 
>> DisasContext *s, uint32_t insn)
>>                  rm = VFP_SREG_M(insn);
>>              }
>>
>> +            if ((insn & 0x0f800e50) == 0x0e000a00) {
>> +                /* vsel */
>> +                uint32_t cc = (insn >> 20) & 3;
>> +                TCGv_i32 tmp, zero;
>> +
>> +                /* ARMv8 VFP.  */
>> +                if (!arm_feature(env, ARM_FEATURE_V8))
>> +                    return 1;
>
> scripts/checkpatch.pl will tell you that omitting the braces
> is a coding style violation.

Ok, I'll fix that.

>> +
>> +                zero = tcg_const_tl(0);
>> +
>> +                if (dp) {
>> +                    TCGv_i64 ftmp1, ftmp2, ftmp3;
>> +
>> +                    ftmp1 = tcg_temp_new_i64();
>> +                    ftmp2 = tcg_temp_new_i64();
>> +                    ftmp3 = tcg_temp_new_i64();
>> +                    tcg_gen_ld_f64(ftmp1, cpu_env, vfp_reg_offset(dp, rn));
>> +                    tcg_gen_ld_f64(ftmp2, cpu_env, vfp_reg_offset(dp, rm));
>> +                    switch (cc) {
>> +                    case 0: /* eq: Z */
>> +                        tcg_gen_movcond_i64(TCG_COND_EQ, ftmp3, cpu_ZF, 
>> zero,
>> +                                            ftmp1, ftmp2);
>> +                        break;
>> +                    case 1: /* vs: V */
>> +                        tcg_gen_movcond_i64(TCG_COND_LT, ftmp3, cpu_VF, 
>> zero,
>> +                                            ftmp1, ftmp2);
>> +                        break;
>> +                    case 2: /* ge: N == V -> N ^ V == 0 */
>> +                        tmp = tcg_temp_new_i32();
>> +                        tcg_gen_xor_i32(tmp, cpu_VF, cpu_NF);
>> +                        tcg_gen_movcond_i64(TCG_COND_GE, ftmp3, tmp, zero,
>> +                                            ftmp1, ftmp2);
>> +                        tcg_temp_free_i32(tmp);
>> +                        break;
>> +                    case 3: /* gt: !Z && N == V */
>> +                        tcg_gen_movcond_i64(TCG_COND_NE, ftmp3, cpu_ZF, 
>> zero,
>> +                                            ftmp1, ftmp2);
>> +                        tmp = tcg_temp_new_i32();
>> +                        tcg_gen_xor_i32(tmp, cpu_VF, cpu_NF);
>> +                        tcg_gen_movcond_i64(TCG_COND_GE, ftmp3, tmp, zero,
>> +                                            ftmp3, ftmp2);
>> +                        tcg_temp_free_i32(tmp);
>> +                        break;
>> +                    }
>> +                    tcg_gen_st_f64(ftmp3, cpu_env, vfp_reg_offset(dp, rd));
>> +                    tcg_temp_free_i64(ftmp1);
>> +                    tcg_temp_free_i64(ftmp2);
>> +                    tcg_temp_free_i64(ftmp3);
>> +                } else {
>> +                    TCGv_i32 ftmp1, ftmp2, ftmp3;
>> +
>> +                    ftmp1 = tcg_temp_new_i32();
>> +                    ftmp2 = tcg_temp_new_i32();
>> +                    ftmp3 = tcg_temp_new_i32();
>> +                    tcg_gen_ld_f32(ftmp1, cpu_env, vfp_reg_offset(dp, rn));
>> +                    tcg_gen_ld_f32(ftmp2, cpu_env, vfp_reg_offset(dp, rm));
>> +                    switch (cc) {
>> +                    case 0: /* eq: Z */
>> +                        tcg_gen_movcond_i32(TCG_COND_EQ, ftmp3, cpu_ZF, 
>> zero,
>> +                                            ftmp1, ftmp2);
>> +                        break;
>> +                    case 1: /* vs: V */
>> +                        tcg_gen_movcond_i32(TCG_COND_LT, ftmp3, cpu_VF, 
>> zero,
>> +                                            ftmp1, ftmp2);
>> +                        break;
>> +                    case 2: /* ge: N == V -> N ^ V == 0 */
>> +                        tmp = tcg_temp_new_i32();
>> +                        tcg_gen_xor_i32(tmp, cpu_VF, cpu_NF);
>> +                        tcg_gen_movcond_i32(TCG_COND_GE, ftmp3, tmp, zero,
>> +                                            ftmp1, ftmp2);
>> +                        tcg_temp_free_i32(tmp);
>> +                        break;
>> +                    case 3: /* gt: !Z && N == V */
>> +                        tcg_gen_movcond_i32(TCG_COND_NE, ftmp3, cpu_ZF, 
>> zero,
>> +                                            ftmp1, ftmp2);
>> +                        tmp = tcg_temp_new_i32();
>> +                        tcg_gen_xor_i32(tmp, cpu_VF, cpu_NF);
>> +                        tcg_gen_movcond_i32(TCG_COND_GE, ftmp3, tmp, zero,
>> +                                            ftmp3, ftmp2);
>> +                        tcg_temp_free_i32(tmp);
>> +                        break;
>> +                    }
>> +                    tcg_gen_st_f32(ftmp3, cpu_env, vfp_reg_offset(dp, rd));
>> +                    tcg_temp_free_i32(ftmp1);
>> +                    tcg_temp_free_i32(ftmp2);
>> +                    tcg_temp_free_i32(ftmp3);
>> +                }
>> +
>> +                return 0;
>> +            }
>> +
>>              veclen = s->vec_len;
>>              if (op == 15 && rn > 3)
>>                  veclen = 0;
>> @@ -6756,6 +6848,13 @@ static void disas_arm_insn(CPUARMState * env, 
>> DisasContext *s)
>>                  goto illegal_op;
>>              return;
>>          }
>> +       if ((insn & 0x0f800e50) == 0x0e000a00) {
>> +           /* ARMv8 VFP.  */
>> +           ARCH(8);
>> +
>> +           if (disas_vfp_insn(env, s, insn))
>> +               goto illegal_op;
>> +       }
>
> This isn't what I meant. If our decoding matches up with the ARM ARM
> then this instruction pattern should already fall into disas_vfp_insn(),
> and we shouldn't need an extra check and call. (If it's not correct then
> we should adjust our decode so it does.)

I'll respin the patch pulling the calls to disas_vfp_insn up a level
which I think you alluded to in the original review. It still needs an
additional call to disas_vfp_insn in the ARM case as condition code ==
0xf is dealt with separately from the others. Let me know if this is
not what you were looking for.

Thanks,

-- 
Will Newton
Toolchain Working Group, Linaro



reply via email to

[Prev in Thread] Current Thread [Next in Thread]