12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088 |
- From 6a82af7b913410fe993671d033d1ee4402d10d4a Mon Sep 17 00:00:00 2001
- From: Frank Chang <frank.chang@sifive.com>
- Date: Mon, 28 Sep 2020 16:23:30 +0800
- Subject: [PATCH 022/107] target/riscv: rvv-1.0: update check functions
- Update check functions with RVV 1.0 rules.
- Signed-off-by: Frank Chang <frank.chang@sifive.com>
- Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
- ---
- target/riscv/insn_trans/trans_rvv.c.inc | 739 ++++++++++++++++--------
- 1 file changed, 505 insertions(+), 234 deletions(-)
- diff --git a/target/riscv/insn_trans/trans_rvv.c.inc b/target/riscv/insn_trans/trans_rvv.c.inc
- index 46e18a62b5..f666c64bbe 100644
- --- a/target/riscv/insn_trans/trans_rvv.c.inc
- +++ b/target/riscv/insn_trans/trans_rvv.c.inc
- @@ -19,11 +19,124 @@
- #include "tcg/tcg-gvec-desc.h"
- #include "internals.h"
-
- +static inline bool is_overlapped(const int8_t astart, int8_t asize,
- + const int8_t bstart, int8_t bsize)
- +{
- + const int8_t aend = astart + asize;
- + const int8_t bend = bstart + bsize;
- +
- + return MAX(aend, bend) - MIN(astart, bstart) < asize + bsize;
- +}
- +
- +static bool require_rvv(DisasContext *s)
- +{
- + return s->mstatus_vs != 0;
- +}
- +
- +static bool require_rvf(DisasContext *s)
- +{
- + if (s->mstatus_fs == 0) {
- + return false;
- + }
- +
- + switch (s->sew) {
- + case MO_16:
- + case MO_32:
- + return has_ext(s, RVF);
- + case MO_64:
- + return has_ext(s, RVD);
- + default:
- + return false;
- + }
- +}
- +
- +static bool require_scale_rvf(DisasContext *s)
- +{
- + if (s->mstatus_fs == 0) {
- + return false;
- + }
- +
- + switch (s->sew) {
- + case MO_8:
- + case MO_16:
- + return has_ext(s, RVF);
- + case MO_32:
- + return has_ext(s, RVD);
- + default:
- + return false;
- + }
- +}
- +
- +/* Destination vector register group cannot overlap source mask register. */
- +static bool require_vm(int vm, int vd)
- +{
- + return (vm != 0 || vd != 0);
- +}
- +
- +static bool require_nf(int vd, int nf, int lmul)
- +{
- + int size = nf << MAX(lmul, 0);
- + return size <= 8 && vd + size <= 32;
- +}
- +
- +/*
- + * Vector register should aligned with the passed-in LMUL (EMUL).
- + * If LMUL < 0, i.e. fractional LMUL, any vector register is allowed.
- + */
- +static bool require_align(const int8_t val, const int8_t lmul)
- +{
- + return lmul <= 0 || extract32(val, 0, lmul) == 0;
- +}
- +
- +/*
- + * A destination vector register group can overlap a source vector
- + * register group only if one of the following holds:
- + * 1. The destination EEW equals the source EEW.
- + * 2. The destination EEW is smaller than the source EEW and the overlap
- + * is in the lowest-numbered part of the source register group.
- + * 3. The destination EEW is greater than the source EEW, the source EMUL
- + * is at least 1, and the overlap is in the highest-numbered part of
- + * the destination register group.
- + * (Section 5.2)
- + *
- + * This function returns true if one of the following holds:
- + * * Destination vector register group does not overlap a source vector
- + * register group.
- + * * Rule 3 met.
- + * For rule 1, overlap is allowed so this function doesn't need to be called.
- + * For rule 2, (vd == vs). Caller has to check whether: (vd != vs) before
- + * calling this function.
- + */
- +static bool require_noover(const int8_t dst, const int8_t dst_lmul,
- + const int8_t src, const int8_t src_lmul)
- +{
- + int8_t dst_size = dst_lmul <= 0 ? 1 : 1 << dst_lmul;
- + int8_t src_size = src_lmul <= 0 ? 1 : 1 << src_lmul;
- +
- + /* Destination EEW is greater than the source EEW, check rule 3. */
- + if (dst_size > src_size) {
- + if (dst < src &&
- + src_lmul >= 0 &&
- + is_overlapped(dst, dst_size, src, src_size) &&
- + !is_overlapped(dst, dst_size, src + src_size, src_size)) {
- + return true;
- + }
- + }
- +
- + return !is_overlapped(dst, dst_size, src, src_size);
- +}
- +
- +static bool require_noover_seg(const int8_t dst, const int8_t nf,
- + const int8_t src)
- +{
- + return !is_overlapped(dst, nf, src, 1);
- +}
- +
- static bool trans_vsetvl(DisasContext *ctx, arg_vsetvl *a)
- {
- TCGv s1, s2, dst;
-
- - if (!has_ext(ctx, RVV)) {
- + if (!require_rvv(ctx) || !has_ext(ctx, RVV)) {
- return false;
- }
-
- @@ -56,7 +169,7 @@ static bool trans_vsetvli(DisasContext *ctx, arg_vsetvli *a)
- {
- TCGv s1, s2, dst;
-
- - if (!has_ext(ctx, RVV)) {
- + if (!require_rvv(ctx) || !has_ext(ctx, RVV)) {
- return false;
- }
-
- @@ -100,54 +213,246 @@ static bool vext_check_isa_ill(DisasContext *s)
- return !s->vill;
- }
-
- +static bool vext_check_ss(DisasContext *s, int vd, int vs, int vm)
- +{
- + return require_vm(vm, vd) &&
- + require_align(vd, s->lmul) &&
- + require_align(vs, s->lmul);
- +}
- +
- /*
- - * There are two rules check here.
- + * Check function for vector instruction with format:
- + * single-width result and single-width sources (SEW = SEW op SEW)
- *
- - * 1. Vector register numbers are multiples of LMUL. (Section 3.2)
- + * Rules to be checked here:
- + * 1. Destination vector register group for a masked vector
- + * instruction cannot overlap the source mask register (v0).
- + * (Section 5.3)
- + * 2. Destination vector register number is multiples of LMUL.
- + * (Section 3.3.2)
- + * 3. Source (vs2, vs1) vector register number are multiples of LMUL.
- + * (Section 3.3.2)
- + */
- +static bool vext_check_sss(DisasContext *s, int vd, int vs1, int vs2, int vm)
- +{
- + return vext_check_ss(s, vd, vs2, vm) &&
- + require_align(vs1, s->lmul);
- +}
- +
- +static bool vext_check_ms(DisasContext *s, int vd, int vs)
- +{
- + bool ret = require_align(vs, s->lmul);
- + if (vd != vs) {
- + ret &= require_noover(vd, 0, vs, s->lmul);
- + }
- + return ret;
- +}
- +
- +/*
- + * Check function for maskable vector instruction with format:
- + * single-width result and single-width sources (SEW = SEW op SEW)
- *
- - * 2. For all widening instructions, the destination LMUL value must also be
- - * a supported LMUL value. (Section 11.2)
- + * Rules to be checked here:
- + * 1. Source (vs2, vs1) vector register number are multiples of LMUL.
- + * (Section 3.3.2)
- + * 2. Destination vector register cannot overlap a source vector
- + * register (vs2, vs1) group.
- + * (Section 5.2)
- + * 3. The destination vector register group for a masked vector
- + * instruction cannot overlap the source mask register (v0),
- + * unless the destination vector register is being written
- + * with a mask value (e.g., comparisons) or the scalar result
- + * of a reduction. (Section 5.3)
- */
- -static bool vext_check_reg(DisasContext *s, uint32_t reg, bool widen)
- +static bool vext_check_mss(DisasContext *s, int vd, int vs1, int vs2)
- {
- - /*
- - * The destination vector register group results are arranged as if both
- - * SEW and LMUL were at twice their current settings. (Section 11.2).
- - */
- - int legal = widen ? 2 << s->lmul : 1 << s->lmul;
- + bool ret = vext_check_ms(s, vd, vs2) &&
- + require_align(vs1, s->lmul);
- + if (vd != vs1) {
- + ret &= require_noover(vd, 0, vs1, s->lmul);
- + }
- + return ret;
- +}
-
- - return !((s->lmul == 0x3 && widen) || (reg % legal));
- +/*
- + * Common check function for vector widening instructions
- + * of double-width result (2*SEW).
- + *
- + * Rules to be checked here:
- + * 1. The largest vector register group used by an instruction
- + * can not be greater than 8 vector registers (Section 5.2):
- + * => LMUL < 8.
- + * => SEW < 64.
- + * 2. Destination vector register number is multiples of 2 * LMUL.
- + * (Section 3.3.2, 11.2)
- + * 3. Destination vector register group for a masked vector
- + * instruction cannot overlap the source mask register (v0).
- + * (Section 5.3)
- + */
- +static bool vext_wide_check_common(DisasContext *s, int vd, int vm)
- +{
- + return (s->lmul <= 2) &&
- + (s->sew < MO_64) &&
- + require_align(vd, s->lmul + 1) &&
- + require_vm(vm, vd);
- }
-
- /*
- - * There are two rules check here.
- + * Common check function for vector narrowing instructions
- + * of single-width result (SEW) and double-width source (2*SEW).
- + *
- + * Rules to be checked here:
- + * 1. The largest vector register group used by an instruction
- + * can not be greater than 8 vector registers (Section 5.2):
- + * => LMUL < 8.
- + * => SEW < 64.
- + * 2. Source vector register number is multiples of 2 * LMUL.
- + * (Section 3.3.2, 11.3)
- + * 3. Destination vector register number is multiples of LMUL.
- + * (Section 3.3.2, 11.3)
- + * 4. Destination vector register group for a masked vector
- + * instruction cannot overlap the source mask register (v0).
- + * (Section 5.3)
- + */
- +static bool vext_narrow_check_common(DisasContext *s, int vd, int vs2,
- + int vm)
- +{
- + return (s->lmul <= 2) &&
- + (s->sew < MO_64) &&
- + require_align(vs2, s->lmul + 1) &&
- + require_align(vd, s->lmul) &&
- + require_vm(vm, vd);
- +}
- +
- +static bool vext_check_ds(DisasContext *s, int vd, int vs, int vm)
- +{
- + return vext_wide_check_common(s, vd, vm) &&
- + require_align(vs, s->lmul) &&
- + require_noover(vd, s->lmul + 1, vs, s->lmul);
- +}
- +
- +static bool vext_check_dd(DisasContext *s, int vd, int vs, int vm)
- +{
- + return vext_wide_check_common(s, vd, vm) &&
- + require_align(vs, s->lmul + 1);
- +}
- +
- +/*
- + * Check function for vector instruction with format:
- + * double-width result and single-width sources (2*SEW = SEW op SEW)
- *
- - * 1. The destination vector register group for a masked vector instruction can
- - * only overlap the source mask register (v0) when LMUL=1. (Section 5.3)
- + * Rules to be checked here:
- + * 1. All rules in defined in widen common rules are applied.
- + * 2. Source (vs2, vs1) vector register number are multiples of LMUL.
- + * (Section 3.3.2)
- + * 3. Destination vector register cannot overlap a source vector
- + * register (vs2, vs1) group.
- + * (Section 5.2)
- + */
- +static bool vext_check_dss(DisasContext *s, int vd, int vs1, int vs2, int vm)
- +{
- + return vext_check_ds(s, vd, vs2, vm) &&
- + require_align(vs1, s->lmul) &&
- + require_noover(vd, s->lmul + 1, vs1, s->lmul);
- +}
- +
- +/*
- + * Check function for vector instruction with format:
- + * double-width result and double-width source1 and single-width
- + * source2 (2*SEW = 2*SEW op SEW)
- *
- - * 2. In widen instructions and some other insturctions, like vslideup.vx,
- - * there is no need to check whether LMUL=1.
- + * Rules to be checked here:
- + * 1. All rules in defined in widen common rules are applied.
- + * 2. Source 1 (vs2) vector register number is multiples of 2 * LMUL.
- + * (Section 3.3.2)
- + * 3. Source 2 (vs1) vector register number is multiples of LMUL.
- + * (Section 3.3.2)
- + * 4. Destination vector register cannot overlap a source vector
- + * register (vs1) group.
- + * (Section 5.2)
- */
- -static bool vext_check_overlap_mask(DisasContext *s, uint32_t vd, bool vm,
- - bool force)
- +static bool vext_check_dds(DisasContext *s, int vd, int vs1, int vs2, int vm)
- {
- - return (vm != 0 || vd != 0) || (!force && (s->lmul == 0));
- + return vext_check_ds(s, vd, vs1, vm) &&
- + require_align(vs2, s->lmul + 1);
- }
-
- -/* The LMUL setting must be such that LMUL * NFIELDS <= 8. (Section 7.8) */
- -static bool vext_check_nf(DisasContext *s, uint32_t nf)
- +static bool vext_check_sd(DisasContext *s, int vd, int vs, int vm)
- {
- - return (1 << s->lmul) * nf <= 8;
- + bool ret = vext_narrow_check_common(s, vd, vs, vm);
- + if (vd != vs) {
- + ret &= require_noover(vd, s->lmul, vs, s->lmul + 1);
- + }
- + return ret;
- }
-
- /*
- - * The destination vector register group cannot overlap a source vector register
- - * group of a different element width. (Section 11.2)
- + * Check function for vector instruction with format:
- + * single-width result and double-width source 1 and single-width
- + * source 2 (SEW = 2*SEW op SEW)
- + *
- + * Rules to be checked here:
- + * 1. All rules in defined in narrow common rules are applied.
- + * 2. Destination vector register cannot overlap a source vector
- + * register (vs2) group.
- + * (Section 5.2)
- + * 3. Source 2 (vs1) vector register number is multiples of LMUL.
- + * (Section 3.3.2)
- */
- -static inline bool vext_check_overlap_group(int rd, int dlen, int rs, int slen)
- +static bool vext_check_sds(DisasContext *s, int vd, int vs1, int vs2, int vm)
- {
- - return ((rd >= rs + slen) || (rs >= rd + dlen));
- + return vext_check_sd(s, vd, vs2, vm) &&
- + require_align(vs1, s->lmul);
- }
- +
- +/*
- + * Check function for vector reduction instructions.
- + *
- + * Rules to be checked here:
- + * 1. Source 1 (vs2) vector register number is multiples of LMUL.
- + * (Section 3.3.2)
- + */
- +static bool vext_check_reduction(DisasContext *s, int vs2)
- +{
- + return require_align(vs2, s->lmul) && (s->vstart == 0);
- +}
- +
- +/*
- + * Check function for vector slide instructions.
- + *
- + * Rules to be checked here:
- + * 1. Source 1 (vs2) vector register number is multiples of LMUL.
- + * (Section 3.3.2)
- + * 2. Destination vector register number is multiples of LMUL.
- + * (Section 3.3.2)
- + * 3. Destination vector register group for a masked vector
- + * instruction cannot overlap the source mask register (v0).
- + * (Section 5.3)
- + * 4. The destination vector register group for vslideup, vslide1up,
- + * vfslide1up, cannot overlap the source vector register (vs2) group.
- + * (Section 5.2, 17.3.1, 17.3.3)
- + */
- +static bool vext_check_slide(DisasContext *s, int vd, int vs2,
- + int vm, bool is_over)
- +{
- + bool ret = require_align(vs2, s->lmul) &&
- + require_align(vd, s->lmul) &&
- + require_vm(vm, vd);
- + if (is_over) {
- + ret &= (vd != vs2);
- + }
- + return ret;
- +}
- +
- +/*
- + * In cpu_get_tb_cpu_state(), set VILL if RVV was not present.
- + * So RVV is also be checked in this function.
- + */
- +static bool vext_check_isa_ill(DisasContext *s)
- +{
- + return !s->vill;
- +}
- +
- /* common translation macro */
- #define GEN_VEXT_TRANS(NAME, SEQ, ARGTYPE, OP, CHECK) \
- static bool trans_##NAME(DisasContext *s, arg_##ARGTYPE *a)\
- @@ -803,11 +1108,9 @@ GEN_VEXT_TRANS(vamomaxud_v, 17, rwdvm, amo_op, amo_check)
-
- static bool opivv_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_sss(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- typedef void GVecGen3Fn(unsigned, uint32_t, uint32_t,
- @@ -898,10 +1201,9 @@ static bool opivx_trans(uint32_t vd, uint32_t rs1, uint32_t vs2, uint32_t vm,
-
- static bool opivx_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ss(s, a->rd, a->rs2, a->vm);
- }
-
- typedef void GVecGen2sFn(unsigned, uint32_t, uint32_t, TCGv_i64,
- @@ -1098,16 +1400,9 @@ GEN_OPIVI_GVEC_TRANS(vrsub_vi, 0, vrsub_vx, rsubi)
- /* OPIVV with WIDEN */
- static bool opivv_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs2,
- - 1 << s->lmul) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs1,
- - 1 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_dss(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- static bool do_opivv_widen(DisasContext *s, arg_rmrr *a,
- @@ -1152,13 +1447,9 @@ GEN_OPIVV_WIDEN_TRANS(vwsub_vv, opivv_widen_check)
- /* OPIVX with WIDEN */
- static bool opivx_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs2,
- - 1 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ds(s, a->rd, a->rs2, a->vm);
- }
-
- static bool do_opivx_widen(DisasContext *s, arg_rmrr *a,
- @@ -1189,14 +1480,9 @@ GEN_OPIVX_WIDEN_TRANS(vwsub_vx)
- /* WIDEN OPIVV with WIDEN */
- static bool opiwv_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, true) &&
- - vext_check_reg(s, a->rs1, false) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs1,
- - 1 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_dds(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- static bool do_opiwv_widen(DisasContext *s, arg_rmrr *a,
- @@ -1239,11 +1525,9 @@ GEN_OPIWV_WIDEN_TRANS(vwsub_wv)
- /* WIDEN OPIVX with WIDEN */
- static bool opiwx_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, true) &&
- - (s->lmul < 0x3) && (s->sew < 0x3));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_dd(s, a->rd, a->rs2, a->vm);
- }
-
- static bool do_opiwx_widen(DisasContext *s, arg_rmrr *a,
- @@ -1304,11 +1588,10 @@ static bool trans_##NAME(DisasContext *s, arg_rmrr *a) \
- */
- static bool opivv_vadc_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - ((a->rd != 0) || (s->lmul == 0)));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + (a->rd != 0) &&
- + vext_check_sss(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- GEN_OPIVV_TRANS(vadc_vvm, opivv_vadc_check)
- @@ -1320,11 +1603,9 @@ GEN_OPIVV_TRANS(vsbc_vvm, opivv_vadc_check)
- */
- static bool opivv_vmadc_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - vext_check_overlap_group(a->rd, 1, a->rs1, 1 << s->lmul) &&
- - vext_check_overlap_group(a->rd, 1, a->rs2, 1 << s->lmul));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_mss(s, a->rd, a->rs1, a->rs2);
- }
-
- GEN_OPIVV_TRANS(vmadc_vvm, opivv_vmadc_check)
- @@ -1332,10 +1613,10 @@ GEN_OPIVV_TRANS(vmsbc_vvm, opivv_vmadc_check)
-
- static bool opivx_vadc_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - ((a->rd != 0) || (s->lmul == 0)));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + (a->rd != 0) &&
- + vext_check_ss(s, a->rd, a->rs2, a->vm);
- }
-
- /* OPIVX without GVEC IR */
- @@ -1358,9 +1639,9 @@ GEN_OPIVX_TRANS(vsbc_vxm, opivx_vadc_check)
-
- static bool opivx_vmadc_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_overlap_group(a->rd, 1, a->rs2, 1 << s->lmul));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ms(s, a->rd, a->rs2);
- }
-
- GEN_OPIVX_TRANS(vmadc_vxm, opivx_vmadc_check)
- @@ -1451,14 +1732,9 @@ GEN_OPIVI_GVEC_TRANS(vsra_vi, 1, vsra_vx, sari)
- /* Vector Narrowing Integer Right Shift Instructions */
- static bool opivv_narrow_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, true) &&
- - vext_check_reg(s, a->rs1, false) &&
- - vext_check_overlap_group(a->rd, 1 << s->lmul, a->rs2,
- - 2 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_sds(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- /* OPIVV with NARROW */
- @@ -1492,13 +1768,9 @@ GEN_OPIVV_NARROW_TRANS(vnsrl_vv)
-
- static bool opivx_narrow_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, true) &&
- - vext_check_overlap_group(a->rd, 1 << s->lmul, a->rs2,
- - 2 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_sd(s, a->rd, a->rs2, a->vm);
- }
-
- /* OPIVX with NARROW */
- @@ -1546,13 +1818,11 @@ GEN_OPIVI_NARROW_TRANS(vnsrl_vi, 1, vnsrl_vx)
- */
- static bool opivv_cmp_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - ((vext_check_overlap_group(a->rd, 1, a->rs1, 1 << s->lmul) &&
- - vext_check_overlap_group(a->rd, 1, a->rs2, 1 << s->lmul)) ||
- - (s->lmul == 0)));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_mss(s, a->rd, a->rs1, a->rs2);
- }
- +
- GEN_OPIVV_TRANS(vmseq_vv, opivv_cmp_check)
- GEN_OPIVV_TRANS(vmsne_vv, opivv_cmp_check)
- GEN_OPIVV_TRANS(vmsltu_vv, opivv_cmp_check)
- @@ -1562,10 +1832,9 @@ GEN_OPIVV_TRANS(vmsle_vv, opivv_cmp_check)
-
- static bool opivx_cmp_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rs2, false) &&
- - (vext_check_overlap_group(a->rd, 1, a->rs2, 1 << s->lmul) ||
- - (s->lmul == 0)));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ms(s, a->rd, a->rs2);
- }
-
- GEN_OPIVX_TRANS(vmseq_vx, opivx_cmp_check)
- @@ -1644,10 +1913,10 @@ GEN_OPIVX_WIDEN_TRANS(vwmaccus_vx)
- /* Vector Integer Merge and Move Instructions */
- static bool trans_vmv_v_v(DisasContext *s, arg_vmv_v_v *a)
- {
- - if (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs1, false)) {
- -
- + if (require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + /* vmv.v.v has rs2 = 0 and vm = 1 */
- + vext_check_sss(s, a->rd, a->rs1, 0, 1)) {
- if (s->vl_eq_vlmax) {
- tcg_gen_gvec_mov(s->sew, vreg_ofs(s, a->rd),
- vreg_ofs(s, a->rs1),
- @@ -1674,9 +1943,10 @@ static bool trans_vmv_v_v(DisasContext *s, arg_vmv_v_v *a)
- typedef void gen_helper_vmv_vx(TCGv_ptr, TCGv_i64, TCGv_env, TCGv_i32);
- static bool trans_vmv_v_x(DisasContext *s, arg_vmv_v_x *a)
- {
- - if (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false)) {
- -
- + if (require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + /* vmv.v.x has rs2 = 0 and vm = 1 */
- + vext_check_ss(s, a->rd, 0, 1)) {
- TCGv s1;
- TCGLabel *over = gen_new_label();
- tcg_gen_brcondi_tl(TCG_COND_EQ, cpu_vl, 0, over);
- @@ -1717,9 +1987,10 @@ static bool trans_vmv_v_x(DisasContext *s, arg_vmv_v_x *a)
-
- static bool trans_vmv_v_i(DisasContext *s, arg_vmv_v_i *a)
- {
- - if (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false)) {
- -
- + if (require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + /* vmv.v.i has rs2 = 0 and vm = 1 */
- + vext_check_ss(s, a->rd, 0, 1)) {
- int64_t simm = sextract64(a->rs1, 0, 5);
- if (s->vl_eq_vlmax) {
- tcg_gen_gvec_dup_imm(s->sew, vreg_ofs(s, a->rd),
- @@ -1821,12 +2092,10 @@ GEN_OPIVI_NARROW_TRANS(vnclip_vi, 1, vnclip_vx)
- */
- static bool opfvv_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - (s->sew != 0));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_sss(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- /* OPFVV without GVEC IR */
- @@ -1891,17 +2160,16 @@ static bool opfvf_trans(uint32_t vd, uint32_t rs1, uint32_t vs2,
- return true;
- }
-
- -static bool opfvf_check(DisasContext *s, arg_rmrr *a)
- -{
- /*
- * If the current SEW does not correspond to a supported IEEE floating-point
- * type, an illegal instruction exception is raised
- */
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - (s->sew != 0));
- +static bool opfvf_check(DisasContext *s, arg_rmrr *a)
- +{
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ss(s, a->rd, a->rs2, a->vm);
- }
-
- /* OPFVF without GVEC IR */
- @@ -1931,16 +2199,10 @@ GEN_OPFVF_TRANS(vfrsub_vf, opfvf_check)
- /* Vector Widening Floating-Point Add/Subtract Instructions */
- static bool opfvv_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs2,
- - 1 << s->lmul) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs1,
- - 1 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3) && (s->sew != 0));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_dss(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- /* OPFVV with WIDEN */
- @@ -1974,13 +2236,10 @@ GEN_OPFVV_WIDEN_TRANS(vfwsub_vv, opfvv_widen_check)
-
- static bool opfvf_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs2,
- - 1 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3) && (s->sew != 0));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ds(s, a->rd, a->rs2, a->vm);
- }
-
- /* OPFVF with WIDEN */
- @@ -2006,14 +2265,10 @@ GEN_OPFVF_WIDEN_TRANS(vfwsub_vf)
-
- static bool opfwv_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, true) &&
- - vext_check_reg(s, a->rs1, false) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs1,
- - 1 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3) && (s->sew != 0));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_dds(s, a->rd, a->rs1, a->rs2, a->vm);
- }
-
- /* WIDEN OPFVV with WIDEN */
- @@ -2047,11 +2302,10 @@ GEN_OPFWV_WIDEN_TRANS(vfwsub_wv)
-
- static bool opfwf_widen_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, true) &&
- - (s->lmul < 0x3) && (s->sew < 0x3) && (s->sew != 0));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_dd(s, a->rd, a->rs2, a->vm);
- }
-
- /* WIDEN OPFVF with WIDEN */
- @@ -2122,11 +2376,11 @@ GEN_OPFVF_WIDEN_TRANS(vfwnmsac_vf)
- */
- static bool opfv_check(DisasContext *s, arg_rmr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - (s->sew != 0));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + /* OPFV instructions ignore vs1 check */
- + vext_check_ss(s, a->rd, a->rs2, a->vm);
- }
-
- #define GEN_OPFV_TRANS(NAME, CHECK) \
- @@ -2174,13 +2428,10 @@ GEN_OPFVF_TRANS(vfsgnjx_vf, opfvf_check)
- /* Vector Floating-Point Compare Instructions */
- static bool opfvv_cmp_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - (s->sew != 0) &&
- - ((vext_check_overlap_group(a->rd, 1, a->rs1, 1 << s->lmul) &&
- - vext_check_overlap_group(a->rd, 1, a->rs2, 1 << s->lmul)) ||
- - (s->lmul == 0)));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_mss(s, a->rd, a->rs1, a->rs2);
- }
-
- GEN_OPFVV_TRANS(vmfeq_vv, opfvv_cmp_check)
- @@ -2191,11 +2442,10 @@ GEN_OPFVV_TRANS(vmford_vv, opfvv_cmp_check)
-
- static bool opfvf_cmp_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rs2, false) &&
- - (s->sew != 0) &&
- - (vext_check_overlap_group(a->rd, 1, a->rs2, 1 << s->lmul) ||
- - (s->lmul == 0)));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ms(s, a->rd, a->rs2);
- }
-
- GEN_OPFVF_TRANS(vmfeq_vf, opfvf_cmp_check)
- @@ -2214,10 +2464,10 @@ GEN_OPFVF_TRANS(vfmerge_vfm, opfvf_check)
-
- static bool trans_vfmv_v_f(DisasContext *s, arg_vfmv_v_f *a)
- {
- - if (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false) &&
- - (s->sew != 0)) {
- -
- + if (require_rvv(s) &&
- + require_rvf(s) &&
- + vext_check_isa_ill(s) &&
- + require_align(a->rd, s->lmul)) {
- if (s->vl_eq_vlmax) {
- tcg_gen_gvec_dup_i64(s->sew, vreg_ofs(s, a->rd),
- MAXSZ(s), MAXSZ(s), cpu_fpr[a->rs1]);
- @@ -2263,13 +2513,11 @@ GEN_OPFV_TRANS(vfcvt_f_x_v, opfv_check)
- */
- static bool opfv_widen_check(DisasContext *s, arg_rmr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, true) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_overlap_group(a->rd, 2 << s->lmul, a->rs2,
- - 1 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3) && (s->sew != 0));
- + return require_rvv(s) &&
- + require_scale_rvf(s) &&
- + (s->sew != MO_8) &&
- + vext_check_isa_ill(s) &&
- + vext_check_ds(s, a->rd, a->rs2, a->vm);
- }
-
- #define GEN_OPFV_WIDEN_TRANS(NAME) \
- @@ -2311,13 +2559,12 @@ GEN_OPFV_WIDEN_TRANS(vfwcvt_f_f_v)
- */
- static bool opfv_narrow_check(DisasContext *s, arg_rmr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, true) &&
- - vext_check_overlap_group(a->rd, 1 << s->lmul, a->rs2,
- - 2 << s->lmul) &&
- - (s->lmul < 0x3) && (s->sew < 0x3) && (s->sew != 0));
- + return require_rvv(s) &&
- + require_rvf(s) &&
- + (s->sew != MO_64) &&
- + vext_check_isa_ill(s) &&
- + /* OPFV narrowing instructions ignore vs1 check */
- + vext_check_sd(s, a->rd, a->rs2, a->vm);
- }
-
- #define GEN_OPFV_NARROW_TRANS(NAME) \
- @@ -2357,7 +2604,9 @@ GEN_OPFV_NARROW_TRANS(vfncvt_f_f_v)
- /* Vector Single-Width Integer Reduction Instructions */
- static bool reduction_check(DisasContext *s, arg_rmrr *a)
- {
- - return vext_check_isa_ill(s) && vext_check_reg(s, a->rs2, false);
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_reduction(s, a->rs2);
- }
-
- GEN_OPIVV_TRANS(vredsum_vs, reduction_check)
- @@ -2370,8 +2619,13 @@ GEN_OPIVV_TRANS(vredor_vs, reduction_check)
- GEN_OPIVV_TRANS(vredxor_vs, reduction_check)
-
- /* Vector Widening Integer Reduction Instructions */
- -GEN_OPIVV_WIDEN_TRANS(vwredsum_vs, reduction_check)
- -GEN_OPIVV_WIDEN_TRANS(vwredsumu_vs, reduction_check)
- +static bool reduction_widen_check(DisasContext *s, arg_rmrr *a)
- +{
- + return reduction_check(s, a) && (s->sew < MO_64);
- +}
- +
- +GEN_OPIVV_WIDEN_TRANS(vwredsum_vs, reduction_widen_check)
- +GEN_OPIVV_WIDEN_TRANS(vwredsumu_vs, reduction_widen_check)
-
- /* Vector Single-Width Floating-Point Reduction Instructions */
- GEN_OPFVV_TRANS(vfredsum_vs, reduction_check)
- @@ -2419,7 +2673,8 @@ GEN_MM_TRANS(vmxnor_mm)
- /* Vector mask population count vmpopc */
- static bool trans_vmpopc_m(DisasContext *s, arg_rmr *a)
- {
- - if (vext_check_isa_ill(s)) {
- + if (require_rvv(s) &&
- + vext_check_isa_ill(s)) {
- TCGv_ptr src2, mask;
- TCGv dst;
- TCGv_i32 desc;
- @@ -2450,7 +2705,8 @@ static bool trans_vmpopc_m(DisasContext *s, arg_rmr *a)
- /* vmfirst find-first-set mask bit */
- static bool trans_vmfirst_m(DisasContext *s, arg_rmr *a)
- {
- - if (vext_check_isa_ill(s)) {
- + if (require_rvv(s) &&
- + vext_check_isa_ill(s)) {
- TCGv_ptr src2, mask;
- TCGv dst;
- TCGv_i32 desc;
- @@ -2509,10 +2765,11 @@ GEN_M_TRANS(vmsof_m)
- /* Vector Iota Instruction */
- static bool trans_viota_m(DisasContext *s, arg_viota_m *a)
- {
- - if (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_overlap_group(a->rd, 1 << s->lmul, a->rs2, 1) &&
- - (a->vm != 0 || a->rd != 0)) {
- + if (require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + require_noover(a->rd, s->lmul, a->rs2, 0) &&
- + require_vm(a->vm, a->rd) &&
- + require_align(a->rd, s->lmul)) {
- uint32_t data = 0;
- TCGLabel *over = gen_new_label();
- tcg_gen_brcondi_tl(TCG_COND_EQ, cpu_vl, 0, over);
- @@ -2536,9 +2793,10 @@ static bool trans_viota_m(DisasContext *s, arg_viota_m *a)
- /* Vector Element Index Instruction */
- static bool trans_vid_v(DisasContext *s, arg_vid_v *a)
- {
- - if (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, false)) {
- + if (require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + require_align(a->rd, s->lmul) &&
- + require_vm(a->vm, a->rd)) {
- uint32_t data = 0;
- TCGLabel *over = gen_new_label();
- tcg_gen_brcondi_tl(TCG_COND_EQ, cpu_vl, 0, over);
- @@ -2788,41 +3046,48 @@ static bool trans_vfmv_s_f(DisasContext *s, arg_vfmv_s_f *a)
- /* Vector Slide Instructions */
- static bool slideup_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - (a->rd != a->rs2));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_slide(s, a->rd, a->rs2, a->vm, true);
- }
-
- GEN_OPIVX_TRANS(vslideup_vx, slideup_check)
- GEN_OPIVX_TRANS(vslide1up_vx, slideup_check)
- GEN_OPIVI_TRANS(vslideup_vi, 1, vslideup_vx, slideup_check)
-
- -GEN_OPIVX_TRANS(vslidedown_vx, opivx_check)
- -GEN_OPIVX_TRANS(vslide1down_vx, opivx_check)
- -GEN_OPIVI_TRANS(vslidedown_vi, 1, vslidedown_vx, opivx_check)
- +static bool slidedown_check(DisasContext *s, arg_rmrr *a)
- +{
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + vext_check_slide(s, a->rd, a->rs2, a->vm, false);
- +}
- +
- +GEN_OPIVX_TRANS(vslidedown_vx, slidedown_check)
- +GEN_OPIVX_TRANS(vslide1down_vx, slidedown_check)
- +GEN_OPIVI_TRANS(vslidedown_vi, 1, vslidedown_vx, slidedown_check)
-
- /* Vector Register Gather Instruction */
- static bool vrgather_vv_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs1, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - (a->rd != a->rs2) && (a->rd != a->rs1));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + require_align(a->rd, s->lmul) &&
- + require_align(a->rs1, s->lmul) &&
- + require_align(a->rs2, s->lmul) &&
- + (a->rd != a->rs2 && a->rd != a->rs1) &&
- + require_vm(a->vm, a->rd);
- }
-
- GEN_OPIVV_TRANS(vrgather_vv, vrgather_vv_check)
-
- static bool vrgather_vx_check(DisasContext *s, arg_rmrr *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_overlap_mask(s, a->rd, a->vm, true) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - (a->rd != a->rs2));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + require_align(a->rd, s->lmul) &&
- + require_align(a->rs2, s->lmul) &&
- + (a->rd != a->rs2) &&
- + require_vm(a->vm, a->rd);
- }
-
- /* vrgather.vx vd, vs2, rs1, vm # vd[i] = (x[rs1] >= VLMAX) ? 0 : vs2[rs1] */
- @@ -2883,14 +3148,20 @@ static bool trans_vrgather_vi(DisasContext *s, arg_rmrr *a)
- return true;
- }
-
- -/* Vector Compress Instruction */
- +/*
- + * Vector Compress Instruction
- + *
- + * The destination vector register group cannot overlap the
- + * source vector register group or the source mask register.
- + */
- static bool vcompress_vm_check(DisasContext *s, arg_r *a)
- {
- - return (vext_check_isa_ill(s) &&
- - vext_check_reg(s, a->rd, false) &&
- - vext_check_reg(s, a->rs2, false) &&
- - vext_check_overlap_group(a->rd, 1 << s->lmul, a->rs1, 1) &&
- - (a->rd != a->rs2));
- + return require_rvv(s) &&
- + vext_check_isa_ill(s) &&
- + require_align(a->rd, s->lmul) &&
- + require_align(a->rs2, s->lmul) &&
- + (a->rd != a->rs2) &&
- + !is_overlapped(a->rd, 1 << MAX(s->lmul, 0), a->rs1, 1) &&
- }
-
- static bool trans_vcompress_vm(DisasContext *s, arg_r *a)
- --
- 2.33.1
|