target/riscv: rvv-1.0: update vext_max_elems() for load/store insns
Signed-off-by: Frank Chang <frank.chang@sifive.com> Reviewed-by: Alistair Francis <alistair.francis@wdc.com> Message-Id: <20211210075704.23951-26-frank.chang@sifive.com> Signed-off-by: Alistair Francis <alistair.francis@wdc.com>
This commit is contained in:
parent
30206bd842
commit
5a9f8e1552
@ -540,6 +540,12 @@ static bool trans_##NAME(DisasContext *s, arg_##ARGTYPE * a) \
|
|||||||
return false; \
|
return false; \
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static uint8_t vext_get_emul(DisasContext *s, uint8_t eew)
|
||||||
|
{
|
||||||
|
int8_t emul = eew - s->sew + s->lmul;
|
||||||
|
return emul < 0 ? 0 : emul;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
*** unit stride load and store
|
*** unit stride load and store
|
||||||
*/
|
*/
|
||||||
@ -604,8 +610,14 @@ static bool ld_us_op(DisasContext *s, arg_r2nfvm *a, uint8_t eew)
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Vector load/store instructions have the EEW encoded
|
||||||
|
* directly in the instructions. The maximum vector size is
|
||||||
|
* calculated with EMUL rather than LMUL.
|
||||||
|
*/
|
||||||
|
uint8_t emul = vext_get_emul(s, eew);
|
||||||
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
||||||
data = FIELD_DP32(data, VDATA, LMUL, s->lmul);
|
data = FIELD_DP32(data, VDATA, LMUL, emul);
|
||||||
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
||||||
return ldst_us_trans(a->rd, a->rs1, data, fn, s, false);
|
return ldst_us_trans(a->rd, a->rs1, data, fn, s, false);
|
||||||
}
|
}
|
||||||
@ -640,8 +652,9 @@ static bool st_us_op(DisasContext *s, arg_r2nfvm *a, uint8_t eew)
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
uint8_t emul = vext_get_emul(s, eew);
|
||||||
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
||||||
data = FIELD_DP32(data, VDATA, LMUL, s->lmul);
|
data = FIELD_DP32(data, VDATA, LMUL, emul);
|
||||||
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
||||||
return ldst_us_trans(a->rd, a->rs1, data, fn, s, true);
|
return ldst_us_trans(a->rd, a->rs1, data, fn, s, true);
|
||||||
}
|
}
|
||||||
@ -711,8 +724,9 @@ static bool ld_stride_op(DisasContext *s, arg_rnfvm *a, uint8_t eew)
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
uint8_t emul = vext_get_emul(s, eew);
|
||||||
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
||||||
data = FIELD_DP32(data, VDATA, LMUL, s->lmul);
|
data = FIELD_DP32(data, VDATA, LMUL, emul);
|
||||||
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
||||||
return ldst_stride_trans(a->rd, a->rs1, a->rs2, data, fn, s, false);
|
return ldst_stride_trans(a->rd, a->rs1, a->rs2, data, fn, s, false);
|
||||||
}
|
}
|
||||||
@ -739,8 +753,9 @@ static bool st_stride_op(DisasContext *s, arg_rnfvm *a, uint8_t eew)
|
|||||||
gen_helper_vsse32_v, gen_helper_vsse64_v
|
gen_helper_vsse32_v, gen_helper_vsse64_v
|
||||||
};
|
};
|
||||||
|
|
||||||
|
uint8_t emul = vext_get_emul(s, eew);
|
||||||
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
||||||
data = FIELD_DP32(data, VDATA, LMUL, s->lmul);
|
data = FIELD_DP32(data, VDATA, LMUL, emul);
|
||||||
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
||||||
fn = fns[eew];
|
fn = fns[eew];
|
||||||
if (fn == NULL) {
|
if (fn == NULL) {
|
||||||
@ -836,8 +851,9 @@ static bool ld_index_op(DisasContext *s, arg_rnfvm *a, uint8_t eew)
|
|||||||
|
|
||||||
fn = fns[eew][s->sew];
|
fn = fns[eew][s->sew];
|
||||||
|
|
||||||
|
uint8_t emul = vext_get_emul(s, s->sew);
|
||||||
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
||||||
data = FIELD_DP32(data, VDATA, LMUL, s->lmul);
|
data = FIELD_DP32(data, VDATA, LMUL, emul);
|
||||||
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
||||||
return ldst_index_trans(a->rd, a->rs1, a->rs2, data, fn, s, false);
|
return ldst_index_trans(a->rd, a->rs1, a->rs2, data, fn, s, false);
|
||||||
}
|
}
|
||||||
@ -887,8 +903,9 @@ static bool st_index_op(DisasContext *s, arg_rnfvm *a, uint8_t eew)
|
|||||||
|
|
||||||
fn = fns[eew][s->sew];
|
fn = fns[eew][s->sew];
|
||||||
|
|
||||||
|
uint8_t emul = vext_get_emul(s, s->sew);
|
||||||
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
||||||
data = FIELD_DP32(data, VDATA, LMUL, s->lmul);
|
data = FIELD_DP32(data, VDATA, LMUL, emul);
|
||||||
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
||||||
return ldst_index_trans(a->rd, a->rs1, a->rs2, data, fn, s, true);
|
return ldst_index_trans(a->rd, a->rs1, a->rs2, data, fn, s, true);
|
||||||
}
|
}
|
||||||
@ -949,8 +966,9 @@ static bool ldff_op(DisasContext *s, arg_r2nfvm *a, uint8_t eew)
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
uint8_t emul = vext_get_emul(s, eew);
|
||||||
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
data = FIELD_DP32(data, VDATA, VM, a->vm);
|
||||||
data = FIELD_DP32(data, VDATA, LMUL, s->lmul);
|
data = FIELD_DP32(data, VDATA, LMUL, emul);
|
||||||
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
data = FIELD_DP32(data, VDATA, NF, a->nf);
|
||||||
return ldff_trans(a->rd, a->rs1, data, fn, s);
|
return ldff_trans(a->rd, a->rs1, data, fn, s);
|
||||||
}
|
}
|
||||||
|
@ -17,6 +17,7 @@
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
#include "qemu/osdep.h"
|
#include "qemu/osdep.h"
|
||||||
|
#include "qemu/host-utils.h"
|
||||||
#include "cpu.h"
|
#include "cpu.h"
|
||||||
#include "exec/memop.h"
|
#include "exec/memop.h"
|
||||||
#include "exec/exec-all.h"
|
#include "exec/exec-all.h"
|
||||||
@ -116,14 +117,21 @@ static inline int32_t vext_lmul(uint32_t desc)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Get vector group length in bytes. Its range is [64, 2048].
|
* Get the maximum number of elements can be operated.
|
||||||
*
|
*
|
||||||
* As simd_desc support at most 256, the max vlen is 512 bits.
|
* esz: log2 of element size in bytes.
|
||||||
* So vlen in bytes is encoded as maxsz.
|
|
||||||
*/
|
*/
|
||||||
static inline uint32_t vext_maxsz(uint32_t desc)
|
static inline uint32_t vext_max_elems(uint32_t desc, uint32_t esz)
|
||||||
{
|
{
|
||||||
return simd_maxsz(desc) << vext_lmul(desc);
|
/*
|
||||||
|
* As simd_desc support at most 256 bytes, the max vlen is 256 bits.
|
||||||
|
* so vlen in bytes (vlenb) is encoded as maxsz.
|
||||||
|
*/
|
||||||
|
uint32_t vlenb = simd_maxsz(desc);
|
||||||
|
|
||||||
|
/* Return VLMAX */
|
||||||
|
int scale = vext_lmul(desc) - esz;
|
||||||
|
return scale < 0 ? vlenb >> -scale : vlenb << scale;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -216,14 +224,14 @@ vext_ldst_stride(void *vd, void *v0, target_ulong base,
|
|||||||
{
|
{
|
||||||
uint32_t i, k;
|
uint32_t i, k;
|
||||||
uint32_t nf = vext_nf(desc);
|
uint32_t nf = vext_nf(desc);
|
||||||
uint32_t vlmax = vext_maxsz(desc) / esz;
|
uint32_t max_elems = vext_max_elems(desc, esz);
|
||||||
|
|
||||||
/* probe every access*/
|
/* probe every access*/
|
||||||
for (i = 0; i < env->vl; i++) {
|
for (i = 0; i < env->vl; i++) {
|
||||||
if (!vm && !vext_elem_mask(v0, i)) {
|
if (!vm && !vext_elem_mask(v0, i)) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
probe_pages(env, base + stride * i, nf * esz, ra, access_type);
|
probe_pages(env, base + stride * i, nf << esz, ra, access_type);
|
||||||
}
|
}
|
||||||
/* do real access */
|
/* do real access */
|
||||||
for (i = 0; i < env->vl; i++) {
|
for (i = 0; i < env->vl; i++) {
|
||||||
@ -232,8 +240,8 @@ vext_ldst_stride(void *vd, void *v0, target_ulong base,
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
while (k < nf) {
|
while (k < nf) {
|
||||||
target_ulong addr = base + stride * i + k * esz;
|
target_ulong addr = base + stride * i + (k << esz);
|
||||||
ldst_elem(env, addr, i + k * vlmax, vd, ra);
|
ldst_elem(env, addr, i + k * max_elems, vd, ra);
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -246,7 +254,7 @@ void HELPER(NAME)(void *vd, void * v0, target_ulong base, \
|
|||||||
{ \
|
{ \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
vext_ldst_stride(vd, v0, base, stride, env, desc, vm, LOAD_FN, \
|
vext_ldst_stride(vd, v0, base, stride, env, desc, vm, LOAD_FN, \
|
||||||
sizeof(ETYPE), GETPC(), MMU_DATA_LOAD); \
|
ctzl(sizeof(ETYPE)), GETPC(), MMU_DATA_LOAD); \
|
||||||
}
|
}
|
||||||
|
|
||||||
GEN_VEXT_LD_STRIDE(vlse8_v, int8_t, lde_b)
|
GEN_VEXT_LD_STRIDE(vlse8_v, int8_t, lde_b)
|
||||||
@ -261,7 +269,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
|||||||
{ \
|
{ \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
vext_ldst_stride(vd, v0, base, stride, env, desc, vm, STORE_FN, \
|
vext_ldst_stride(vd, v0, base, stride, env, desc, vm, STORE_FN, \
|
||||||
sizeof(ETYPE), GETPC(), MMU_DATA_STORE); \
|
ctzl(sizeof(ETYPE)), GETPC(), MMU_DATA_STORE); \
|
||||||
}
|
}
|
||||||
|
|
||||||
GEN_VEXT_ST_STRIDE(vsse8_v, int8_t, ste_b)
|
GEN_VEXT_ST_STRIDE(vsse8_v, int8_t, ste_b)
|
||||||
@ -281,16 +289,16 @@ vext_ldst_us(void *vd, target_ulong base, CPURISCVState *env, uint32_t desc,
|
|||||||
{
|
{
|
||||||
uint32_t i, k;
|
uint32_t i, k;
|
||||||
uint32_t nf = vext_nf(desc);
|
uint32_t nf = vext_nf(desc);
|
||||||
uint32_t vlmax = vext_maxsz(desc) / esz;
|
uint32_t max_elems = vext_max_elems(desc, esz);
|
||||||
|
|
||||||
/* probe every access */
|
/* probe every access */
|
||||||
probe_pages(env, base, env->vl * nf * esz, ra, access_type);
|
probe_pages(env, base, env->vl * (nf << esz), ra, access_type);
|
||||||
/* load bytes from guest memory */
|
/* load bytes from guest memory */
|
||||||
for (i = 0; i < env->vl; i++) {
|
for (i = 0; i < env->vl; i++) {
|
||||||
k = 0;
|
k = 0;
|
||||||
while (k < nf) {
|
while (k < nf) {
|
||||||
target_ulong addr = base + (i * nf + k) * esz;
|
target_ulong addr = base + ((i * nf + k) << esz);
|
||||||
ldst_elem(env, addr, i + k * vlmax, vd, ra);
|
ldst_elem(env, addr, i + k * max_elems, vd, ra);
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -305,16 +313,16 @@ vext_ldst_us(void *vd, target_ulong base, CPURISCVState *env, uint32_t desc,
|
|||||||
void HELPER(NAME##_mask)(void *vd, void *v0, target_ulong base, \
|
void HELPER(NAME##_mask)(void *vd, void *v0, target_ulong base, \
|
||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
uint32_t stride = vext_nf(desc) * sizeof(ETYPE); \
|
uint32_t stride = vext_nf(desc) << ctzl(sizeof(ETYPE)); \
|
||||||
vext_ldst_stride(vd, v0, base, stride, env, desc, false, LOAD_FN, \
|
vext_ldst_stride(vd, v0, base, stride, env, desc, false, LOAD_FN, \
|
||||||
sizeof(ETYPE), GETPC(), MMU_DATA_LOAD); \
|
ctzl(sizeof(ETYPE)), GETPC(), MMU_DATA_LOAD); \
|
||||||
} \
|
} \
|
||||||
\
|
\
|
||||||
void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
vext_ldst_us(vd, base, env, desc, LOAD_FN, \
|
vext_ldst_us(vd, base, env, desc, LOAD_FN, \
|
||||||
sizeof(ETYPE), GETPC(), MMU_DATA_LOAD); \
|
ctzl(sizeof(ETYPE)), GETPC(), MMU_DATA_LOAD); \
|
||||||
}
|
}
|
||||||
|
|
||||||
GEN_VEXT_LD_US(vle8_v, int8_t, lde_b)
|
GEN_VEXT_LD_US(vle8_v, int8_t, lde_b)
|
||||||
@ -326,16 +334,16 @@ GEN_VEXT_LD_US(vle64_v, int64_t, lde_d)
|
|||||||
void HELPER(NAME##_mask)(void *vd, void *v0, target_ulong base, \
|
void HELPER(NAME##_mask)(void *vd, void *v0, target_ulong base, \
|
||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
uint32_t stride = vext_nf(desc) * sizeof(ETYPE); \
|
uint32_t stride = vext_nf(desc) << ctzl(sizeof(ETYPE)); \
|
||||||
vext_ldst_stride(vd, v0, base, stride, env, desc, false, STORE_FN, \
|
vext_ldst_stride(vd, v0, base, stride, env, desc, false, STORE_FN, \
|
||||||
sizeof(ETYPE), GETPC(), MMU_DATA_STORE); \
|
ctzl(sizeof(ETYPE)), GETPC(), MMU_DATA_STORE); \
|
||||||
} \
|
} \
|
||||||
\
|
\
|
||||||
void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
vext_ldst_us(vd, base, env, desc, STORE_FN, \
|
vext_ldst_us(vd, base, env, desc, STORE_FN, \
|
||||||
sizeof(ETYPE), GETPC(), MMU_DATA_STORE); \
|
ctzl(sizeof(ETYPE)), GETPC(), MMU_DATA_STORE); \
|
||||||
}
|
}
|
||||||
|
|
||||||
GEN_VEXT_ST_US(vse8_v, int8_t, ste_b)
|
GEN_VEXT_ST_US(vse8_v, int8_t, ste_b)
|
||||||
@ -371,14 +379,14 @@ vext_ldst_index(void *vd, void *v0, target_ulong base,
|
|||||||
uint32_t i, k;
|
uint32_t i, k;
|
||||||
uint32_t nf = vext_nf(desc);
|
uint32_t nf = vext_nf(desc);
|
||||||
uint32_t vm = vext_vm(desc);
|
uint32_t vm = vext_vm(desc);
|
||||||
uint32_t vlmax = vext_maxsz(desc) / esz;
|
uint32_t max_elems = vext_max_elems(desc, esz);
|
||||||
|
|
||||||
/* probe every access*/
|
/* probe every access*/
|
||||||
for (i = 0; i < env->vl; i++) {
|
for (i = 0; i < env->vl; i++) {
|
||||||
if (!vm && !vext_elem_mask(v0, i)) {
|
if (!vm && !vext_elem_mask(v0, i)) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
probe_pages(env, get_index_addr(base, i, vs2), nf * esz, ra,
|
probe_pages(env, get_index_addr(base, i, vs2), nf << esz, ra,
|
||||||
access_type);
|
access_type);
|
||||||
}
|
}
|
||||||
/* load bytes from guest memory */
|
/* load bytes from guest memory */
|
||||||
@ -388,8 +396,8 @@ vext_ldst_index(void *vd, void *v0, target_ulong base,
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
while (k < nf) {
|
while (k < nf) {
|
||||||
abi_ptr addr = get_index_addr(base, i, vs2) + k * esz;
|
abi_ptr addr = get_index_addr(base, i, vs2) + (k << esz);
|
||||||
ldst_elem(env, addr, i + k * vlmax, vd, ra);
|
ldst_elem(env, addr, i + k * max_elems, vd, ra);
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -400,7 +408,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
|||||||
void *vs2, CPURISCVState *env, uint32_t desc) \
|
void *vs2, CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \
|
vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \
|
||||||
LOAD_FN, sizeof(ETYPE), GETPC(), MMU_DATA_LOAD); \
|
LOAD_FN, ctzl(sizeof(ETYPE)), GETPC(), MMU_DATA_LOAD); \
|
||||||
}
|
}
|
||||||
|
|
||||||
GEN_VEXT_LD_INDEX(vlxei8_8_v, int8_t, idx_b, lde_b)
|
GEN_VEXT_LD_INDEX(vlxei8_8_v, int8_t, idx_b, lde_b)
|
||||||
@ -425,7 +433,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
|||||||
void *vs2, CPURISCVState *env, uint32_t desc) \
|
void *vs2, CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \
|
vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \
|
||||||
STORE_FN, sizeof(ETYPE), \
|
STORE_FN, ctzl(sizeof(ETYPE)), \
|
||||||
GETPC(), MMU_DATA_STORE); \
|
GETPC(), MMU_DATA_STORE); \
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -459,7 +467,7 @@ vext_ldff(void *vd, void *v0, target_ulong base,
|
|||||||
uint32_t i, k, vl = 0;
|
uint32_t i, k, vl = 0;
|
||||||
uint32_t nf = vext_nf(desc);
|
uint32_t nf = vext_nf(desc);
|
||||||
uint32_t vm = vext_vm(desc);
|
uint32_t vm = vext_vm(desc);
|
||||||
uint32_t vlmax = vext_maxsz(desc) / esz;
|
uint32_t max_elems = vext_max_elems(desc, esz);
|
||||||
target_ulong addr, offset, remain;
|
target_ulong addr, offset, remain;
|
||||||
|
|
||||||
/* probe every access*/
|
/* probe every access*/
|
||||||
@ -467,24 +475,24 @@ vext_ldff(void *vd, void *v0, target_ulong base,
|
|||||||
if (!vm && !vext_elem_mask(v0, i)) {
|
if (!vm && !vext_elem_mask(v0, i)) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
addr = base + nf * i * esz;
|
addr = base + i * (nf << esz);
|
||||||
if (i == 0) {
|
if (i == 0) {
|
||||||
probe_pages(env, addr, nf * esz, ra, MMU_DATA_LOAD);
|
probe_pages(env, addr, nf << esz, ra, MMU_DATA_LOAD);
|
||||||
} else {
|
} else {
|
||||||
/* if it triggers an exception, no need to check watchpoint */
|
/* if it triggers an exception, no need to check watchpoint */
|
||||||
remain = nf * esz;
|
remain = nf << esz;
|
||||||
while (remain > 0) {
|
while (remain > 0) {
|
||||||
offset = -(addr | TARGET_PAGE_MASK);
|
offset = -(addr | TARGET_PAGE_MASK);
|
||||||
host = tlb_vaddr_to_host(env, addr, MMU_DATA_LOAD,
|
host = tlb_vaddr_to_host(env, addr, MMU_DATA_LOAD,
|
||||||
cpu_mmu_index(env, false));
|
cpu_mmu_index(env, false));
|
||||||
if (host) {
|
if (host) {
|
||||||
#ifdef CONFIG_USER_ONLY
|
#ifdef CONFIG_USER_ONLY
|
||||||
if (page_check_range(addr, nf * esz, PAGE_READ) < 0) {
|
if (page_check_range(addr, nf << esz, PAGE_READ) < 0) {
|
||||||
vl = i;
|
vl = i;
|
||||||
goto ProbeSuccess;
|
goto ProbeSuccess;
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
probe_pages(env, addr, nf * esz, ra, MMU_DATA_LOAD);
|
probe_pages(env, addr, nf << esz, ra, MMU_DATA_LOAD);
|
||||||
#endif
|
#endif
|
||||||
} else {
|
} else {
|
||||||
vl = i;
|
vl = i;
|
||||||
@ -509,8 +517,8 @@ ProbeSuccess:
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
while (k < nf) {
|
while (k < nf) {
|
||||||
target_ulong addr = base + (i * nf + k) * esz;
|
target_ulong addr = base + ((i * nf + k) << esz);
|
||||||
ldst_elem(env, addr, i + k * vlmax, vd, ra);
|
ldst_elem(env, addr, i + k * max_elems, vd, ra);
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -521,7 +529,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
|
|||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
vext_ldff(vd, v0, base, env, desc, LOAD_FN, \
|
vext_ldff(vd, v0, base, env, desc, LOAD_FN, \
|
||||||
sizeof(ETYPE), GETPC()); \
|
ctzl(sizeof(ETYPE)), GETPC()); \
|
||||||
}
|
}
|
||||||
|
|
||||||
GEN_VEXT_LDFF(vle8ff_v, int8_t, lde_b)
|
GEN_VEXT_LDFF(vle8ff_v, int8_t, lde_b)
|
||||||
@ -986,7 +994,8 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint32_t vlmax = vext_maxsz(desc) / sizeof(ETYPE); \
|
uint32_t vlmax = vext_max_elems(desc, \
|
||||||
|
ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t i; \
|
uint32_t i; \
|
||||||
\
|
\
|
||||||
for (i = 0; i < vl; i++) { \
|
for (i = 0; i < vl; i++) { \
|
||||||
@ -1016,7 +1025,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
|
|||||||
void *vs2, CPURISCVState *env, uint32_t desc) \
|
void *vs2, CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint32_t vlmax = vext_maxsz(desc) / sizeof(ETYPE); \
|
uint32_t vlmax = vext_max_elems(desc, ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t i; \
|
uint32_t i; \
|
||||||
\
|
\
|
||||||
for (i = 0; i < vl; i++) { \
|
for (i = 0; i < vl; i++) { \
|
||||||
@ -1190,7 +1199,8 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|||||||
{ \
|
{ \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint32_t vlmax = vext_maxsz(desc) / sizeof(ETYPE); \
|
uint32_t vlmax = vext_max_elems(desc, \
|
||||||
|
ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t i; \
|
uint32_t i; \
|
||||||
\
|
\
|
||||||
for (i = 0; i < vl; i++) { \
|
for (i = 0; i < vl; i++) { \
|
||||||
@ -1242,7 +1252,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
|||||||
{ \
|
{ \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint32_t vlmax = vext_maxsz(desc) / sizeof(ETYPE); \
|
uint32_t vlmax = vext_max_elems(desc, ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t i; \
|
uint32_t i; \
|
||||||
\
|
\
|
||||||
for (i = 0; i < vl; i++) { \
|
for (i = 0; i < vl; i++) { \
|
||||||
@ -3644,7 +3654,8 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|||||||
{ \
|
{ \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint32_t vlmax = vext_maxsz(desc) / sizeof(ETYPE); \
|
uint32_t vlmax = vext_max_elems(desc, \
|
||||||
|
ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t i; \
|
uint32_t i; \
|
||||||
\
|
\
|
||||||
for (i = 0; i < vl; i++) { \
|
for (i = 0; i < vl; i++) { \
|
||||||
@ -3671,7 +3682,7 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
|
|||||||
{ \
|
{ \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint32_t vlmax = vext_maxsz(desc) / sizeof(ETYPE); \
|
uint32_t vlmax = vext_max_elems(desc, ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t i; \
|
uint32_t i; \
|
||||||
\
|
\
|
||||||
for (i = 0; i < vl; i++) { \
|
for (i = 0; i < vl; i++) { \
|
||||||
@ -4457,7 +4468,7 @@ GEN_VEXT_VSLIDE1DOWN_VX(vslide1down_vx_d, uint64_t, H8)
|
|||||||
void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
uint32_t vlmax = env_archcpu(env)->cfg.vlen; \
|
uint32_t vlmax = vext_max_elems(desc, ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint64_t index; \
|
uint64_t index; \
|
||||||
@ -4486,7 +4497,7 @@ GEN_VEXT_VRGATHER_VV(vrgather_vv_d, uint64_t, H8)
|
|||||||
void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
||||||
CPURISCVState *env, uint32_t desc) \
|
CPURISCVState *env, uint32_t desc) \
|
||||||
{ \
|
{ \
|
||||||
uint32_t vlmax = env_archcpu(env)->cfg.vlen; \
|
uint32_t vlmax = vext_max_elems(desc, ctzl(sizeof(ETYPE))); \
|
||||||
uint32_t vm = vext_vm(desc); \
|
uint32_t vm = vext_vm(desc); \
|
||||||
uint32_t vl = env->vl; \
|
uint32_t vl = env->vl; \
|
||||||
uint64_t index = s1; \
|
uint64_t index = s1; \
|
||||||
|
Loading…
x
Reference in New Issue
Block a user