Lines Matching refs:v0

156 static inline void vext_set_elem_mask(void *v0, int index,  in vext_set_elem_mask()  argument
161 uint64_t old = ((uint64_t *)v0)[idx]; in vext_set_elem_mask()
162 ((uint64_t *)v0)[idx] = deposit64(old, pos, 1, value); in vext_set_elem_mask()
272 vext_ldst_stride(void *vd, void *v0, target_ulong base, target_ulong stride, in vext_ldst_stride() argument
288 if (!vm && !vext_elem_mask(v0, i)) { in vext_ldst_stride()
306 void HELPER(NAME)(void *vd, void * v0, target_ulong base, \
311 vext_ldst_stride(vd, v0, base, stride, env, desc, vm, LOAD_FN, \
321 void HELPER(NAME)(void *vd, void *v0, target_ulong base, \ in GEN_VEXT_LD_STRIDE()
326 vext_ldst_stride(vd, v0, base, stride, env, desc, vm, STORE_FN, \ in GEN_VEXT_LD_STRIDE()
470 void HELPER(NAME##_mask)(void *vd, void *v0, target_ulong base, \
474 vext_ldst_stride(vd, v0, base, stride, env, desc, false, \
478 void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
491 void HELPER(NAME##_mask)(void *vd, void *v0, target_ulong base, \ in GEN_VEXT_LD_US()
495 vext_ldst_stride(vd, v0, base, stride, env, desc, false, \ in GEN_VEXT_LD_US()
499 void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
514 void HELPER(vlm_v)(void *vd, void *v0, target_ulong base,
523 void HELPER(vsm_v)(void *vd, void *v0, target_ulong base, in HELPER()
551 vext_ldst_index(void *vd, void *v0, target_ulong base, in GEN_VEXT_GET_INDEX_ADDR()
570 if (!vm && !vext_elem_mask(v0, i)) { in GEN_VEXT_GET_INDEX_ADDR()
588 void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
591 vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \
613 void HELPER(NAME)(void *vd, void *v0, target_ulong base, \ in GEN_VEXT_LD_INDEX()
616 vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \ in GEN_VEXT_LD_INDEX()
642 vext_ldff(void *vd, void *v0, target_ulong base, CPURISCVState *env,
683 if (!vm && !vext_elem_mask(v0, i)) {
759 if (!vext_elem_mask(v0, i)) {
780 void HELPER(NAME)(void *vd, void *v0, target_ulong base, \
783 vext_ldff(vd, v0, base, env, desc, LOAD_FN_TLB, \
1126 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \ in RVVCALL()
1141 ETYPE carry = vext_elem_mask(v0, i); \ in RVVCALL()
1161 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
1174 ETYPE carry = vext_elem_mask(v0, i); \
1198 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
1212 ETYPE carry = !vm && vext_elem_mask(v0, i); \
1238 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
1251 ETYPE carry = !vm && vext_elem_mask(v0, i); \
1334 void HELPER(NAME)(void *vd, void *v0, void *vs1, \
1348 if (!vm && !vext_elem_mask(v0, i)) { \
1381 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
1397 if (!vm && !vext_elem_mask(v0, i)) { \
1448 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
1463 if (!vm && !vext_elem_mask(v0, i)) { \
1515 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
1529 if (!vm && !vext_elem_mask(v0, i)) { \
2116 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
2128 ETYPE *vt = (!vext_elem_mask(v0, i) ? vs2 : vs1); \
2142 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
2155 ETYPE d = (!vext_elem_mask(v0, i) ? s2 : \
2193 vext_vv_rm_1(void *vd, void *v0, void *vs1, void *vs2, in vext_vv_rm_1() argument
2199 if (!vm && !vext_elem_mask(v0, i)) { in vext_vv_rm_1()
2210 vext_vv_rm_2(void *vd, void *v0, void *vs1, void *vs2, in vext_vv_rm_2() argument
2225 vext_vv_rm_1(vd, v0, vs1, vs2, in vext_vv_rm_2()
2229 vext_vv_rm_1(vd, v0, vs1, vs2, in vext_vv_rm_2()
2233 vext_vv_rm_1(vd, v0, vs1, vs2, in vext_vv_rm_2()
2237 vext_vv_rm_1(vd, v0, vs1, vs2, in vext_vv_rm_2()
2247 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
2250 vext_vv_rm_2(vd, v0, vs1, vs2, env, desc, \
2320 vext_vx_rm_1(void *vd, void *v0, target_long s1, void *vs2, in vext_vx_rm_1() argument
2326 if (!vm && !vext_elem_mask(v0, i)) { in vext_vx_rm_1()
2337 vext_vx_rm_2(void *vd, void *v0, target_long s1, void *vs2, in vext_vx_rm_2() argument
2352 vext_vx_rm_1(vd, v0, s1, vs2, in vext_vx_rm_2()
2356 vext_vx_rm_1(vd, v0, s1, vs2, in vext_vx_rm_2()
2360 vext_vx_rm_1(vd, v0, s1, vs2, in vext_vx_rm_2()
2364 vext_vx_rm_1(vd, v0, s1, vs2, in vext_vx_rm_2()
2374 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
2378 vext_vx_rm_2(vd, v0, s1, vs2, env, desc, \
3112 void HELPER(NAME)(void *vd, void *v0, void *vs1, \
3127 if (!vm && !vext_elem_mask(v0, i)) { \
3157 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, \
3172 if (!vm && !vext_elem_mask(v0, i)) { \
3746 void HELPER(NAME)(void *vd, void *v0, void *vs2, \
3763 if (!vm && !vext_elem_mask(v0, i)) { \
4270 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \ in RVVCALL()
4285 if (!vm && !vext_elem_mask(v0, i)) { \ in RVVCALL()
4312 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
4326 if (!vm && !vext_elem_mask(v0, i)) { \
4506 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \ in RVVCALL()
4522 (!vm && !vext_elem_mask(v0, i) ? s2 : s1); \ in RVVCALL()
4673 void HELPER(NAME)(void *vd, void *v0, void *vs1, \ in RVVCALL()
4689 if (!vm && !vext_elem_mask(v0, i)) { \ in RVVCALL()
4763 void HELPER(NAME)(void *vd, void *v0, void *vs1, \
4779 if (!vm && !vext_elem_mask(v0, i)) { \
4841 void HELPER(NAME)(void *vd, void *v0, void *vs1, \ in GEN_VEXT_FRED()
4886 target_ulong HELPER(vcpop_m)(void *v0, void *vs2, CPURISCVState *env,
4895 if (vm || vext_elem_mask(v0, i)) {
4906 target_ulong HELPER(vfirst_m)(void *v0, void *vs2, CPURISCVState *env, in HELPER()
4914 if (vm || vext_elem_mask(v0, i)) { in HELPER()
4930 static void vmsetm(void *vd, void *v0, void *vs2, CPURISCVState *env, in vmsetm() argument
4944 if (!vm && !vext_elem_mask(v0, i)) { in vmsetm()
4983 void HELPER(vmsbf_m)(void *vd, void *v0, void *vs2, CPURISCVState *env, in HELPER()
4986 vmsetm(vd, v0, vs2, env, desc, BEFORE_FIRST); in HELPER()
4989 void HELPER(vmsif_m)(void *vd, void *v0, void *vs2, CPURISCVState *env, in HELPER()
4992 vmsetm(vd, v0, vs2, env, desc, INCLUDE_FIRST); in HELPER()
4995 void HELPER(vmsof_m)(void *vd, void *v0, void *vs2, CPURISCVState *env, in HELPER()
4998 vmsetm(vd, v0, vs2, env, desc, ONLY_FIRST); in HELPER()
5003 void HELPER(NAME)(void *vd, void *v0, void *vs2, CPURISCVState *env, \
5018 if (!vm && !vext_elem_mask(v0, i)) { \
5040 void HELPER(NAME)(void *vd, void *v0, CPURISCVState *env, uint32_t desc) \ in GEN_VEXT_VIOTA_M()
5053 if (!vm && !vext_elem_mask(v0, i)) { \ in GEN_VEXT_VIOTA_M()
5076 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5091 if (!vm && !vext_elem_mask(v0, i)) { \
5110 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5127 if (!vm && !vext_elem_mask(v0, i)) { \
5136 if (!vm && !vext_elem_mask(v0, i)) { \
5155 static void vslide1up_##BITWIDTH(void *vd, void *v0, uint64_t s1, \
5171 if (!vm && !vext_elem_mask(v0, i)) { \
5193 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5196 vslide1up_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5206 static void vslide1down_##BITWIDTH(void *vd, void *v0, uint64_t s1, \
5222 if (!vm && !vext_elem_mask(v0, i)) { \
5244 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5247 vslide1down_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5258 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
5261 vslide1up_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5270 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
5273 vslide1down_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5283 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
5299 if (!vm && !vext_elem_mask(v0, i)) { \
5328 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5344 if (!vm && !vext_elem_mask(v0, i)) { \
5368 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
5428 void HELPER(NAME)(void *vd, void *v0, void *vs2, \
5442 if (!vm && !vext_elem_mask(v0, i)) { \