Lines Matching full:v1

139 static __always_inline void fpu_vab(u8 v1, u8 v2, u8 v3)
141 asm volatile("VAB %[v1],%[v2],%[v3]"
143 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3)
147 static __always_inline void fpu_vcksm(u8 v1, u8 v2, u8 v3)
149 asm volatile("VCKSM %[v1],%[v2],%[v3]"
151 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3)
155 static __always_inline void fpu_vesravb(u8 v1, u8 v2, u8 v3)
157 asm volatile("VESRAVB %[v1],%[v2],%[v3]"
159 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3)
163 static __always_inline void fpu_vgfmag(u8 v1, u8 v2, u8 v3, u8 v4)
165 asm volatile("VGFMAG %[v1],%[v2],%[v3],%[v4]"
167 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3), [v4] "I" (v4)
171 static __always_inline void fpu_vgfmg(u8 v1, u8 v2, u8 v3)
173 asm volatile("VGFMG %[v1],%[v2],%[v3]"
175 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3)
181 static __always_inline void fpu_vl(u8 v1, const void *vxr)
184 asm volatile("VL %[v1],%O[vxr],,%R[vxr]\n"
187 [v1] "I" (v1)
193 static __always_inline void fpu_vl(u8 v1, const void *vxr)
198 " VL %[v1],0,,1\n"
201 [v1] "I" (v1)
236 static __always_inline void fpu_vll(u8 v1, u32 index, const void *vxr)
242 asm volatile("VLL %[v1],%[index],%O[vxr],%R[vxr]\n"
246 [v1] "I" (v1)
252 static __always_inline void fpu_vll(u8 v1, u32 index, const void *vxr)
260 " VLL %[v1],%[index],0,1\n"
264 [v1] "I" (v1)
280 asm volatile("VLM %[v1],%[v3],%O[vxrs],%R[vxrs]\n" \
283 [v1] "I" (_v1), [v3] "I" (_v3) \
300 " VLM %[v1],%[v3],0,1\n" \
303 [v1] "I" (_v1), [v3] "I" (_v3) \
310 static __always_inline void fpu_vlr(u8 v1, u8 v2)
312 asm volatile("VLR %[v1],%[v2]"
314 : [v1] "I" (v1), [v2] "I" (v2)
326 static __always_inline void fpu_vn(u8 v1, u8 v2, u8 v3)
328 asm volatile("VN %[v1],%[v2],%[v3]"
330 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3)
334 static __always_inline void fpu_vperm(u8 v1, u8 v2, u8 v3, u8 v4)
336 asm volatile("VPERM %[v1],%[v2],%[v3],%[v4]"
338 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3), [v4] "I" (v4)
342 static __always_inline void fpu_vrepib(u8 v1, s16 i2)
344 asm volatile("VREPIB %[v1],%[i2]"
346 : [v1] "I" (v1), [i2] "K" (i2)
350 static __always_inline void fpu_vsrlb(u8 v1, u8 v2, u8 v3)
352 asm volatile("VSRLB %[v1],%[v2],%[v3]"
354 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3)
360 static __always_inline void fpu_vst(u8 v1, const void *vxr)
363 asm volatile("VST %[v1],%O[vxr],,%R[vxr]\n"
365 : [v1] "I" (v1)
371 static __always_inline void fpu_vst(u8 v1, const void *vxr)
376 " VST %[v1],0,,1\n"
378 : [v1] "I" (v1)
386 static __always_inline void fpu_vstl(u8 v1, u32 index, const void *vxr)
392 asm volatile("VSTL %[v1],%[index],%O[vxr],%R[vxr]\n"
394 : [index] "d" (index), [v1] "I" (v1)
400 static __always_inline void fpu_vstl(u8 v1, u32 index, const void *vxr)
408 " VSTL %[v1],%[index],0,1\n"
410 : [index] "d" (index), [v1] "I" (v1)
426 asm volatile("VSTM %[v1],%[v3],%O[vxrs],%R[vxrs]\n" \
428 : [v1] "I" (_v1), [v3] "I" (_v3) \
445 " VSTM %[v1],%[v3],0,1\n" \
447 : [v1] "I" (_v1), [v3] "I" (_v3) \
454 static __always_inline void fpu_vupllf(u8 v1, u8 v2)
456 asm volatile("VUPLLF %[v1],%[v2]"
458 : [v1] "I" (v1), [v2] "I" (v2)
462 static __always_inline void fpu_vx(u8 v1, u8 v2, u8 v3)
464 asm volatile("VX %[v1],%[v2],%[v3]"
466 : [v1] "I" (v1), [v2] "I" (v2), [v3] "I" (v3)