Home
last modified time | relevance | path

Searched refs:fp8 (Results 1 – 9 of 9) sorted by relevance

/CafeSDK-2.12.13-1/system/src/lib/mtx/
Dmtx44Vec_asm.s48 psq_l fp8, 32(m), 0, 0
54 ps_mul fp8, fp0, fp8
56 ps_madd fp9, fp1, fp9, fp8
87 psq_l fp8, 0(srcBase), 0, 0 // fp8 <- src.x, src.y
92 ps_mul fp13, fp6, fp8
98 ps_mul fp10, fp0, fp8
100 ps_mul fp11, fp2, fp8
102 ps_mul fp12, fp4, fp8
116 psq_lu fp8, 4(srcBase), 0, 0
126 ps_mul fp13, fp6, fp8
[all …]
DmtxVec_asm.s31 #define D2_XY fp8
69 ps_msub fp8, fp0, fp6, fp7
74 ps_merge01 fp10, fp5, fp8
538 psq_l fp8, 16(m), 0, 0
544 ps_mul fp10, fp8, fp0
616 ps_madds0 fp8, fp0, fp6, fp3
620 ps_madds1 fp8, fp1, fp6, fp8
630 ps_madds0 fp12, fp2, fp7, fp8
636 ps_madds0 fp8, fp0, fp6, fp3
642 ps_madds1 fp8, fp1, fp6, fp8
[all …]
Dmtx44Vec.c77 f32x2 fp0, fp1, fp2, fp3, fp4, fp5, fp6, fp7, fp8, fp9, fp12, fp13; //fp10, fp11, in PSMTX44MultVec() local
140 fp8 = __PSQ_LX(m, 32, 0, 0); in PSMTX44MultVec()
160 fp8 = __PS_MUL(fp0, fp8); in PSMTX44MultVec()
166 fp9 = __PS_MADD(fp1, fp9, fp8); in PSMTX44MultVec()
301 f32x2 fp0, fp1, fp2, fp3, fp4, fp5, fp6, fp7, fp8, fp9, fp10, fp11, fp12, fp13; in PSMTX44MultVecSR() local
321 fp8 = __PS_MUL(fp0, fp6); in PSMTX44MultVecSR()
347 fp8 = __PS_SUM0(fp8, fp8, fp8); in PSMTX44MultVecSR()
366 fp9 = __PS_MADD(fp1, fp7, fp8); in PSMTX44MultVecSR()
Dmtx_asm.s112 #define B10_B11 fp8
550 #define trns2 fp8
620 ps_merge10 fp8, fp5, fp4
626 ps_mul fp12, fp1, fp8
628 ps_msub fp13, fp3, fp8, fp13
638 ps_mul fp8, fp1, fp2
650 ps_msub fp8, fp0, fp3, fp8
703 ps_muls0 fp8, fp8, fp0
713 psq_st fp8, 40(inv), 1, 0
721 ps_nmadd fp7, fp8, fp3, fp7
[all …]
Dmtx.c640 f32x2 fp8; in PSMTXInverse() local
675 fp8 = __PS_MERGE10(fp5, fp4); in PSMTXInverse()
684 fp12 = __PS_MUL(fp1, fp8); in PSMTXInverse()
687 fp13 = __PS_MSUB(fp3, fp8, fp13); in PSMTXInverse()
702 fp8 = __PS_MUL(fp1, fp2); in PSMTXInverse()
720 fp8 = __PS_MSUB(fp0, fp3, fp8); in PSMTXInverse()
796 fp8 = __PS_MULS0(fp8, fp0); in PSMTXInverse()
813 __PSQ_STX(inv, 40, fp8, 1, 0); in PSMTXInverse()
827 fp7 = __PS_NMADD(fp8, fp3, fp7); in PSMTXInverse()
945 f32x2 fp8; in PSMTXInvXpose() local
[all …]
Dmtx44_asm.s165 ps_muls0 fp8 , fp2, fp0 // b00a10,b01a10
166 ps_madds1 fp8 , fp3, fp0, fp8 // b00a10+b10a11,b01a11+b11a11
170 ps_madds0 fp8 , fp4, fp1, fp8
173 ps_madds1 fp8 , fp5, fp1, fp8
220 psq_st fp8 ,16(ab), 0, 0 // ab10,ab11
275 #define tmp3 fp8
427 psq_l fp8, 32(src), 0, 0 // fp8 <- src20,src21
436 ps_muls0 fp8, fp8, zS // fp8 <- src20*zS,src21*zS
440 psq_st fp8, 32(dst), 0, 0 // dst20,dst21
521 psq_l fp8, 32(src), 0, 0
[all …]
DmtxVec.c610 f32x2 fp4, fp5, fp6, fp7, fp8, fp9, fp10; in PSVECCrossProduct() local
635 fp8 = __PS_MSUB(fp0, fp6, fp7); in PSVECCrossProduct()
641 fp10 = __PS_MERGE01(fp5, fp8); in PSVECCrossProduct()
973 f32x2 fp0, fp1, fp2, fp3, fp4, fp5, fp6, fp8, fp9, fp10, fp11, fp12; // fp7, in PSMTXMultVec() local
997 fp8 = __PSQ_LX(m, 16, 0, 0); in PSMTXMultVec()
1007 fp10 = __PS_MUL(fp8, fp0); in PSMTXMultVec()
1164 f32x2 fp0, fp1, fp2, fp3, fp4, fp5, fp6, fp7, fp8, fp9, fp10, fp11, fp12, fp13; in PSMTXMultVecSR() local
1176 fp8 = __PS_MUL(fp0, fp6); in PSMTXMultVecSR()
1195 fp8 = __PS_SUM0(fp8, fp8, fp8); in PSMTXMultVecSR()
1206 fp9 = __PS_MADD(fp1, fp7, fp8); in PSMTXMultVecSR()
Dmtx44.c454 f32x2 fp0, fp1, fp2, fp3, fp4, fp5, fp6, fp7, fp8, fp9, fp10, fp11, fp12, fp13; in PSMTX44Concat() local
511 fp8 = __PS_MULS0(fp2, fp0); in PSMTX44Concat()
514 fp8 = __PS_MADDS1(fp3, fp0, fp8); in PSMTX44Concat()
523 fp8 = __PS_MADDS0(fp4, fp1, fp8); in PSMTX44Concat()
527 fp8 = __PS_MADDS1(fp5, fp1, fp8); in PSMTX44Concat()
649 __PSQ_STX(ab, 16, fp8, 0, 0); in PSMTX44Concat()
1053 f32x2 fp4, fp5, fp6, fp7, fp8; //fp0, fp1, fp2, fp3, in PSMTX44TransApply() local
1089 fp8 = __PSQ_LX(src, 32, 0, 0); in PSMTX44TransApply()
1101 __PSQ_STX(dst, 32, fp8, 0, 0); in PSMTX44TransApply()
1213 f32x2 fp4, fp5, fp6, fp7, fp8, fp9, fp10, fp11; //fp0, fp1, fp2, fp3, in PSMTX44ScaleApply() local
[all …]
DmtxQuat_asm.s115 #define pnznw fp8
205 #define nninv fp8
357 #define nwork1 fp8
441 #define tmp3 fp8