Lines Matching full:v1
12 static inline __vr _vel_approx_vfdivs_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_vfdivs_vvvl() argument
15 v5 = _vel_vrcps_vvl(v1, l); in _vel_approx_vfdivs_vvvl()
17 v4 = _vel_vfnmsbs_vsvvl(s0, v1, v5, l); in _vel_approx_vfdivs_vvvl()
20 v4 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl()
22 v0 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl()
27 static inline __vr _vel_approx_pvfdiv_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_pvfdiv_vvvl() argument
30 v5 = _vel_pvrcp_vvl(v1, l); in _vel_approx_pvfdiv_vvvl()
32 v4 = _vel_pvfnmsb_vsvvl(s0, v1, v5, l); in _vel_approx_pvfdiv_vvvl()
35 v4 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl()
37 v0 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl()
44 __vr v1, v2, v3, v4; in _vel_approx_vfdivs_vsvl() local
49 v1 = _vel_vfmuls_vsvl(s0, v2, l); in _vel_approx_vfdivs_vsvl()
50 v3 = _vel_vfnmsbs_vsvvl(s0, v1, v0, l); in _vel_approx_vfdivs_vsvl()
51 v1 = _vel_vfmads_vvvvl(v1, v4, v3, l); in _vel_approx_vfdivs_vsvl()
52 v3 = _vel_vfnmsbs_vsvvl(s0, v1, v0, l); in _vel_approx_vfdivs_vsvl()
53 v0 = _vel_vfmads_vvvvl(v1, v2, v3, l); in _vel_approx_vfdivs_vsvl()
59 __vr v1, v2; in _vel_approx_vfdivs_vvsl() local
61 v1 = _vel_vfmuls_vsvl(s1, v0, l); in _vel_approx_vfdivs_vvsl()
62 v2 = _vel_vfnmsbs_vvsvl(v0, s0, v1, l); in _vel_approx_vfdivs_vvsl()
63 v0 = _vel_vfmads_vvsvl(v1, s1, v2, l); in _vel_approx_vfdivs_vvsl()
68 __vr v1, v2, v3; in _vel_approx_vfdivd_vsvl() local
73 v1 = _vel_vfnmsbd_vsvvl(s1, v0, v2, l); in _vel_approx_vfdivd_vsvl()
74 v1 = _vel_vfmadd_vvvvl(v2, v2, v1, l); in _vel_approx_vfdivd_vsvl()
75 v1 = _vel_vaddul_vsvl(1, v1, l); in _vel_approx_vfdivd_vsvl()
76 v3 = _vel_vfnmsbd_vsvvl(s1, v0, v1, l); in _vel_approx_vfdivd_vsvl()
77 v3 = _vel_vfmadd_vvvvl(v1, v1, v3, l); in _vel_approx_vfdivd_vsvl()
78 v1 = _vel_vfmuld_vsvl(s0, v3, l); in _vel_approx_vfdivd_vsvl()
79 v0 = _vel_vfnmsbd_vsvvl(s0, v1, v0, l); in _vel_approx_vfdivd_vsvl()
80 v0 = _vel_vfmadd_vvvvl(v1, v3, v0, l); in _vel_approx_vfdivd_vsvl()
86 __vr v1, v2, v3; in _vel_approx_vfsqrtd_vvl() local
88 v1 = _vel_vfmuld_vvvl(v0, v2, l); in _vel_approx_vfsqrtd_vvl()
91 v3 = _vel_vfnmsbd_vsvvl(s0, v1, v2, l); in _vel_approx_vfsqrtd_vvl()
94 v1 = _vel_vfmuld_vvvl(v0, v2, l); in _vel_approx_vfsqrtd_vvl()
95 v3 = _vel_vfnmsbd_vsvvl(s0, v1, v2, l); in _vel_approx_vfsqrtd_vvl()
97 v0 = _vel_vfmadd_vvvvl(v1, v1, v3, l); in _vel_approx_vfsqrtd_vvl()
103 __vr v1, v2, v3; in _vel_approx_vfsqrts_vvl() local
106 v1 = _vel_vfmuld_vvvl(v0, v2, l); in _vel_approx_vfsqrts_vvl()
109 v3 = _vel_vfnmsbd_vsvvl(s0, v1, v2, l); in _vel_approx_vfsqrts_vvl()
112 v1 = _vel_vfmuld_vvvl(v0, v2, l); in _vel_approx_vfsqrts_vvl()
113 v3 = _vel_vfnmsbd_vsvvl(s0, v1, v2, l); in _vel_approx_vfsqrts_vvl()
115 v0 = _vel_vfmadd_vvvvl(v1, v1, v3, l); in _vel_approx_vfsqrts_vvl()