1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
|
#include "macros.inc"
#include "fpu.h"
test_suite fp0_arith
#if XCHAL_HAVE_DFP
.macro movfp fr, v
movi a2, ((\v) >> 32) & 0xffffffff
movi a3, ((\v) & 0xffffffff)
wfrd \fr, a2, a3
.endm
.macro check_res fr, r, sr
rfrd a2, \fr
dump a2
movi a3, ((\r) >> 32) & 0xffffffff
assert eq, a2, a3
rfr a2, \fr
dump a2
movi a3, ((\r) & 0xffffffff)
assert eq, a2, a3
rur a2, fsr
movi a3, \sr
assert eq, a2, a3
.endm
test add_d
movi a2, 1
wsr a2, cpenable
/* MAX_FLOAT + MAX_FLOAT = +inf/MAX_FLOAT */
test_op2 add.d, f6, f7, f8, F64_MAX, F64_MAX, \
F64_PINF, F64_MAX, F64_PINF, F64_MAX, \
FSR_OI, FSR_OI, FSR_OI, FSR_OI
test_end
test add_d_inf
/* 1 + +inf = +inf */
test_op2 add.d, f6, f7, f8, F64_1, F64_PINF, \
F64_PINF, F64_PINF, F64_PINF, F64_PINF, \
FSR__, FSR__, FSR__, FSR__
/* +inf + -inf = default NaN */
test_op2 add.d, f0, f1, f2, F64_PINF, F64_NINF, \
F64_DNAN, F64_DNAN, F64_DNAN, F64_DNAN, \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
test add_d_nan_dfpu
/* 1 + QNaN = QNaN */
test_op2 add.d, f9, f10, f11, F64_1, F64_QNAN(1), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
/* 1 + SNaN = QNaN */
test_op2 add.d, f12, f13, f14, F64_1, F64_SNAN(1), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* SNaN1 + SNaN2 = QNaN2 */
test_op2 add.d, f15, f0, f1, F64_SNAN(1), F64_SNAN(2), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* QNaN1 + SNaN2 = QNaN2 */
test_op2 add.d, f5, f6, f7, F64_QNAN(1), F64_SNAN(2), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* SNaN1 + QNaN2 = QNaN2 */
test_op2 add.d, f8, f9, f10, F64_SNAN(1), F64_QNAN(2), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
test sub_d
/* norm - norm = denorm */
test_op2 sub.d, f6, f7, f8, F64_MIN_NORM | 1, F64_MIN_NORM, \
0x00000001, 0x00000001, 0x00000001, 0x00000001, \
FSR__, FSR__, FSR__, FSR__
test_end
test mul_d
test_op2 mul.d, f0, f1, f2, F64_1 | 1, F64_1 | 1, \
F64_1 | 2, F64_1 | 2, F64_1 | 3, F64_1 | 2, \
FSR_I, FSR_I, FSR_I, FSR_I
/* MAX_FLOAT/2 * MAX_FLOAT/2 = +inf/MAX_FLOAT */
test_op2 mul.d, f6, f7, f8, F64_MAX_2, F64_MAX_2, \
F64_PINF, F64_MAX, F64_PINF, F64_MAX, \
FSR_OI, FSR_OI, FSR_OI, FSR_OI
/* min norm * min norm = 0/denorm */
test_op2 mul.d, f6, f7, f8, F64_MIN_NORM, F64_MIN_NORM, \
F64_0, F64_0, 0x00000001, F64_0, \
FSR_UI, FSR_UI, FSR_UI, FSR_UI
/* inf * 0 = default NaN */
test_op2 mul.d, f6, f7, f8, F64_PINF, F64_0, \
F64_DNAN, F64_DNAN, F64_DNAN, F64_DNAN, \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
test madd_d
test_op3 madd.d, f0, f1, f2, f0, F64_0, F64_1 | 1, F64_1 | 1, \
F64_1 | 2, F64_1 | 2, F64_1 | 3, F64_1 | 2, \
FSR_I, FSR_I, FSR_I, FSR_I
test_end
test madd_d_precision
test_op3 madd.d, f0, f1, f2, f0, \
F64_MINUS | F64_1 | 2, F64_1 | 1, F64_1 | 1, \
0x3970000000000000, 0x3970000000000000, 0x3970000000000000, 0x3970000000000000, \
FSR__, FSR__, FSR__, FSR__
test_end
test madd_d_nan_dfpu
/* DFPU madd/msub NaN1, NaN2, NaN3 priority: NaN1, NaN3, NaN2 */
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_1, F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_QNAN(2), F64_1, \
F64_QNAN(2), F64_QNAN(2), F64_QNAN(2), F64_QNAN(2), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_1, F64_QNAN(3), \
F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_QNAN(2), F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_1, F64_QNAN(3), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_QNAN(2), F64_QNAN(3), \
F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_QNAN(2), F64_QNAN(3), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
/* inf * 0 = default NaN */
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_PINF, F64_0, \
F64_DNAN, F64_DNAN, F64_DNAN, F64_DNAN, \
FSR_V, FSR_V, FSR_V, FSR_V
/* inf * 0 + SNaN1 = QNaN1 */
test_op3 madd.d, f0, f1, f2, f0, F64_SNAN(1), F64_PINF, F64_0, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* inf * 0 + QNaN1 = QNaN1 */
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_PINF, F64_0, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* madd/msub SNaN turns to QNaN and sets Invalid flag */
test_op3 madd.d, f0, f1, f2, f0, F64_SNAN(1), F64_1, F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_SNAN(2), F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
#endif
test_suite_end
|