Lines Matching full:mask

18 	int mask;  in __kernel_fpu_begin()  local
24 flags &= state->hdr.mask; in __kernel_fpu_begin()
32 mask = flags & KERNEL_VXR; in __kernel_fpu_begin()
33 if (mask == KERNEL_VXR) { in __kernel_fpu_begin()
38 if (mask == KERNEL_VXR_MID) { in __kernel_fpu_begin()
42 mask = flags & KERNEL_VXR_LOW; in __kernel_fpu_begin()
43 if (mask) { in __kernel_fpu_begin()
44 if (mask == KERNEL_VXR_LOW) in __kernel_fpu_begin()
46 else if (mask == KERNEL_VXR_V0V7) in __kernel_fpu_begin()
51 mask = flags & KERNEL_VXR_HIGH; in __kernel_fpu_begin()
52 if (mask) { in __kernel_fpu_begin()
53 if (mask == KERNEL_VXR_HIGH) in __kernel_fpu_begin()
55 else if (mask == KERNEL_VXR_V16V23) in __kernel_fpu_begin()
66 int mask; in __kernel_fpu_end() local
73 flags &= state->hdr.mask; in __kernel_fpu_end()
81 mask = flags & KERNEL_VXR; in __kernel_fpu_end()
82 if (mask == KERNEL_VXR) { in __kernel_fpu_end()
87 if (mask == KERNEL_VXR_MID) { in __kernel_fpu_end()
91 mask = flags & KERNEL_VXR_LOW; in __kernel_fpu_end()
92 if (mask) { in __kernel_fpu_end()
93 if (mask == KERNEL_VXR_LOW) in __kernel_fpu_end()
95 else if (mask == KERNEL_VXR_V0V7) in __kernel_fpu_end()
100 mask = flags & KERNEL_VXR_HIGH; in __kernel_fpu_end()
101 if (mask) { in __kernel_fpu_end()
102 if (mask == KERNEL_VXR_HIGH) in __kernel_fpu_end()
104 else if (mask == KERNEL_VXR_V16V23) in __kernel_fpu_end()
115 int mask; in load_fpu_state() local
124 mask = flags & KERNEL_VXR; in load_fpu_state()
125 if (mask == KERNEL_VXR) { in load_fpu_state()
130 if (mask == KERNEL_VXR_MID) { in load_fpu_state()
134 mask = flags & KERNEL_VXR_LOW; in load_fpu_state()
135 if (mask) { in load_fpu_state()
136 if (mask == KERNEL_VXR_LOW) in load_fpu_state()
138 else if (mask == KERNEL_VXR_V0V7) in load_fpu_state()
143 mask = flags & KERNEL_VXR_HIGH; in load_fpu_state()
144 if (mask) { in load_fpu_state()
145 if (mask == KERNEL_VXR_HIGH) in load_fpu_state()
147 else if (mask == KERNEL_VXR_V16V23) in load_fpu_state()
157 int mask; in save_fpu_state() local
166 mask = flags & KERNEL_VXR; in save_fpu_state()
167 if (mask == KERNEL_VXR) { in save_fpu_state()
172 if (mask == KERNEL_VXR_MID) { in save_fpu_state()
176 mask = flags & KERNEL_VXR_LOW; in save_fpu_state()
177 if (mask) { in save_fpu_state()
178 if (mask == KERNEL_VXR_LOW) in save_fpu_state()
180 else if (mask == KERNEL_VXR_V0V7) in save_fpu_state()
185 mask = flags & KERNEL_VXR_HIGH; in save_fpu_state()
186 if (mask) { in save_fpu_state()
187 if (mask == KERNEL_VXR_HIGH) in save_fpu_state()
189 else if (mask == KERNEL_VXR_V16V23) in save_fpu_state()