Lines Matching +full:3 +full:v

26 	uint8x16_t v[5];  member
44 vst1q_u8(state, st.v[0]); in aegis128_save_state_neon()
45 vst1q_u8(state + 16, st.v[1]); in aegis128_save_state_neon()
46 vst1q_u8(state + 32, st.v[2]); in aegis128_save_state_neon()
47 vst1q_u8(state + 48, st.v[3]); in aegis128_save_state_neon()
48 vst1q_u8(state + 64, st.v[4]); in aegis128_save_state_neon()
66 uint8x16_t v; in aegis_aes_round() local
73 v = vqtbl4q_u8(vld1q_u8_x4(crypto_aes_sbox), w); in aegis_aes_round()
74 v = vqtbx4q_u8(v, vld1q_u8_x4(crypto_aes_sbox + 0x40), w - 0x40); in aegis_aes_round()
75 v = vqtbx4q_u8(v, vld1q_u8_x4(crypto_aes_sbox + 0x80), w - 0x80); in aegis_aes_round()
76 v = vqtbx4q_u8(v, vld1q_u8_x4(crypto_aes_sbox + 0xc0), w - 0xc0); in aegis_aes_round()
78 asm("tbl %0.16b, {v16.16b-v19.16b}, %1.16b" : "=w"(v) : "w"(w)); in aegis_aes_round()
80 asm("tbx %0.16b, {v20.16b-v23.16b}, %1.16b" : "+w"(v) : "w"(w)); in aegis_aes_round()
82 asm("tbx %0.16b, {v24.16b-v27.16b}, %1.16b" : "+w"(v) : "w"(w)); in aegis_aes_round()
84 asm("tbx %0.16b, {v28.16b-v31.16b}, %1.16b" : "+w"(v) : "w"(w)); in aegis_aes_round()
88 w = (v << 1) ^ (uint8x16_t)(((int8x16_t)v >> 7) & 0x1b); in aegis_aes_round()
89 w ^= (uint8x16_t)vrev32q_u16((uint16x8_t)v); in aegis_aes_round()
90 w ^= vqtbl1q_u8(v ^ w, vld1q_u8(ror32by8)); in aegis_aes_round()
110 m ^= aegis_aes_round(st.v[4]); in aegis128_update_neon()
111 st.v[4] ^= aegis_aes_round(st.v[3]); in aegis128_update_neon()
112 st.v[3] ^= aegis_aes_round(st.v[2]); in aegis128_update_neon()
113 st.v[2] ^= aegis_aes_round(st.v[1]); in aegis128_update_neon()
114 st.v[1] ^= aegis_aes_round(st.v[0]); in aegis128_update_neon()
115 st.v[0] ^= m; in aegis128_update_neon()
193 static uint8x16_t vqtbx1q_u8(uint8x16_t v, uint8x16_t a, uint8x16_t b) in vqtbx1q_u8() argument
200 return vcombine_u8(vtbx2_u8(vget_low_u8(v), __a.pair, vget_low_u8(b)), in vqtbx1q_u8()
201 vtbx2_u8(vget_high_u8(v), __a.pair, vget_high_u8(b))); in vqtbx1q_u8()
204 static int8_t vminvq_s8(int8x16_t v) in vminvq_s8() argument
206 int8x8_t s = vpmin_s8(vget_low_s8(v), vget_high_s8(v)); in vminvq_s8()
218 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15,
232 uint8x16_t s = st.v[1] ^ (st.v[2] & st.v[3]) ^ st.v[4]; in crypto_aegis128_encrypt_chunk_neon()
245 uint8x16_t s = st.v[1] ^ (st.v[2] & st.v[3]) ^ st.v[4]; in crypto_aegis128_encrypt_chunk_neon()
281 msg = vld1q_u8(src) ^ st.v[1] ^ (st.v[2] & st.v[3]) ^ st.v[4]; in crypto_aegis128_decrypt_chunk_neon()
291 uint8x16_t s = st.v[1] ^ (st.v[2] & st.v[3]) ^ st.v[4]; in crypto_aegis128_decrypt_chunk_neon()
323 uint8x16_t v; in crypto_aegis128_final_neon() local
328 v = st.v[3] ^ (uint8x16_t)vcombine_u64(vmov_n_u64(8ULL * assoclen), in crypto_aegis128_final_neon()
332 st = aegis128_update_neon(st, v); in crypto_aegis128_final_neon()
334 v = st.v[0] ^ st.v[1] ^ st.v[2] ^ st.v[3] ^ st.v[4]; in crypto_aegis128_final_neon()
337 v = vqtbl1q_u8(~vceqq_u8(v, vld1q_u8(tag_xor)), in crypto_aegis128_final_neon()
340 return vminvq_s8((int8x16_t)v); in crypto_aegis128_final_neon()
343 vst1q_u8(tag_xor, v); in crypto_aegis128_final_neon()