Lines Matching +full:1 +full:q
29 br_aes_ct64_bitslice_invSbox(uint64_t *q) in br_aes_ct64_bitslice_invSbox() argument
37 q0 = ~q[0]; in br_aes_ct64_bitslice_invSbox()
38 q1 = ~q[1]; in br_aes_ct64_bitslice_invSbox()
39 q2 = q[2]; in br_aes_ct64_bitslice_invSbox()
40 q3 = q[3]; in br_aes_ct64_bitslice_invSbox()
41 q4 = q[4]; in br_aes_ct64_bitslice_invSbox()
42 q5 = ~q[5]; in br_aes_ct64_bitslice_invSbox()
43 q6 = ~q[6]; in br_aes_ct64_bitslice_invSbox()
44 q7 = q[7]; in br_aes_ct64_bitslice_invSbox()
45 q[7] = q1 ^ q4 ^ q6; in br_aes_ct64_bitslice_invSbox()
46 q[6] = q0 ^ q3 ^ q5; in br_aes_ct64_bitslice_invSbox()
47 q[5] = q7 ^ q2 ^ q4; in br_aes_ct64_bitslice_invSbox()
48 q[4] = q6 ^ q1 ^ q3; in br_aes_ct64_bitslice_invSbox()
49 q[3] = q5 ^ q0 ^ q2; in br_aes_ct64_bitslice_invSbox()
50 q[2] = q4 ^ q7 ^ q1; in br_aes_ct64_bitslice_invSbox()
51 q[1] = q3 ^ q6 ^ q0; in br_aes_ct64_bitslice_invSbox()
52 q[0] = q2 ^ q5 ^ q7; in br_aes_ct64_bitslice_invSbox()
54 br_aes_ct64_bitslice_Sbox(q); in br_aes_ct64_bitslice_invSbox()
56 q0 = ~q[0]; in br_aes_ct64_bitslice_invSbox()
57 q1 = ~q[1]; in br_aes_ct64_bitslice_invSbox()
58 q2 = q[2]; in br_aes_ct64_bitslice_invSbox()
59 q3 = q[3]; in br_aes_ct64_bitslice_invSbox()
60 q4 = q[4]; in br_aes_ct64_bitslice_invSbox()
61 q5 = ~q[5]; in br_aes_ct64_bitslice_invSbox()
62 q6 = ~q[6]; in br_aes_ct64_bitslice_invSbox()
63 q7 = q[7]; in br_aes_ct64_bitslice_invSbox()
64 q[7] = q1 ^ q4 ^ q6; in br_aes_ct64_bitslice_invSbox()
65 q[6] = q0 ^ q3 ^ q5; in br_aes_ct64_bitslice_invSbox()
66 q[5] = q7 ^ q2 ^ q4; in br_aes_ct64_bitslice_invSbox()
67 q[4] = q6 ^ q1 ^ q3; in br_aes_ct64_bitslice_invSbox()
68 q[3] = q5 ^ q0 ^ q2; in br_aes_ct64_bitslice_invSbox()
69 q[2] = q4 ^ q7 ^ q1; in br_aes_ct64_bitslice_invSbox()
70 q[1] = q3 ^ q6 ^ q0; in br_aes_ct64_bitslice_invSbox()
71 q[0] = q2 ^ q5 ^ q7; in br_aes_ct64_bitslice_invSbox()
75 add_round_key(uint64_t *q, const uint64_t *sk) in add_round_key() argument
80 q[i] ^= sk[i]; in add_round_key()
85 inv_shift_rows(uint64_t *q) in inv_shift_rows() argument
92 x = q[i]; in inv_shift_rows()
93 q[i] = (x & (uint64_t)0x000000000000FFFF) in inv_shift_rows()
110 inv_mix_columns(uint64_t *q) in inv_mix_columns() argument
115 q0 = q[0]; in inv_mix_columns()
116 q1 = q[1]; in inv_mix_columns()
117 q2 = q[2]; in inv_mix_columns()
118 q3 = q[3]; in inv_mix_columns()
119 q4 = q[4]; in inv_mix_columns()
120 q5 = q[5]; in inv_mix_columns()
121 q6 = q[6]; in inv_mix_columns()
122 q7 = q[7]; in inv_mix_columns()
132 q[0] = q5 ^ q6 ^ q7 ^ r0 ^ r5 ^ r7 ^ rotr32(q0 ^ q5 ^ q6 ^ r0 ^ r5); in inv_mix_columns()
133 q[1] = q0 ^ q5 ^ r0 ^ r1 ^ r5 ^ r6 ^ r7 ^ rotr32(q1 ^ q5 ^ q7 ^ r1 ^ r5 ^ r6); in inv_mix_columns()
134 q[2] = q0 ^ q1 ^ q6 ^ r1 ^ r2 ^ r6 ^ r7 ^ rotr32(q0 ^ q2 ^ q6 ^ r2 ^ r6 ^ r7); in inv_mix_columns()
135 …q[3] = q0 ^ q1 ^ q2 ^ q5 ^ q6 ^ r0 ^ r2 ^ r3 ^ r5 ^ rotr32(q0 ^ q1 ^ q3 ^ q5 ^ q6 ^ q7 ^ r0 ^ r3 ^… in inv_mix_columns()
136 …q[4] = q1 ^ q2 ^ q3 ^ q5 ^ r1 ^ r3 ^ r4 ^ r5 ^ r6 ^ r7 ^ rotr32(q1 ^ q2 ^ q4 ^ q5 ^ q7 ^ r1 ^ r4 ^… in inv_mix_columns()
137 q[5] = q2 ^ q3 ^ q4 ^ q6 ^ r2 ^ r4 ^ r5 ^ r6 ^ r7 ^ rotr32(q2 ^ q3 ^ q5 ^ q6 ^ r2 ^ r5 ^ r6 ^ r7); in inv_mix_columns()
138 q[6] = q3 ^ q4 ^ q5 ^ q7 ^ r3 ^ r5 ^ r6 ^ r7 ^ rotr32(q3 ^ q4 ^ q6 ^ q7 ^ r3 ^ r6 ^ r7); in inv_mix_columns()
139 q[7] = q4 ^ q5 ^ q6 ^ r4 ^ r6 ^ r7 ^ rotr32(q4 ^ q5 ^ q7 ^ r4 ^ r7); in inv_mix_columns()
145 const uint64_t *skey, uint64_t *q) in br_aes_ct64_bitslice_decrypt() argument
149 add_round_key(q, skey + (num_rounds << 3)); in br_aes_ct64_bitslice_decrypt()
150 for (u = num_rounds - 1; u > 0; u --) { in br_aes_ct64_bitslice_decrypt()
151 inv_shift_rows(q); in br_aes_ct64_bitslice_decrypt()
152 br_aes_ct64_bitslice_invSbox(q); in br_aes_ct64_bitslice_decrypt()
153 add_round_key(q, skey + (u << 3)); in br_aes_ct64_bitslice_decrypt()
154 inv_mix_columns(q); in br_aes_ct64_bitslice_decrypt()
156 inv_shift_rows(q); in br_aes_ct64_bitslice_decrypt()
157 br_aes_ct64_bitslice_invSbox(q); in br_aes_ct64_bitslice_decrypt()
158 add_round_key(q, skey); in br_aes_ct64_bitslice_decrypt()