| /linux/crypto/ |
| H A D | sm4.c | 104 static inline u32 sm4_round(u32 x0, u32 x1, u32 x2, u32 x3, u32 rk) in sm4_round() argument 106 return x0 ^ sm4_enc_sub(x1 ^ x2 ^ x3 ^ rk); in sm4_round() 122 u32 rk[4]; in sm4_expandkey() local 129 rk[0] = get_unaligned_be32(&key[0]) ^ fk[0]; in sm4_expandkey() 130 rk[1] = get_unaligned_be32(&key[1]) ^ fk[1]; in sm4_expandkey() 131 rk[2] = get_unaligned_be32(&key[2]) ^ fk[2]; in sm4_expandkey() 132 rk[3] = get_unaligned_be32(&key[3]) ^ fk[3]; in sm4_expandkey() 135 rk[0] ^= sm4_key_sub(rk[1] ^ rk[2] ^ rk[3] ^ ck[i + 0]); in sm4_expandkey() 136 rk[1] ^= sm4_key_sub(rk[2] ^ rk[3] ^ rk[0] ^ ck[i + 1]); in sm4_expandkey() 137 rk[2] ^= sm4_key_sub(rk[3] ^ rk[0] ^ rk[1] ^ ck[i + 2]); in sm4_expandkey() [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | aes-neonbs-glue.c | 27 asmlinkage void aesbs_convert_key(u8 out[], u32 const rk[], int rounds); 29 asmlinkage void aesbs_ecb_encrypt(u8 out[], u8 const in[], u8 const rk[], 31 asmlinkage void aesbs_ecb_decrypt(u8 out[], u8 const in[], u8 const rk[], 34 asmlinkage void aesbs_cbc_decrypt(u8 out[], u8 const in[], u8 const rk[], 37 asmlinkage void aesbs_ctr_encrypt(u8 out[], u8 const in[], u8 const rk[], 40 asmlinkage void aesbs_xts_encrypt(u8 out[], u8 const in[], u8 const rk[], 42 asmlinkage void aesbs_xts_decrypt(u8 out[], u8 const in[], u8 const rk[], 46 asmlinkage void neon_aes_ecb_encrypt(u8 out[], u8 const in[], u32 const rk[], 48 asmlinkage void neon_aes_cbc_encrypt(u8 out[], u8 const in[], u32 const rk[], 50 asmlinkage void neon_aes_ctr_encrypt(u8 out[], u8 const in[], u32 const rk[], [all …]
|
| H A D | aes-ce.S | 28 .macro load_round_keys, rk, nr, tmp 29 add \tmp, \rk, \nr, sxtw #4 31 ld1 {v17.4s-v20.4s}, [\rk] 38 .macro enc_prepare, rounds, rk, temp 39 load_round_keys \rk, \rounds, \temp 43 .macro enc_switch_key, rounds, rk, temp 44 load_round_keys \rk, \rounds, \temp 48 .macro dec_prepare, rounds, rk, temp 49 load_round_keys \rk, \rounds, \temp
|
| H A D | aes-neon.S | 98 .macro do_block, enc, in, rounds, rk, rkp, i 99 ld1 {v15.4s}, [\rk] 100 add \rkp, \rk, #16 114 .macro encrypt_block, in, rounds, rk, rkp, i 115 do_block 1, \in, \rounds, \rk, \rkp, \i 118 .macro decrypt_block, in, rounds, rk, rkp, i 119 do_block 0, \in, \rounds, \rk, \rkp, \i 205 .macro do_block_4x, enc, in0, in1, in2, in3, rounds, rk, rkp, i 206 ld1 {v15.4s}, [\rk] 207 add \rkp, \rk, #16 [all …]
|
| H A D | aes-ce-ccm-core.S | 17 .macro load_round_keys, rk, nr, tmp 19 add \tmp, \rk, w\tmp, sxtw #4 20 ld1 {v10.4s-v13.4s}, [\rk]
|
| H A D | sm4-ce-cipher-glue.c | 18 asmlinkage void sm4_ce_do_crypt(const u32 *rk, void *out, const void *in);
|
| H A D | ghash-ce-core.S | 386 .macro load_round_keys, rounds, rk, tmp 387 add \tmp, \rk, #64 388 ld1 {K0.4s-K3.4s}, [\rk] 390 add \tmp, \rk, \rounds, lsl #4 407 .macro enc_block, state, rounds, rk, tmp 408 add \tmp, \rk, #96
|
| /linux/include/crypto/ |
| H A D | aria.h | 407 static inline void aria_add_round_key(u32 *rk, u32 *t0, u32 *t1, u32 *t2, in aria_add_round_key() argument 410 *t0 ^= rk[0]; in aria_add_round_key() 411 *t1 ^= rk[1]; in aria_add_round_key() 412 *t2 ^= rk[2]; in aria_add_round_key() 413 *t3 ^= rk[3]; in aria_add_round_key() 434 static inline void aria_gsrk(u32 *rk, u32 *x, u32 *y, u32 n) in aria_gsrk() argument 439 rk[0] = (x[0]) ^ in aria_gsrk() 442 rk[1] = (x[1]) ^ in aria_gsrk() 445 rk[2] = (x[2]) ^ in aria_gsrk() 448 rk[3] = (x[3]) ^ in aria_gsrk()
|
| H A D | sm4.h | 46 void sm4_crypt_block(const u32 *rk, u8 *out, const u8 *in);
|
| /linux/lib/crypto/ |
| H A D | aes.c | 328 static __always_inline u32 enc_quarterround(const u32 w[4], int i, u32 rk) in enc_quarterround() argument 330 return rk ^ aes_enc_tab[(u8)w[i]] ^ in enc_quarterround() 336 static __always_inline u32 enclast_quarterround(const u32 w[4], int i, u32 rk) in enclast_quarterround() argument 338 return rk ^ ((aes_enc_tab[(u8)w[i]] & 0x0000ff00) >> 8) ^ in enclast_quarterround() 385 static __always_inline u32 dec_quarterround(const u32 w[4], int i, u32 rk) in dec_quarterround() argument 387 return rk ^ aes_dec_tab[(u8)w[i]] ^ in dec_quarterround() 393 static __always_inline u32 declast_quarterround(const u32 w[4], int i, u32 rk) in declast_quarterround() argument 395 return rk ^ aes_inv_sbox[(u8)w[i]] ^ in declast_quarterround()
|
| /linux/lib/crypto/arm/ |
| H A D | sha256-ce.S | 18 rk .req r3 35 vld1.32 {k\ev}, [rk, :128]! 90 adr rk, .Lsha256_rcon 91 vld1.32 {k0}, [rk, :128]!
|
| /linux/arch/x86/crypto/ |
| H A D | aria-gfni-avx512-asm_64.S | 272 t0, rk, round) \ 274 vpbroadcastb ((round * 16) + 3)(rk), t0; \ 276 vpbroadcastb ((round * 16) + 2)(rk), t0; \ 278 vpbroadcastb ((round * 16) + 1)(rk), t0; \ 280 vpbroadcastb ((round * 16) + 0)(rk), t0; \ 282 vpbroadcastb ((round * 16) + 7)(rk), t0; \ 284 vpbroadcastb ((round * 16) + 6)(rk), t0; \ 286 vpbroadcastb ((round * 16) + 5)(rk), t0; \ 288 vpbroadcastb ((round * 16) + 4)(rk), t0; \ 290 vpbroadcastb ((round * 16) + 11)(rk), t0; \ [all …]
|
| H A D | aria-aesni-avx2-asm_64.S | 286 t0, rk, idx, round) \ 288 vpbroadcastb ((round * 16) + idx + 3)(rk), t0; \ 290 vpbroadcastb ((round * 16) + idx + 2)(rk), t0; \ 292 vpbroadcastb ((round * 16) + idx + 1)(rk), t0; \ 294 vpbroadcastb ((round * 16) + idx + 0)(rk), t0; \ 296 vpbroadcastb ((round * 16) + idx + 7)(rk), t0; \ 298 vpbroadcastb ((round * 16) + idx + 6)(rk), t0; \ 300 vpbroadcastb ((round * 16) + idx + 5)(rk), t0; \ 302 vpbroadcastb ((round * 16) + idx + 4)(rk), t0; \ 462 mem_tmp, rk, round) \ argument [all …]
|
| H A D | aria-aesni-avx-asm_64.S | 270 t0, t1, t2, rk, \ 273 vbroadcastss ((round * 16) + idx + 0)(rk), t0; \ 285 vbroadcastss ((round * 16) + idx + 4)(rk), t0; \ 420 mem_tmp, rk, round) \ argument 423 y0, y7, y2, rk, 8, round); \ 438 y0, y7, y2, rk, 0, round); \ 475 mem_tmp, rk, round) \ argument 478 y0, y7, y2, rk, 8, round); \ 493 y0, y7, y2, rk, 0, round); \ 530 mem_tmp, rk, round, last_round) \ argument [all …]
|
| H A D | sm4-avx.h | 8 typedef void (*sm4_crypt_func)(const u32 *rk, u8 *dst, const u8 *src, u8 *iv);
|
| /linux/arch/loongarch/include/asm/ |
| H A D | inst.h | 348 unsigned int rk : 5; member 355 unsigned int rk : 5; member 517 u32 larch_insn_gen_or(enum loongarch_gpr rd, enum loongarch_gpr rj, enum loongarch_gpr rk); 730 enum loongarch_gpr rk) \ 735 insn->reg3_format.rk = rk; \ 781 enum loongarch_gpr rk, \ 788 insn->reg3sa2_format.rk = rk; \
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | metafmt-rkisp1.rst | 4 .. _v4l2-meta-fmt-rk-isp1-stat-3a: 22 .. _v4l2-meta-fmt-rk-isp1-params: 42 .. _v4l2-meta-fmt-rk-isp1-ext-params:
|
| /linux/arch/arm/boot/dts/broadcom/ |
| H A D | bcm4708-linksys-ea6500-v2.dts | 4 * Copyright (C) 2018 Rene Kjellerup <rk.katana.steel@gmail.com>
|
| /linux/arch/loongarch/kernel/ |
| H A D | traps.c | 657 upper = regs->regs[insn.reg3_format.rk]; in do_bce() 664 lower = regs->regs[insn.reg3_format.rk]; in do_bce() 680 upper = regs->regs[insn.reg3_format.rk]; in do_bce() 696 lower = regs->regs[insn.reg3_format.rk]; in do_bce()
|
| H A D | inst.c | 347 u32 larch_insn_gen_or(enum loongarch_gpr rd, enum loongarch_gpr rj, enum loongarch_gpr rk) in larch_insn_gen_or() argument 351 emit_or(&insn, rd, rj, rk); in larch_insn_gen_or()
|
| /linux/Documentation/admin-guide/media/ |
| H A D | rkisp1.rst | 105 :ref:`V4L2_META_FMT_RK_ISP1_STAT_3A <v4l2-meta-fmt-rk-isp1-stat-3a>` as the 122 :ref:`V4L2_META_FMT_RK_ISP1_PARAMS <v4l2-meta-fmt-rk-isp1-params>` as the 127 :ref:`V4L2_META_FMT_RK_ISP1_EXT_PARAMS <v4l2-meta-fmt-rk-isp1-ext-params>` as
|
| /linux/arch/arm/boot/dts/microchip/ |
| H A D | at91-dvk_su60_somc.dtsi | 55 atmel,clk-from-rk-pin;
|
| H A D | sama5d3xmb.dtsi | 37 atmel,clk-from-rk-pin;
|
| /linux/drivers/input/joystick/iforce/ |
| H A D | iforce-ff.c | 126 __u16 rsat, __u16 lsat, __s16 rk, __s16 lk, u16 db, __s16 center) in make_condition_modifier() argument 143 …data[2] = (100 * rk) >> 15; /* Dangerous: the sign is extended by gcc on plateforms providing an a… in make_condition_modifier()
|
| /linux/arch/arm/crypto/ |
| H A D | aes-ce-core.S | 162 .macro prepare_key, rk, rounds 163 add ip, \rk, \rounds, lsl #4 164 vld1.32 {q8-q9}, [\rk] @ load first 2 round keys
|