/linux/arch/mips/kernel/ |
H A D | genex.S | 35 mfc0 k1, CP0_CAUSE 36 andi k1, k1, 0x7c 38 dsll k1, k1, 1 40 PTR_L k0, exception_handlers(k1) 55 mfc0 k1, CP0_CAUSE 57 andi k1, k1, 0x7c 61 beq k1, k0, handle_vced 63 beq k1, k0, handle_vcei 65 dsll k1, k1, 1 68 PTR_L k0, exception_handlers(k1) [all …]
|
H A D | bmips_vec.S | 42 li k1, CKSEG1 43 or k0, k1 58 mfc0 k1, $22, 3 59 srl k1, 16 60 andi k1, 0x8000 61 or k1, CKSEG1 | BMIPS_RELO_VECTOR_CONTROL_0 62 or k0, k1 63 li k1, 0xa0080000 64 sw k1, 0(k0) 70 li k1, CKSEG1 [all …]
|
/linux/drivers/md/dm-vdo/ |
H A D | murmurhash3.c | 53 u64 k1 = get_unaligned_le64(&blocks[i * 2]); in murmurhash3_128() local 56 k1 *= c1; in murmurhash3_128() 57 k1 = ROTL64(k1, 31); in murmurhash3_128() 58 k1 *= c2; in murmurhash3_128() 59 h1 ^= k1; in murmurhash3_128() 80 u64 k1 = 0; in murmurhash3_128() local 111 k1 ^= ((u64)tail[7]) << 56; in murmurhash3_128() 114 k1 ^= ((u64)tail[6]) << 48; in murmurhash3_128() 117 k1 ^= ((u64)tail[5]) << 40; in murmurhash3_128() 120 k1 ^= ((u64)tail[4]) << 32; in murmurhash3_128() [all …]
|
/linux/include/linux/ |
H A D | btree-128.h | 22 static inline void *btree_lookup128(struct btree_head128 *head, u64 k1, u64 k2) in btree_lookup128() argument 24 u64 key[2] = {k1, k2}; in btree_lookup128() 29 u64 *k1, u64 *k2) in btree_get_prev128() argument 31 u64 key[2] = {*k1, *k2}; in btree_get_prev128() 36 *k1 = key[0]; in btree_get_prev128() 41 static inline int btree_insert128(struct btree_head128 *head, u64 k1, u64 k2, in btree_insert128() argument 44 u64 key[2] = {k1, k2}; in btree_insert128() 49 static inline int btree_update128(struct btree_head128 *head, u64 k1, u64 k2, in btree_update128() argument 52 u64 key[2] = {k1, k2}; in btree_update128() 57 static inline void *btree_remove128(struct btree_head128 *head, u64 k1, u64 k2) in btree_remove128() argument [all …]
|
/linux/arch/mips/cavium-octeon/executive/ |
H A D | cvmx-boot-vector.c | 18 dmtc0 $k1, $31, 3 # Save $k1 to KScratch2 21 mfc0 $k1, $15, 1 # Ebase 25 andi $k1, 0x3ff # mask out core ID 28 sll $k1, 5 38 daddu $k0, $k0, $k1 42 dmfc0 $k1, $31, 3 # Restore $k1 from KScratch2 66 4: 40bbf803 dmtc0 k1,c0_kscratch2 69 c: 401b7801 mfc0 k1,c0_ebase 72 14: 337b03ff andi k1,k1,0x3ff 75 1c: 001bd940 sll k1,k1,0x5 [all …]
|
/linux/arch/arm/crypto/ |
H A D | nh-neon-core.S | 33 K1 .req q5 49 .macro _nh_stride k0, k1, k2, k3 59 vadd.u32 T1, T3, \k1 78 vld1.32 {K0,K1}, [KEY]! 88 _nh_stride K0, K1, K2, K3 89 _nh_stride K1, K2, K3, K0 90 _nh_stride K2, K3, K0, K1 91 _nh_stride K3, K0, K1, K2 98 _nh_stride K0, K1, K2, K3 102 _nh_stride K1, K2, K3, K0 [all …]
|
H A D | sha1-ce-core.S | 17 k1 .req q1 68 vld1.32 {k0-k1}, [ip, :128]! 94 add_update c, 0, k1, 8, 9, 10, 11 96 add_update p, 1, k1, 9, 10, 11, 8 97 add_update p, 0, k1, 10, 11, 8, 9 98 add_update p, 1, k1, 11, 8, 9, 10 99 add_update p, 0, k1, 8, 9, 10, 11
|
/linux/arch/sh/boards/mach-hp6xx/ |
H A D | pm_wakeup.S | 12 * k1 scratch 17 #define k1 r1 macro 21 mov #-126, k1 23 mov.b k0, @k1 25 mov.l 5f, k1 27 mov.w k0, @k1 29 mov.l 4f, k1 30 jmp @k1
|
/linux/arch/mips/mm/ |
H A D | cex-oct.S | 31 PTR_LA k1, cache_err_dcache 33 PTR_ADDU k1, k0, k1 /* k1 = &cache_err_dcache[core_id] */ 36 sd k0, (k1) 40 mfc0 k1, CP0_STATUS 41 andi k1, k1, ST0_EXL 42 beqz k1, 1f
|
H A D | cex-sb1.S | 35 * (0x170-0x17f) are used to preserve k0, k1, and ra. 42 * save/restore k0 and k1 from low memory (Useg is direct 49 sd k1,0x178($0) 59 mfc0 k1,C0_ERRCTL 60 bgtz k1,attempt_recovery 61 sll k0,k1,1 98 li k1,1 << 26 /* ICACHE_EXTERNAL */ 99 and k1,k0 100 bnez k1,unrecoverable 118 ld k1,0x178($0) [all …]
|
/linux/arch/x86/crypto/ |
H A D | nh-sse2-x86_64.S | 18 #define K1 %xmm5 macro 34 .macro _nh_stride k0, k1, k2, k3, offset 46 paddd \k1, T1 78 movdqu 0x10(KEY), K1 89 _nh_stride K0, K1, K2, K3, 0x00 90 _nh_stride K1, K2, K3, K0, 0x10 91 _nh_stride K2, K3, K0, K1, 0x20 92 _nh_stride K3, K0, K1, K2, 0x30 101 _nh_stride K0, K1, K2, K3, 0x00 105 _nh_stride K1, K2, K3, K0, 0x10 [all …]
|
H A D | nh-avx2-x86_64.S | 19 #define K1 %ymm5 macro 40 .macro _nh_2xstride k0, k1, k2, k3 44 vpaddd \k1, T3, T1 76 vmovdqu 0x10(KEY), K1 89 _nh_2xstride K0, K1, K2, K3 93 vmovdqu 0x30(KEY), K1 94 _nh_2xstride K2, K3, K0, K1 112 _nh_2xstride K0, K1, K2, K3 118 vmovdqa K3, K1 127 _nh_2xstride K0, K1, K2, K3
|
/linux/arch/arm64/crypto/ |
H A D | nh-neon-core.S | 23 K1 .req v5 35 .macro _nh_stride k0, k1, k2, k3 45 add T1.4s, T3.4s, \k1\().4s 68 ld1 {K0.4s,K1.4s}, [KEY], #32 78 _nh_stride K0, K1, K2, K3 79 _nh_stride K1, K2, K3, K0 80 _nh_stride K2, K3, K0, K1 81 _nh_stride K3, K0, K1, K2 88 _nh_stride K0, K1, K2, K3 92 _nh_stride K1, K2, K3, K0 [all …]
|
H A D | sha1-ce-core.S | 15 k1 .req v1 71 loadrc k1.4s, 0x6ed9eba1, w6 99 add_update c, ev, k1, 8, 9, 10, 11 101 add_update p, od, k1, 9, 10, 11, 8 102 add_update p, ev, k1, 10, 11, 8, 9 103 add_update p, od, k1, 11, 8, 9, 10 104 add_update p, ev, k1, 8, 9, 10, 11
|
/linux/arch/sh/kernel/cpu/sh3/ |
H A D | entry.S | 34 * ldc k1, ssr ! delay slot 68 #define k1 r1 macro 82 * k1 scratch 257 mov #0xfffffff0, k1 258 extu.b k1, k1 259 not k1, k1 260 and k1, k2 ! Mask original SR value 303 ! k1 trashed 314 stc spc, k1 316 cmp/hs k0, k1 ! test k1 (saved PC) >= k0 (saved r0) [all …]
|
/linux/arch/s390/crypto/ |
H A D | chacha-s390.S | 53 #define K1 %v17 macro 90 VL K1,0,,KEY # load key 99 VREPF XB0,K1,0 # smash the key 100 VREPF XB1,K1,1 101 VREPF XB2,K1,2 102 VREPF XB3,K1,3 284 VAF XB0,XB0,K1 307 VAF XB0,XB1,K1 334 VAF XB0,XB2,K1 361 VAF XB0,XB3,K1 [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/dccg/dcn32/ |
H A D | dcn32_dccg.c | 61 uint32_t *k1, in dccg32_get_pixel_rate_div() argument 67 *k1 = PIXEL_RATE_DIV_NA; in dccg32_get_pixel_rate_div() 96 *k1 = val_k1; in dccg32_get_pixel_rate_div() 103 enum pixel_rate_div k1, in dccg32_set_pixel_rate_div() argument 111 // K1 / K2 field is only 1 / 2 bits wide in dccg32_set_pixel_rate_div() 112 if (k1 == PIXEL_RATE_DIV_NA || k2 == PIXEL_RATE_DIV_NA) { in dccg32_set_pixel_rate_div() 118 if (k1 == cur_k1 && k2 == cur_k2) in dccg32_set_pixel_rate_div() 124 OTG0_PIXEL_RATE_DIVK1, k1, in dccg32_set_pixel_rate_div() 129 OTG1_PIXEL_RATE_DIVK1, k1, in dccg32_set_pixel_rate_div() 134 OTG2_PIXEL_RATE_DIVK1, k1, in dccg32_set_pixel_rate_div() [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/dccg/dcn314/ |
H A D | dcn314_dccg.c | 61 uint32_t *k1, in dccg314_get_pixel_rate_div() argument 67 *k1 = PIXEL_RATE_DIV_NA; in dccg314_get_pixel_rate_div() 96 *k1 = val_k1; in dccg314_get_pixel_rate_div() 103 enum pixel_rate_div k1, in dccg314_set_pixel_rate_div() argument 111 // K1 / K2 field is only 1 / 2 bits wide in dccg314_set_pixel_rate_div() 112 if (k1 == PIXEL_RATE_DIV_NA || k2 == PIXEL_RATE_DIV_NA) { in dccg314_set_pixel_rate_div() 118 if (k1 == cur_k1 && k2 == cur_k2) in dccg314_set_pixel_rate_div() 124 OTG0_PIXEL_RATE_DIVK1, k1, in dccg314_set_pixel_rate_div() 129 OTG1_PIXEL_RATE_DIVK1, k1, in dccg314_set_pixel_rate_div() 134 OTG2_PIXEL_RATE_DIVK1, k1, in dccg314_set_pixel_rate_div() [all …]
|
/linux/fs/bcachefs/ |
H A D | sb-clean.c | 103 struct bkey_i *k1, *k2; in bch2_verify_superblock_clean() local 106 k1 = btree_root_find(c, clean, NULL, i, &l1); in bch2_verify_superblock_clean() 109 if (!k1 && !k2) in bch2_verify_superblock_clean() 115 if (k1) in bch2_verify_superblock_clean() 116 bch2_bkey_val_to_text(&buf1, c, bkey_i_to_s_c(k1)); in bch2_verify_superblock_clean() 125 mustfix_fsck_err_on(!k1 || !k2 || in bch2_verify_superblock_clean() 126 IS_ERR(k1) || in bch2_verify_superblock_clean() 128 k1->k.u64s != k2->k.u64s || in bch2_verify_superblock_clean() 129 memcmp(k1, k2, bkey_bytes(&k1->k)) || in bch2_verify_superblock_clean()
|
/linux/fs/ext4/ |
H A D | hash.c | 46 #define K1 0 macro 58 ROUND(F, a, b, c, d, in[0] + K1, 3); in half_md4_transform() 59 ROUND(F, d, a, b, c, in[1] + K1, 7); in half_md4_transform() 60 ROUND(F, c, d, a, b, in[2] + K1, 11); in half_md4_transform() 61 ROUND(F, b, c, d, a, in[3] + K1, 19); in half_md4_transform() 62 ROUND(F, a, b, c, d, in[4] + K1, 3); in half_md4_transform() 63 ROUND(F, d, a, b, c, in[5] + K1, 7); in half_md4_transform() 64 ROUND(F, c, d, a, b, in[6] + K1, 11); in half_md4_transform() 65 ROUND(F, b, c, d, a, in[7] + K1, 19); in half_md4_transform() 95 #undef K1
|
/linux/lib/raid6/ |
H A D | avx512.c | 71 "vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5121_gen_syndrome() 72 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_gen_syndrome() 82 asm volatile("vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5121_gen_syndrome() 83 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_gen_syndrome() 126 "vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t" in raid6_avx5121_xor_syndrome() 127 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_xor_syndrome() 140 "vpcmpgtb %%zmm4,%%zmm5,%%k1\n\t" in raid6_avx5121_xor_syndrome() 141 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5121_xor_syndrome() 201 "vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" in raid6_avx5122_gen_syndrome() 203 "vpmovm2b %%k1,%%zmm5\n\t" in raid6_avx5122_gen_syndrome() [all …]
|
/linux/crypto/ |
H A D | rmd160.c | 26 #define K1 RMD_K1 macro 68 ROUND(aa, bb, cc, dd, ee, F1, K1, in[0], 11); in rmd160_transform() 69 ROUND(ee, aa, bb, cc, dd, F1, K1, in[1], 14); in rmd160_transform() 70 ROUND(dd, ee, aa, bb, cc, F1, K1, in[2], 15); in rmd160_transform() 71 ROUND(cc, dd, ee, aa, bb, F1, K1, in[3], 12); in rmd160_transform() 72 ROUND(bb, cc, dd, ee, aa, F1, K1, in[4], 5); in rmd160_transform() 73 ROUND(aa, bb, cc, dd, ee, F1, K1, in[5], 8); in rmd160_transform() 74 ROUND(ee, aa, bb, cc, dd, F1, K1, in[6], 7); in rmd160_transform() 75 ROUND(dd, ee, aa, bb, cc, F1, K1, in[7], 9); in rmd160_transform() 76 ROUND(cc, dd, ee, aa, bb, F1, K1, in[8], 11); in rmd160_transform() [all …]
|
H A D | khazad.c | 763 u64 K2, K1; in khazad_setkey() local 767 K1 = ((u64)be32_to_cpu(key[2]) << 32) | be32_to_cpu(key[3]); in khazad_setkey() 771 ctx->E[r] = T0[(int)(K1 >> 56) ] ^ in khazad_setkey() 772 T1[(int)(K1 >> 48) & 0xff] ^ in khazad_setkey() 773 T2[(int)(K1 >> 40) & 0xff] ^ in khazad_setkey() 774 T3[(int)(K1 >> 32) & 0xff] ^ in khazad_setkey() 775 T4[(int)(K1 >> 24) & 0xff] ^ in khazad_setkey() 776 T5[(int)(K1 >> 16) & 0xff] ^ in khazad_setkey() 777 T6[(int)(K1 >> 8) & 0xff] ^ in khazad_setkey() 778 T7[(int)(K1 ) & 0xff] ^ in khazad_setkey() [all …]
|
/linux/fs/xfs/libxfs/ |
H A D | xfs_alloc_btree.c | 219 const union xfs_btree_key *k1, in xfs_bnobt_diff_two_keys() argument 225 return (int64_t)be32_to_cpu(k1->alloc.ar_startblock) - in xfs_bnobt_diff_two_keys() 232 const union xfs_btree_key *k1, in xfs_cntbt_diff_two_keys() argument 241 diff = be32_to_cpu(k1->alloc.ar_blockcount) - in xfs_cntbt_diff_two_keys() 246 return be32_to_cpu(k1->alloc.ar_startblock) - in xfs_cntbt_diff_two_keys() 362 const union xfs_btree_key *k1, in xfs_bnobt_keys_inorder() argument 365 return be32_to_cpu(k1->alloc.ar_startblock) < in xfs_bnobt_keys_inorder() 383 const union xfs_btree_key *k1, in xfs_cntbt_keys_inorder() argument 386 return be32_to_cpu(k1->alloc.ar_blockcount) < in xfs_cntbt_keys_inorder() 388 (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && in xfs_cntbt_keys_inorder() [all …]
|
/linux/net/sunrpc/auth_gss/ |
H A D | gss_krb5_keys.c | 433 * K1 = HMAC-SHA(key, 0x00000001 | label | 0x00 | k) 445 u32 outlen, struct xdr_netobj *K1) in krb5_hmac_K1() argument 469 ret = crypto_shash_final(desc, K1->data); in krb5_hmac_K1() 492 * k-truncate(K1) 507 struct xdr_netobj K1 = { in krb5_kdf_hmac_sha2() local 527 K1.len = crypto_shash_digestsize(tfm); in krb5_kdf_hmac_sha2() 528 K1.data = kmalloc(K1.len, gfp_mask); in krb5_kdf_hmac_sha2() 529 if (!K1.data) { in krb5_kdf_hmac_sha2() 534 ret = krb5_hmac_K1(tfm, label, outkey->len, &K1); in krb5_kdf_hmac_sha2() 539 memcpy(outkey->data, K1.data, outkey->len); in krb5_kdf_hmac_sha2() [all …]
|