| /freebsd/crypto/libecc/src/nn/ |
| H A D | nn_mod_pow.c | 44 u8 expbit, rbit; in _nn_exp_monty_ladder_ltr() local 68 ret = nn_getbit(&mask, (bitcnt_t)(explen - 1), &rbit); EG(ret, err); in _nn_exp_monty_ladder_ltr() 74 ret = nn_mod(&T[rbit], base, mod); EG(ret, err); in _nn_exp_monty_ladder_ltr() 77 ret = nn_mul_redc1(&T[rbit], &T[rbit], r_square, mod, mpinv); EG(ret, err); in _nn_exp_monty_ladder_ltr() 83 ret = nn_mul_redc1(&T[rbit], base, r_square, mod, mpinv); EG(ret, err); in _nn_exp_monty_ladder_ltr() 86 ret = nn_copy(&T[rbit], base); EG(ret, err); in _nn_exp_monty_ladder_ltr() 94 ret = nn_mul_redc1(&T[1-rbit], &T[rbit], &T[rbit], mod, mpinv); EG(ret, err); in _nn_exp_monty_ladder_ltr() 97 ret = nn_mod_mul(&T[1-rbit], &T[rbit], &T[rbit], mod); EG(ret, err); in _nn_exp_monty_ladder_ltr() 114 ret = nn_mul_redc1(&T[2], &T[expbit ^ rbit], &T[expbit ^ rbit], mod, mpinv); EG(ret, err); in _nn_exp_monty_ladder_ltr() 117 ret = nn_mod_mul(&T[2], &T[expbit ^ rbit], &T[expbit ^ rbit], mod); EG(ret, err); in _nn_exp_monty_ladder_ltr() [all …]
|
| /freebsd/lib/libc/aarch64/string/ |
| H A D | strlcpy.S | 108 rbit x8, x8 // simulate x86 tzcnt 140 2: rbit x8, x7 156 rbit x8, x5 182 rbit x8, x8 192 rbit x8, x8 224 2: rbit x6, x7 240 rbit x8, x5 266 rbit x8, x5
|
| H A D | memccpy.S | 43 rbit x8, x8 68 rbit x8, x5 138 rbit x8, x8 // simulate x86 tzcnt 161 rbit x8, x5 199 rbit x8, x8 206 rbit x8, x8
|
| H A D | strlen.S | 24 rbit x1, x1 40 rbit x1, x1 // reverse bits as NEON has no ctz
|
| H A D | strcmp.S | 164 rbit x2, x5 201 rbit x2, x5 215 rbit x2, x5 300 rbit x2, x5 331 rbit x2, x5
|
| H A D | strncmp.S | 185 rbit x3, x5 348 rbit x3, x5 364 rbit x3, x5 531 rbit x3, x5 547 rbit x3, x5
|
| /freebsd/crypto/openssl/crypto/sm4/asm/ |
| H A D | vpsm4-armv8.pl | 86 sub rbit() { subroutine 526 &rbit(@vtmp[2],$src,$std); 536 &rbit($des,$des,$std); 1275 &rbit(@tweak[0],@tweak[0],$std); 1301 &rbit(@vtmp[0],@vtmp[0],$std); 1302 &rbit(@vtmp[1],@vtmp[1],$std); 1303 &rbit(@vtmp[2],@vtmp[2],$std); 1304 &rbit(@vtmp[3],@vtmp[3],$std); 1312 &rbit(@vtmpx[0],@vtmpx[0],$std); 1313 &rbit(@vtmpx[1],@vtmpx[1],$std); [all …]
|
| H A D | vpsm4_ex-armv8.pl | 89 sub rbit() { subroutine 527 &rbit(@vtmp[2],$src,$std); 537 &rbit($des,$des,$std); 1275 &rbit(@tweak[0],@tweak[0],$std); 1309 &rbit(@tweak[0],@tweak[0],$std); 1310 &rbit(@tweak[1],@tweak[1],$std); 1311 &rbit(@tweak[2],@tweak[2],$std); 1312 &rbit(@tweak[3],@tweak[3],$std); 1320 &rbit(@tweak[4],@tweak[4],$std); 1321 &rbit(@tweak[5],@tweak[5],$std); [all …]
|
| /freebsd/sys/powerpc/powermac/ |
| H A D | hrowpic.c | 180 u_int rbit; in hrowpic_toggle_irq() local 195 rbit = HPIC_INT_TO_REGBIT(irq); in hrowpic_toggle_irq() 198 sc->sc_softreg[roffset] |= (1 << rbit); in hrowpic_toggle_irq() 200 sc->sc_softreg[roffset] &= ~(1 << rbit); in hrowpic_toggle_irq()
|
| /freebsd/contrib/arm-optimized-routines/string/aarch64/ |
| H A D | strlen-mte.S | 45 rbit synd, synd 58 rbit synd, synd 81 rbit synd, synd
|
| H A D | strcpy.S | 71 rbit synd, synd 86 rbit synd, synd 144 rbit synd, synd
|
| H A D | strchrnul-mte.S | 52 rbit tmp1, tmp1 77 rbit tmp1, tmp1
|
| H A D | strnlen.S | 49 rbit synd, synd 91 rbit synd, synd
|
| H A D | strchr-mte.S | 55 rbit tmp1, tmp1 89 rbit tmp1, tmp1
|
| H A D | memchr-mte.S | 54 rbit synd, synd 95 rbit synd, synd
|
| H A D | strlen.S | 164 rbit synd, synd 187 rbit synd, synd
|
| H A D | memrchr.S | 98 rbit synd, synd
|
| H A D | strchrnul.S | 104 rbit tmp1, tmp1
|
| H A D | strchr.S | 113 rbit tmp1, tmp1
|
| H A D | memchr.S | 126 rbit synd, synd
|
| H A D | strrchr-mte.S | 83 rbit synd, synd
|
| /freebsd/crypto/libecc/src/curves/ |
| H A D | prj_pt.c | 1322 u8 mbit, rbit; in _prj_pt_mul_ltr_monty_dbl_add_always() local 1378 ret = nn_getbit(&r, mlen, &rbit); EG(ret, err); in _prj_pt_mul_ltr_monty_dbl_add_always() 1392 ret = prj_pt_copy(&T[rbit], &T[2]); EG(ret, err); in _prj_pt_mul_ltr_monty_dbl_add_always() 1411 ret_ops |= prj_pt_add(&T[rbit], &T[rbit], &T[rbit]); in _prj_pt_mul_ltr_monty_dbl_add_always() 1413 ret_ops |= prj_pt_dbl(&T[rbit], &T[rbit]); in _prj_pt_mul_ltr_monty_dbl_add_always() 1416 ret_ops |= prj_pt_add(&T[1-rbit], &T[rbit], &T[2]); in _prj_pt_mul_ltr_monty_dbl_add_always() 1423 ret = nn_copy(&(T[rbit_next].X.fp_val), &(T[mbit ^ rbit].X.fp_val)); EG(ret, err); in _prj_pt_mul_ltr_monty_dbl_add_always() 1424 ret = nn_copy(&(T[rbit_next].Y.fp_val), &(T[mbit ^ rbit].Y.fp_val)); EG(ret, err); in _prj_pt_mul_ltr_monty_dbl_add_always() 1425 ret = nn_copy(&(T[rbit_next].Z.fp_val), &(T[mbit ^ rbit].Z.fp_val)); EG(ret, err); in _prj_pt_mul_ltr_monty_dbl_add_always() 1428 rbit = rbit_next; in _prj_pt_mul_ltr_monty_dbl_add_always() [all …]
|
| /freebsd/cddl/contrib/opensolaris/lib/libdtrace/common/ |
| H A D | dt_cg.c | 679 uint_t rbit = dnp->dn_child->dn_flags & DT_NF_REF; in dt_cg_prearith_op() local 691 dnp->dn_left->dn_flags |= rbit; in dt_cg_prearith_op() 733 uint_t rbit = dnp->dn_child->dn_flags & DT_NF_REF; in dt_cg_postarith_op() local 748 dnp->dn_left->dn_flags |= rbit; in dt_cg_postarith_op() 1135 uint_t rbit = dnp->dn_left->dn_flags & DT_NF_REF; in dt_cg_asgn_op() local 1147 dnp->dn_left->dn_flags |= rbit; in dt_cg_asgn_op() 1757 uint_t rbit = dnp->dn_child->dn_flags & DT_NF_REF; in dt_cg_node() local 1764 dnp->dn_child->dn_flags |= rbit; in dt_cg_node()
|
| /freebsd/sys/crypto/openssl/aarch64/ |
| H A D | vpsm4_ex-armv8.S | 2722 rbit v16.16b,v16.16b 2847 rbit v16.16b,v16.16b 2848 rbit v17.16b,v17.16b 2849 rbit v18.16b,v18.16b 2850 rbit v19.16b,v19.16b 2856 rbit v20.16b,v20.16b 2857 rbit v21.16b,v21.16b 2858 rbit v22.16b,v22.16b 2859 rbit v23.16b,v23.16b 2941 rbit v16.16b,v16.16b [all …]
|
| H A D | vpsm4-armv8.S | 2923 rbit v8.16b,v8.16b 3008 rbit v0.16b,v0.16b 3009 rbit v1.16b,v1.16b 3010 rbit v2.16b,v2.16b 3011 rbit v3.16b,v3.16b 3017 rbit v12.16b,v12.16b 3018 rbit v13.16b,v13.16b 3019 rbit v14.16b,v14.16b 3020 rbit v15.16b,v15.16b 3202 rbit v8.16b,v8.16b [all …]
|