| /linux/tools/perf/arch/x86/tests/ | 
| H A D | insn-x86-dat-src.c | 21 	asm volatile("rdtsc"); /* Start here */  in main()25 	asm volatile("vcvtph2ps %xmm3,%ymm5");  in main()
 31 	asm volatile("cmovno %rax,%rbx");  in main()
 32 	asm volatile("cmovno 0x12345678(%rax),%rcx");  in main()
 33 	asm volatile("cmovno 0x12345678(%rax),%cx");  in main()
 35 	asm volatile("cmove  %rax,%rbx");  in main()
 36 	asm volatile("cmove 0x12345678(%rax),%rcx");  in main()
 37 	asm volatile("cmove 0x12345678(%rax),%cx");  in main()
 39 	asm volatile("seto    0x12345678(%rax)");  in main()
 40 	asm volatile("setno   0x12345678(%rax)");  in main()
 [all …]
 
 | 
| /linux/lib/raid6/ | 
| H A D | sse2.c | 48 	asm volatile("movdqa %0,%%xmm0" : : "m" (raid6_sse_constants.x1d[0]));  in raid6_sse21_gen_syndrome()49 	asm volatile("pxor %xmm5,%xmm5");	/* Zero temp */  in raid6_sse21_gen_syndrome()
 52 		asm volatile("prefetchnta %0" : : "m" (dptr[z0][d]));  in raid6_sse21_gen_syndrome()
 53 		asm volatile("movdqa %0,%%xmm2" : : "m" (dptr[z0][d])); /* P[0] */  in raid6_sse21_gen_syndrome()
 54 		asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d]));  in raid6_sse21_gen_syndrome()
 55 		asm volatile("movdqa %xmm2,%xmm4"); /* Q[0] */  in raid6_sse21_gen_syndrome()
 56 		asm volatile("movdqa %0,%%xmm6" : : "m" (dptr[z0-1][d]));  in raid6_sse21_gen_syndrome()
 58 			asm volatile("prefetchnta %0" : : "m" (dptr[z][d]));  in raid6_sse21_gen_syndrome()
 59 			asm volatile("pcmpgtb %xmm4,%xmm5");  in raid6_sse21_gen_syndrome()
 60 			asm volatile("paddb %xmm4,%xmm4");  in raid6_sse21_gen_syndrome()
 [all …]
 
 | 
| H A D | avx2.c | 46 	asm volatile("vmovdqa %0,%%ymm0" : : "m" (raid6_avx2_constants.x1d[0]));  in raid6_avx21_gen_syndrome()47 	asm volatile("vpxor %ymm3,%ymm3,%ymm3");	/* Zero temp */  in raid6_avx21_gen_syndrome()
 50 		asm volatile("prefetchnta %0" : : "m" (dptr[z0][d]));  in raid6_avx21_gen_syndrome()
 51 		asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */  in raid6_avx21_gen_syndrome()
 52 		asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d]));  in raid6_avx21_gen_syndrome()
 53 		asm volatile("vmovdqa %ymm2,%ymm4");/* Q[0] */  in raid6_avx21_gen_syndrome()
 54 		asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z0-1][d]));  in raid6_avx21_gen_syndrome()
 56 			asm volatile("prefetchnta %0" : : "m" (dptr[z][d]));  in raid6_avx21_gen_syndrome()
 57 			asm volatile("vpcmpgtb %ymm4,%ymm3,%ymm5");  in raid6_avx21_gen_syndrome()
 58 			asm volatile("vpaddb %ymm4,%ymm4,%ymm4");  in raid6_avx21_gen_syndrome()
 [all …]
 
 | 
| H A D | recov_ssse3.c | 54 	asm volatile("movdqa %0,%%xmm7" : : "m" (x0f[0]));  in raid6_2data_recov_ssse3()57 	asm volatile("movdqa %0,%%xmm6" : : "m" (qmul[0]));  in raid6_2data_recov_ssse3()
 58 	asm volatile("movdqa %0,%%xmm14" : : "m" (pbmul[0]));  in raid6_2data_recov_ssse3()
 59 	asm volatile("movdqa %0,%%xmm15" : : "m" (pbmul[16]));  in raid6_2data_recov_ssse3()
 67 		asm volatile("movdqa %0,%%xmm1" : : "m" (q[0]));  in raid6_2data_recov_ssse3()
 68 		asm volatile("movdqa %0,%%xmm9" : : "m" (q[16]));  in raid6_2data_recov_ssse3()
 69 		asm volatile("movdqa %0,%%xmm0" : : "m" (p[0]));  in raid6_2data_recov_ssse3()
 70 		asm volatile("movdqa %0,%%xmm8" : : "m" (p[16]));  in raid6_2data_recov_ssse3()
 71 		asm volatile("pxor   %0,%%xmm1" : : "m" (dq[0]));  in raid6_2data_recov_ssse3()
 72 		asm volatile("pxor   %0,%%xmm9" : : "m" (dq[16]));  in raid6_2data_recov_ssse3()
 [all …]
 
 | 
| H A D | recov_loongarch_simd.c | 69 	asm volatile("vld $vr20, %0" : : "m" (qmul[0]));  in raid6_2data_recov_lsx()70 	asm volatile("vld $vr21, %0" : : "m" (qmul[16]));  in raid6_2data_recov_lsx()
 71 	asm volatile("vld $vr22, %0" : : "m" (pbmul[0]));  in raid6_2data_recov_lsx()
 72 	asm volatile("vld $vr23, %0" : : "m" (pbmul[16]));  in raid6_2data_recov_lsx()
 76 		asm volatile("vld $vr4, %0" : : "m" (q[0]));  in raid6_2data_recov_lsx()
 77 		asm volatile("vld $vr5, %0" : : "m" (q[16]));  in raid6_2data_recov_lsx()
 78 		asm volatile("vld $vr6, %0" : : "m" (q[32]));  in raid6_2data_recov_lsx()
 79 		asm volatile("vld $vr7, %0" : : "m" (q[48]));  in raid6_2data_recov_lsx()
 81 		asm volatile("vld $vr8, %0" : : "m" (dq[0]));  in raid6_2data_recov_lsx()
 82 		asm volatile("vld $vr9, %0" : : "m" (dq[16]));  in raid6_2data_recov_lsx()
 [all …]
 
 | 
| H A D | recov_avx2.c | 53 	asm volatile("vpbroadcastb %0, %%ymm7" : : "m" (x0f));  in raid6_2data_recov_avx2()57 		asm volatile("vmovdqa %0, %%ymm1" : : "m" (q[0]));  in raid6_2data_recov_avx2()
 58 		asm volatile("vmovdqa %0, %%ymm9" : : "m" (q[32]));  in raid6_2data_recov_avx2()
 59 		asm volatile("vmovdqa %0, %%ymm0" : : "m" (p[0]));  in raid6_2data_recov_avx2()
 60 		asm volatile("vmovdqa %0, %%ymm8" : : "m" (p[32]));  in raid6_2data_recov_avx2()
 61 		asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0]));  in raid6_2data_recov_avx2()
 62 		asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32]));  in raid6_2data_recov_avx2()
 63 		asm volatile("vpxor %0, %%ymm0, %%ymm0" : : "m" (dp[0]));  in raid6_2data_recov_avx2()
 64 		asm volatile("vpxor %0, %%ymm8, %%ymm8" : : "m" (dp[32]));  in raid6_2data_recov_avx2()
 73 		asm volatile("vbroadcasti128 %0, %%ymm4" : : "m" (qmul[0]));  in raid6_2data_recov_avx2()
 [all …]
 
 | 
| H A D | loongarch_simd.c | 54 		asm volatile("vld $vr0, %0" : : "m"(dptr[z0][d+0*NSIZE]));  in raid6_lsx_gen_syndrome()55 		asm volatile("vld $vr1, %0" : : "m"(dptr[z0][d+1*NSIZE]));  in raid6_lsx_gen_syndrome()
 56 		asm volatile("vld $vr2, %0" : : "m"(dptr[z0][d+2*NSIZE]));  in raid6_lsx_gen_syndrome()
 57 		asm volatile("vld $vr3, %0" : : "m"(dptr[z0][d+3*NSIZE]));  in raid6_lsx_gen_syndrome()
 58 		asm volatile("vori.b $vr4, $vr0, 0");  in raid6_lsx_gen_syndrome()
 59 		asm volatile("vori.b $vr5, $vr1, 0");  in raid6_lsx_gen_syndrome()
 60 		asm volatile("vori.b $vr6, $vr2, 0");  in raid6_lsx_gen_syndrome()
 61 		asm volatile("vori.b $vr7, $vr3, 0");  in raid6_lsx_gen_syndrome()
 64 			asm volatile("vld $vr8, %0" : : "m"(dptr[z][d+0*NSIZE]));  in raid6_lsx_gen_syndrome()
 65 			asm volatile("vld $vr9, %0" : : "m"(dptr[z][d+1*NSIZE]));  in raid6_lsx_gen_syndrome()
 [all …]
 
 | 
| H A D | sse1.c | 52 	asm volatile("movq %0,%%mm0" : : "m" (raid6_mmx_constants.x1d));  in raid6_sse11_gen_syndrome()53 	asm volatile("pxor %mm5,%mm5");	/* Zero temp */  in raid6_sse11_gen_syndrome()
 56 		asm volatile("prefetchnta %0" : : "m" (dptr[z0][d]));  in raid6_sse11_gen_syndrome()
 57 		asm volatile("movq %0,%%mm2" : : "m" (dptr[z0][d])); /* P[0] */  in raid6_sse11_gen_syndrome()
 58 		asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d]));  in raid6_sse11_gen_syndrome()
 59 		asm volatile("movq %mm2,%mm4");	/* Q[0] */  in raid6_sse11_gen_syndrome()
 60 		asm volatile("movq %0,%%mm6" : : "m" (dptr[z0-1][d]));  in raid6_sse11_gen_syndrome()
 62 			asm volatile("prefetchnta %0" : : "m" (dptr[z][d]));  in raid6_sse11_gen_syndrome()
 63 			asm volatile("pcmpgtb %mm4,%mm5");  in raid6_sse11_gen_syndrome()
 64 			asm volatile("paddb %mm4,%mm4");  in raid6_sse11_gen_syndrome()
 [all …]
 
 | 
| H A D | mmx.c | 47 	asm volatile("movq %0,%%mm0" : : "m" (raid6_mmx_constants.x1d));  in raid6_mmx1_gen_syndrome()48 	asm volatile("pxor %mm5,%mm5");	/* Zero temp */  in raid6_mmx1_gen_syndrome()
 51 		asm volatile("movq %0,%%mm2" : : "m" (dptr[z0][d])); /* P[0] */  in raid6_mmx1_gen_syndrome()
 52 		asm volatile("movq %mm2,%mm4");	/* Q[0] */  in raid6_mmx1_gen_syndrome()
 54 			asm volatile("movq %0,%%mm6" : : "m" (dptr[z][d]));  in raid6_mmx1_gen_syndrome()
 55 			asm volatile("pcmpgtb %mm4,%mm5");  in raid6_mmx1_gen_syndrome()
 56 			asm volatile("paddb %mm4,%mm4");  in raid6_mmx1_gen_syndrome()
 57 			asm volatile("pand %mm0,%mm5");  in raid6_mmx1_gen_syndrome()
 58 			asm volatile("pxor %mm5,%mm4");  in raid6_mmx1_gen_syndrome()
 59 			asm volatile("pxor %mm5,%mm5");  in raid6_mmx1_gen_syndrome()
 [all …]
 
 | 
| /linux/arch/x86/kvm/ | 
| H A D | fpu.h | 6 #include <asm/fpu/api.h>21 	case 0: asm("movdqa %%xmm0, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 22 	case 1: asm("movdqa %%xmm1, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 23 	case 2: asm("movdqa %%xmm2, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 24 	case 3: asm("movdqa %%xmm3, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 25 	case 4: asm("movdqa %%xmm4, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 26 	case 5: asm("movdqa %%xmm5, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 27 	case 6: asm("movdqa %%xmm6, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 28 	case 7: asm("movdqa %%xmm7, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 30 	case 8: asm("movdqa %%xmm8, %0" : "=m"(*data)); break;  in _kvm_read_sse_reg()
 [all …]
 
 | 
| /linux/arch/arm64/include/asm/vdso/ | 
| H A D | compat_gettimeofday.h | 10 #include <asm/barrier.h>11 #include <asm/unistd_compat_32.h>
 12 #include <asm/errno.h>
 14 #include <asm/vdso/compat_barrier.h>
 24 	register struct timezone *tz asm("r1") = _tz;  in gettimeofday_fallback()
 25 	register struct __kernel_old_timeval *tv asm("r0") = _tv;  in gettimeofday_fallback()
 26 	register long ret asm ("r0");  in gettimeofday_fallback()
 27 	register long nr asm("r7") = __NR_compat32_gettimeofday;  in gettimeofday_fallback()
 29 	asm volatile(  in gettimeofday_fallback()
 41 	register struct __kernel_timespec *ts asm("r1") = _ts;  in clock_gettime_fallback()
 [all …]
 
 | 
| H A D | gettimeofday.h | 12 #include <asm/alternative.h>13 #include <asm/arch_timer.h>
 14 #include <asm/barrier.h>
 15 #include <asm/unistd.h>
 16 #include <asm/sysreg.h>
 24 	register struct timezone *tz asm("x1") = _tz;  in gettimeofday_fallback()
 25 	register struct __kernel_old_timeval *tv asm("x0") = _tv;  in gettimeofday_fallback()
 26 	register long ret asm ("x0");  in gettimeofday_fallback()
 27 	register long nr asm("x8") = __NR_gettimeofday;  in gettimeofday_fallback()
 29 	asm volatile(  in gettimeofday_fallback()
 [all …]
 
 | 
| /linux/tools/testing/selftests/arm64/abi/ | 
| H A D | hwcap.c | 16 #include <asm/hwcap.h>17 #include <asm/sigcontext.h>
 18 #include <asm/unistd.h>
 46 	asm volatile(".inst 0x4e284800" : : : );  in aes_sigill()
 52 	asm volatile(".inst 0xb82003ff" : : : );  in atomics_sigill()
 63 	asm volatile(".inst 0x1ac14800" : : : );  in crc32_sigill()
 69 	asm volatile(".inst 0xdac01c00" : : : "x0");  in cssc_sigill()
 75 	asm volatile(".inst 0x2ec03c00");  in f8cvt_sigill()
 81 	asm volatile(".inst 0xe40fc00");  in f8dp2_sigill()
 87 	asm volatile(".inst 0xe00fc00");  in f8dp4_sigill()
 [all …]
 
 | 
| /linux/arch/mips/include/asm/vdso/ | 
| H A D | gettimeofday.h | 16 #include <asm/vdso/vdso.h>17 #include <asm/clocksource.h>
 18 #include <asm/unistd.h>
 19 #include <asm/vdso.h>
 33 	register struct timezone *tz asm("a1") = _tz;  in gettimeofday_fallback()
 34 	register struct __kernel_old_timeval *tv asm("a0") = _tv;  in gettimeofday_fallback()
 35 	register long ret asm("v0");  in gettimeofday_fallback()
 36 	register long nr asm("v0") = __NR_gettimeofday;  in gettimeofday_fallback()
 37 	register long error asm("a3");  in gettimeofday_fallback()
 39 	asm volatile(  in gettimeofday_fallback()
 [all …]
 
 | 
| /linux/tools/perf/ | 
| H A D | check-headers.sh | 36   "arch/x86/include/asm/cpufeatures.h"37   "arch/x86/include/asm/inat_types.h"
 38   "arch/x86/include/asm/emulate_prefix.h"
 39   "arch/x86/include/asm/msr-index.h"
 42   "arch/arm/include/uapi/asm/perf_regs.h"
 43   "arch/arm64/include/uapi/asm/perf_regs.h"
 44   "arch/loongarch/include/uapi/asm/perf_regs.h"
 45   "arch/mips/include/uapi/asm/perf_regs.h"
 46   "arch/powerpc/include/uapi/asm/perf_regs.h"
 47   "arch/s390/include/uapi/asm/perf_regs.h"
 [all …]
 
 | 
| /linux/drivers/gpu/drm/vmwgfx/ | 
| H A D | vmwgfx_msg_arm64.h | 53 	register u64 x0 asm("x0") = VMWARE_HYPERVISOR_MAGIC;  in vmware_hypercall1()54 	register u64 x1 asm("x1") = in1;  in vmware_hypercall1()
 55 	register u64 x2 asm("x2") = cmd;  in vmware_hypercall1()
 56 	register u64 x3 asm("x3") = VMWARE_HYPERVISOR_PORT;  in vmware_hypercall1()
 57 	register u64 x7 asm("x7") = ((u64)X86_IO_MAGIC << 32) |  in vmware_hypercall1()
 76 	register u64 x0 asm("x0") = VMWARE_HYPERVISOR_MAGIC;  in vmware_hypercall5()
 77 	register u64 x1 asm("x1") = in1;  in vmware_hypercall5()
 78 	register u64 x2 asm("x2") = cmd;  in vmware_hypercall5()
 79 	register u64 x3 asm("x3") = in3 | VMWARE_HYPERVISOR_PORT;  in vmware_hypercall5()
 80 	register u64 x4 asm("x4") = in4;  in vmware_hypercall5()
 [all …]
 
 | 
| /linux/arch/riscv/include/asm/vdso/ | 
| H A D | gettimeofday.h | 7 #include <asm/barrier.h>8 #include <asm/unistd.h>
 9 #include <asm/csr.h>
 24 	register struct __kernel_old_timeval *tv asm("a0") = _tv;  in gettimeofday_fallback()
 25 	register struct timezone *tz asm("a1") = _tz;  in gettimeofday_fallback()
 26 	register long ret asm("a0");  in gettimeofday_fallback()
 27 	register long nr asm("a7") = __NR_gettimeofday;  in gettimeofday_fallback()
 29 	asm volatile ("ecall\n"  in gettimeofday_fallback()
 40 	register clockid_t clkid asm("a0") = _clkid;  in clock_gettime_fallback()
 41 	register struct __kernel_timespec *ts asm("a1") = _ts;  in clock_gettime_fallback()
 [all …]
 
 | 
| /linux/arch/m68k/kernel/ | 
| H A D | vectors.c | 27 #include <asm/setup.h>28 #include <asm/fpu.h>
 29 #include <asm/traps.h>
 45 asm(".text\n"
 67 		asmlinkage void unimp_vec(void) asm ("_060_isp_unimp");  in base_trap_init()
 98 		asmlinkage void dz_vec(void) asm ("dz");  in trap_init()
 99 		asmlinkage void inex_vec(void) asm ("inex");  in trap_init()
 100 		asmlinkage void ovfl_vec(void) asm ("ovfl");  in trap_init()
 101 		asmlinkage void unfl_vec(void) asm ("unfl");  in trap_init()
 102 		asmlinkage void snan_vec(void) asm ("snan");  in trap_init()
 [all …]
 
 | 
| /linux/arch/arm64/include/asm/ | 
| H A D | barrier.h | 3  * Based on arch/arm/include/asm/barrier.h14 #include <asm/alternative-macros.h>
 17 #define nops(n)		asm volatile(__nops(n))
 19 #define sev()		asm volatile("sev" : : : "memory")
 20 #define wfe()		asm volatile("wfe" : : : "memory")
 21 #define wfet(val)	asm volatile("msr s0_3_c1_c0_0, %0"	\
 23 #define wfi()		asm volatile("wfi" : : : "memory")
 24 #define wfit(val)	asm volatile("msr s0_3_c1_c0_1, %0"	\
 27 #define isb()		asm volatile("isb" : : : "memory")
 28 #define dmb(opt)	asm volatile("dmb " #opt : : : "memory")
 [all …]
 
 | 
| /linux/arch/sh/include/asm/ | 
| H A D | bitops.h | 10 #include <asm/byteorder.h>11 #include <asm/barrier.h>
 14 #include <asm/bitops-grb.h>
 16 #include <asm-generic/bitops/atomic.h>
 17 #include <asm/bitops-op32.h>
 19 #include <asm/bitops-llsc.h>
 21 #include <asm/bitops-cas.h>
 23 #include <asm-generic/bitops/atomic.h>
 24 #include <asm-generic/bitops/non-atomic.h>
 61 #include <asm-generic/bitops/ffs.h>
 [all …]
 
 | 
| /linux/arch/csky/include/asm/ | 
| H A D | bitops.h | 7 #include <asm/barrier.h>10  * asm-generic/bitops/ffs.h
 17 	asm volatile (  in ffs()
 27  * asm-generic/bitops/__ffs.h
 31 	asm volatile (  in __ffs()
 40  * asm-generic/bitops/fls.h
 44 	asm volatile(  in fls()
 53  * asm-generic/bitops/__fls.h
 60 #include <asm-generic/bitops/ffz.h>
 61 #include <asm-generic/bitops/fls64.h>
 [all …]
 
 | 
| H A D | barrier.h | 8 #define nop()	asm volatile ("nop\n":::"memory")44 #define __bar_brw()	asm volatile (".long 0x842cc000\n":::"memory")
 45 #define __bar_br()	asm volatile (".long 0x8424c000\n":::"memory")
 46 #define __bar_bw()	asm volatile (".long 0x8428c000\n":::"memory")
 47 #define __bar_arw()	asm volatile (".long 0x8423c000\n":::"memory")
 48 #define __bar_ar()	asm volatile (".long 0x8421c000\n":::"memory")
 49 #define __bar_aw()	asm volatile (".long 0x8422c000\n":::"memory")
 50 #define __bar_brwarw()	asm volatile (FULL_FENCE:::"memory")
 51 #define __bar_brarw()	asm volatile (ACQUIRE_FENCE:::"memory")
 52 #define __bar_bwarw()	asm volatile (".long 0x842bc000\n":::"memory")
 [all …]
 
 | 
| /linux/tools/testing/selftests/bpf/progs/ | 
| H A D | verifier_sdiv.c | 19 	asm volatile ("					\  in sdiv32_non_zero_imm_1()31 	asm volatile ("					\  in sdiv32_non_zero_imm_2()
 43 	asm volatile ("					\  in sdiv32_non_zero_imm_3()
 55 	asm volatile ("					\  in sdiv32_non_zero_imm_4()
 67 	asm volatile ("					\  in sdiv32_non_zero_imm_5()
 79 	asm volatile ("					\  in sdiv32_non_zero_imm_6()
 91 	asm volatile ("					\  in sdiv32_non_zero_imm_7()
 103 	asm volatile ("					\  in sdiv32_non_zero_imm_8()
 115 	asm volatile ("					\  in sdiv32_non_zero_reg_1()
 128 	asm volatile ("					\  in sdiv32_non_zero_reg_2()
 [all …]
 
 | 
| /linux/arch/loongarch/include/asm/vdso/ | 
| H A D | gettimeofday.h | 12 #include <asm/unistd.h>13 #include <asm/vdso/vdso.h>
 21 	register struct __kernel_old_timeval *tv asm("a0") = _tv;  in gettimeofday_fallback()
 22 	register struct timezone *tz asm("a1") = _tz;  in gettimeofday_fallback()
 23 	register long nr asm("a7") = __NR_gettimeofday;  in gettimeofday_fallback()
 24 	register long ret asm("a0");  in gettimeofday_fallback()
 26 	asm volatile(  in gettimeofday_fallback()
 40 	register clockid_t clkid asm("a0") = _clkid;  in clock_gettime_fallback()
 41 	register struct __kernel_timespec *ts asm("a1") = _ts;  in clock_gettime_fallback()
 42 	register long nr asm("a7") = __NR_clock_gettime;  in clock_gettime_fallback()
 [all …]
 
 | 
| /linux/arch/loongarch/include/asm/ | 
| H A D | bitops.h | 14 #include <asm/barrier.h>16 #include <asm-generic/bitops/builtin-ffs.h>
 17 #include <asm-generic/bitops/builtin-fls.h>
 18 #include <asm-generic/bitops/builtin-__ffs.h>
 19 #include <asm-generic/bitops/builtin-__fls.h>
 21 #include <asm-generic/bitops/ffz.h>
 22 #include <asm-generic/bitops/fls64.h>
 24 #include <asm-generic/bitops/sched.h>
 25 #include <asm-generic/bitops/hweight.h>
 27 #include <asm-generic/bitops/atomic.h>
 [all …]
 
 |