Lines Matching +full:0 +full:x66

18 	asm volatile("mov %%cr0,%0" : "=r" (val));  in native_read_cr0()
25 asm volatile("mov %%cr2,%0" : "=r" (val)); in native_read_cr2()
31 asm volatile("mov %0,%%cr2": : "r" (val) : "memory"); in native_write_cr2()
37 asm volatile("mov %%cr3,%0" : "=r" (val)); in __native_read_cr3()
43 asm volatile("mov %0,%%cr3": : "r" (val) : "memory"); in native_write_cr3()
52 * is functionally equivalent to CR4 == 0. Keep it simple and pretend in native_read_cr4()
53 * that CR4 == 0 on CPUs that don't have CR4. in native_read_cr4()
55 asm volatile("1: mov %%cr4, %0\n" in native_read_cr4()
58 : "=r" (val) : "0" (0)); in native_read_cr4()
61 asm volatile("mov %%cr4,%0" : "=r" (val)); in native_read_cr4()
71 u32 ecx = 0; in rdpkru()
76 * clears EDX and requires that ecx=0. in rdpkru()
78 asm volatile(".byte 0x0f,0x01,0xee\n\t" in rdpkru()
86 u32 ecx = 0, edx = 0; in wrpkru()
90 * requires that ecx = edx = 0. in wrpkru()
92 asm volatile(".byte 0x0f,0x01,0xef\n\t" in wrpkru()
99 return 0; in rdpkru()
120 #define ASM_WBNOINVD _ASM_BYTES(0xf3,0x0f,0x09)
190 asm volatile("clflush %0" : "+m" (*(volatile char __force *)__p)); in clflush()
195 alternative_io("ds clflush %0", in clflushopt()
196 "clflushopt %0", X86_FEATURE_CLFLUSHOPT, in clflushopt()
205 "ds clflush %0", in clwb()
206 "clflushopt %0", X86_FEATURE_CLFLUSHOPT, in clwb()
207 "clwb %0", X86_FEATURE_CLWB) in clwb()
218 return 0; in write_user_shstk_64()
229 asm volatile(".byte 0xf, 0x1, 0xe8" ::: "memory"); in serialize()
249 asm volatile(".byte 0x66, 0x0f, 0x38, 0xf8, 0x02" in movdir64b()
267 * ZF = 0 equates to success, and ZF = 1 indicates retry or error.
272 * returns 0 on success and -EAGAIN on failure.
289 asm volatile(".byte 0xf3, 0x0f, 0x38, 0xf8, 0x02, 0x66, 0x90" in enqcmds()
298 return 0; in enqcmds()
307 asm volatile(".byte 0xc4, 0xe2, 0x78, 0x49, 0xc0"); in tile_release()