/* * include/asm-x86_64/i387.h * * Copyright (C) 1994 Linus Torvalds * * Pentium III FXSR, SSE support * General FPU state handling cleanups * Gareth Hughes , May 2000 * x86-64 work by Andi Kleen 2002 */ #ifndef __ASM_X86_64_I387_H #define __ASM_X86_64_I387_H #include #include #include #include #include #include extern void fpu_init(void); extern unsigned int mxcsr_feature_mask; extern void mxcsr_feature_mask_init(void); extern void init_fpu(struct task_struct *child); extern int save_i387(struct _fpstate __user *buf); static inline int need_signal_i387(struct task_struct *me) { if (!me->used_math) return 0; me->used_math = 0; if (me->thread_info->status & TS_USEDFPU) return 0; return 1; } /* * FPU lazy state save handling... */ #define kernel_fpu_end() stts() #define unlazy_fpu(tsk) do { \ if ((tsk)->thread_info->status & TS_USEDFPU) \ save_init_fpu(tsk); \ } while (0) #define clear_fpu(tsk) do { \ if ((tsk)->thread_info->status & TS_USEDFPU) { \ asm volatile("fnclex ; fwait"); \ (tsk)->thread_info->status &= ~TS_USEDFPU; \ stts(); \ } \ } while (0) /* * ptrace request handers... */ extern int get_fpregs(struct user_i387_struct __user *buf, struct task_struct *tsk); extern int set_fpregs(struct task_struct *tsk, struct user_i387_struct __user *buf); /* * i387 state interaction */ #define get_fpu_mxcsr(t) ((t)->thread.i387.fxsave.mxcsr) #define get_fpu_cwd(t) ((t)->thread.i387.fxsave.cwd) #define get_fpu_fxsr_twd(t) ((t)->thread.i387.fxsave.twd) #define get_fpu_swd(t) ((t)->thread.i387.fxsave.swd) #define set_fpu_cwd(t,val) ((t)->thread.i387.fxsave.cwd = (val)) #define set_fpu_swd(t,val) ((t)->thread.i387.fxsave.swd = (val)) #define set_fpu_fxsr_twd(t,val) ((t)->thread.i387.fxsave.twd = (val)) static inline int restore_fpu_checking(struct i387_fxsave_struct *fx) { int err; asm volatile("1: rex64 ; fxrstor (%[fx])\n\t" "2:\n" ".section .fixup,\"ax\"\n" "3: movl $-1,%[err]\n" " jmp 2b\n" ".previous\n" ".section __ex_table,\"a\"\n" " .align 8\n" " .quad 1b,3b\n" ".previous" : [err] "=r" (err) : [fx] "r" (fx), "0" (0)); if (unlikely(err)) init_fpu(current); return err; } static inline int save_i387_checking(struct i387_fxsave_struct __user *fx) { int err; asm volatile("1: rex64 ; fxsave (%[fx])\n\t" "2:\n" ".section .fixup,\"ax\"\n" "3: movl $-1,%[err]\n" " jmp 2b\n" ".previous\n" ".section __ex_table,\"a\"\n" " .align 8\n" " .quad 1b,3b\n" ".previous" : [err] "=r" (err) : [fx] "r" (fx), "0" (0)); if (unlikely(err)) __clear_user(fx, sizeof(struct i387_fxsave_struct)); return err; } static inline void kernel_fpu_begin(void) { struct thread_info *me = current_thread_info(); if (me->status & TS_USEDFPU) { asm volatile("rex64 ; fxsave %0 ; fnclex" : "=m" (me->task->thread.i387.fxsave)); me->status &= ~TS_USEDFPU; return; } clts(); } static inline void save_init_fpu( struct task_struct *tsk ) { asm volatile( "fxsave %0 ; fnclex" : "=m" (tsk->thread.i387.fxsave)); tsk->thread_info->status &= ~TS_USEDFPU; stts(); } /* * This restores directly out of user space. Exceptions are handled. */ static inline int restore_i387(struct _fpstate __user *buf) { return restore_fpu_checking((__force struct i387_fxsave_struct *)buf); } #endif /* __ASM_X86_64_I387_H */