Rik,

I can't review this series, I forgot almost everything I learned about
this code. The only thing I can recall is that it needs cleanups and
fixes ;) Just a couple of random questions.

On 01/11, r...@redhat.com wrote:
>
> +static inline void switch_fpu_prepare(struct task_struct *old, struct 
> task_struct *new, int cpu)
>  {
> -     fpu_switch_t fpu;
> -
>       /*
>        * If the task has used the math, pre-load the FPU on xsave processors
>        * or if the past 5 consecutive context-switches used math.
>        */
> -     fpu.preload = tsk_used_math(new) && (use_eager_fpu() ||
> +     bool preload = tsk_used_math(new) && (use_eager_fpu() ||
>                                            new->thread.fpu_counter > 5);
>       if (__thread_has_fpu(old)) {
>               if (!__save_init_fpu(old))
> @@ -433,8 +417,9 @@ static inline fpu_switch_t switch_fpu_prepare(struct 
> task_struct *old, struct ta
>               old->thread.fpu.has_fpu = 0;    /* But leave fpu_owner_task! */
>
>               /* Don't change CR0.TS if we just switch! */
> -             if (fpu.preload) {
> +             if (preload) {
>                       new->thread.fpu_counter++;
> +                     set_thread_flag(TIF_LOAD_FPU);
>                       __thread_set_has_fpu(new);
>                       prefetch(new->thread.fpu.state);
>               } else if (!use_eager_fpu())
> @@ -442,16 +427,19 @@ static inline fpu_switch_t switch_fpu_prepare(struct 
> task_struct *old, struct ta
>       } else {
>               old->thread.fpu_counter = 0;
>               old->thread.fpu.last_cpu = ~0;
> -             if (fpu.preload) {
> +             if (preload) {
>                       new->thread.fpu_counter++;
>                       if (!use_eager_fpu() && fpu_lazy_restore(new, cpu))
> -                             fpu.preload = 0;
> -                     else
> +                             /* XXX: is this safe against ptrace??? */

Could you explain your concerns?

> +                             __thread_fpu_begin(new);

this looks strange/unnecessary, there is another unconditonal
__thread_fpu_begin(new) below.

OK, the next patch moves it to switch_fpu_finish(), so perhaps this change
should go into 3/11.

And I am not sure I understand set_thread_flag(TIF_LOAD_FPU). This is called
before __switch_to() updates kernel_stack, so it seems that the old thread
gets this flag set, not new?

Even if this is correct, perhaps set_tsk_thread_flag(new) will look better?

The same for switch_fpu_finish(). I guess this should actually work after
this patch, because switch_fpu_finish() is called before
this_cpu_write(kernel_stack) too and thus both prepare/finish will use the
same thread_info, but this looks confusing at least.

> --- a/arch/x86/include/asm/thread_info.h
> +++ b/arch/x86/include/asm/thread_info.h
> @@ -91,6 +91,7 @@ struct thread_info {
>  #define TIF_SYSCALL_TRACEPOINT       28      /* syscall tracepoint 
> instrumentation */
>  #define TIF_ADDR32           29      /* 32-bit address space on 64 bits */
>  #define TIF_X32                      30      /* 32-bit native x86-64 binary 
> */
> +#define TIF_LOAD_FPU         31      /* load FPU on return to userspace */

Well, the comment is wrong after this patch, but I see 4/11...

>  #define _TIF_SYSCALL_TRACE   (1 << TIF_SYSCALL_TRACE)
>  #define _TIF_NOTIFY_RESUME   (1 << TIF_NOTIFY_RESUME)
> @@ -115,6 +116,7 @@ struct thread_info {
>  #define _TIF_SYSCALL_TRACEPOINT      (1 << TIF_SYSCALL_TRACEPOINT)
>  #define _TIF_ADDR32          (1 << TIF_ADDR32)
>  #define _TIF_X32             (1 << TIF_X32)
> +#define _TIF_LOAD_FPU                (1 << TIF_LOAD_FPU)
>
>  /* work to do in syscall_trace_enter() */
>  #define _TIF_WORK_SYSCALL_ENTRY      \
> @@ -141,7 +143,7 @@ struct thread_info {
>  /* Only used for 64 bit */
>  #define _TIF_DO_NOTIFY_MASK                                          \
>       (_TIF_SIGPENDING | _TIF_MCE_NOTIFY | _TIF_NOTIFY_RESUME |       \
> -      _TIF_USER_RETURN_NOTIFY | _TIF_UPROBE)
> +      _TIF_USER_RETURN_NOTIFY | _TIF_UPROBE | _TIF_LOAD_FPU)

This too. I mean, this change has no effect until 4/11.

Oleg.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to