On 07/08/2016 07:50 AM, Michael Neuling wrote:
> 
>> diff --git a/arch/powerpc/include/asm/cpuidle.h 
>> b/arch/powerpc/include/asm/cpuidle.h
>> index d2f99ca..3d7fc06 100644
>> --- a/arch/powerpc/include/asm/cpuidle.h
>> +++ b/arch/powerpc/include/asm/cpuidle.h
>> @@ -13,6 +13,8 @@
>>  #ifndef __ASSEMBLY__
>>  extern u32 pnv_fastsleep_workaround_at_entry[];
>>  extern u32 pnv_fastsleep_workaround_at_exit[];
>> +
>> +extern u64 pnv_first_deep_stop_state;
> 
> mpe asked a question about this which you neither answered or addressed.
> "Should this have some safe initial value?"
> 
> I'm thinking we could do this which is what you have in the init call.
>    u64 pnv_first_deep_stop_state = MAX_STOP_STATE;
> 

I missed the comment. I'll make the change.
> 
>> @@ -439,7 +540,18 @@ timebase_resync:
>>       */
>>      bne     cr4,clear_lock
>>  
>> -    /* Restore per core state */
>> +    /*
>> +     * First thread in the core to wake up and its waking up with
>> +     * complete hypervisor state loss. Restore per core hypervisor
>> +     * state.
>> +     */
>> +BEGIN_FTR_SECTION
>> +    ld      r4,_PTCR(r1)
>> +    mtspr   SPRN_PTCR,r4
>> +    ld      r4,_RPR(r1)
>> +    mtspr   SPRN_RPR,r4
> 
> RPR looks wrong here.  This should be on POWER8 too.
> 
> This has changed since v6 and not noted in the v7 comments.  Why are you
> changing this now?
> 
RPR is a per-core resource in P9. So with this patch, RPR will continue
to be restored per-subcore in P8 and will restored once per core in P9.

>> +END_FTR_SECTION_IFSET(CPU_FTR_ARCH_300)
>> +
>>      ld      r4,_TSCR(r1)
>>      mtspr   SPRN_TSCR,r4
>>      ld      r4,_WORC(r1)
>> @@ -461,9 +573,7 @@ common_exit:
>>  
>>      /* Waking up from winkle */
>>  
>> -    /* Restore per thread state */
>> -    bl      __restore_cpu_power8
>> -
>> +BEGIN_MMU_FTR_SECTION
>>      /* Restore SLB  from PACA */
>>      ld      r8,PACA_SLBSHADOWPTR(r13)
>>  
>> @@ -477,6 +587,9 @@ common_exit:
>>      slbmte  r6,r5
>>  1:  addi    r8,r8,16
>>      .endr
>> +END_MMU_FTR_SECTION_IFCLR(MMU_FTR_RADIX)
>> +
>> +    /* Restore per thread state */
> 
> This FTR section is too big  It ends up at 25 instructions with the loop.
> Probably better like this:
> 
> BEGIN_MMU_FTR_SECTION
>       b       no_segments
> END_MMU_FTR_SECTION_IFSET(MMU_FTR_RADIX)
>       /* Restore SLB  from PACA */
>       ld      r8,PACA_SLBSHADOWPTR(r13)
> 
>       .rept   SLB_NUM_BOLTED
>       li      r3, SLBSHADOW_SAVEAREA
>       LDX_BE  r5, r8, r3
>       addi    r3, r3, 8
>       LDX_BE  r6, r8, r3
>       andis.  r7,r5,SLB_ESID_V@h
>       beq     1f
>       slbmte  r6,r5
> 1:    addi    r8,r8,16
>       .endr
> 
> no_segments:
> 
Cool. Will make the change.

Thanks,
Shreyas

Reply via email to