On Wed, 2011-03-02 at 11:20 -0600, Tseng-Hui (Frank) Lin wrote: > +#define CPU_FTR_ICSWX LONG_ASM_CONST(0x1000000000000000)
Do we want a userspace visible feature as well ? Or some other way to inform userspace that we support icswx ? > index acac35d..b0c2549 100644 > --- a/arch/powerpc/include/asm/mmu-hash64.h > +++ b/arch/powerpc/include/asm/mmu-hash64.h > @@ -409,6 +409,14 @@ static inline void subpage_prot_init_new_context(struct > mm_struct *mm) { } > > typedef unsigned long mm_context_id_t; > > +#ifdef CONFIG_ICSWX CONFIG_PPC_ICSWX please. > +/* > + * Use forward declearation to avoid including linux/spinlock_types.h > + * which breaks kernel build. > + */ > +struct spinlock; > +#endif /* CONFIG_ICSWX */ > + Yuck. Instead put all your fields into a structure called something like struct copro_data, make that a fwd declaration and stick a pointer to it in the mm_context. It then only need to be allocated for processes that try to use copros, and the definition of that structure can remain local to whatever header you have dedicated for that. > typedef struct { > mm_context_id_t id; > u16 user_psize; /* page size index */ > @@ -423,6 +431,11 @@ typedef struct { > #ifdef CONFIG_PPC_SUBPAGE_PROT > struct subpage_prot_table spt; > #endif /* CONFIG_PPC_SUBPAGE_PROT */ > +#ifdef CONFIG_ICSWX > + struct spinlock *cop_lockp; /* guard acop and cop_pid */ > + unsigned long acop; /* mask of enabled coprocessor types */ > + unsigned int cop_pid; /* pid value used with coprocessors */ > +#endif /* CONFIG_ICSWX */ > } mm_context_t; > > > diff --git a/arch/powerpc/include/asm/mmu_context.h > b/arch/powerpc/include/asm/mmu_context.h > index 81fb412..fe0a09a 100644 > --- a/arch/powerpc/include/asm/mmu_context.h > +++ b/arch/powerpc/include/asm/mmu_context.h > @@ -32,6 +32,12 @@ extern void __destroy_context(unsigned long context_id); > extern void mmu_context_init(void); > #endif > > +#ifdef CONFIG_ICSWX > +extern void switch_cop(struct mm_struct *next); > +extern int use_cop(unsigned long acop, struct mm_struct *mm); > +extern void drop_cop(unsigned long acop, struct mm_struct *mm); > +#endif /* CONFIG_ICSWX */ No need to ifdef declarations. .../... > + > +#ifdef CONFIG_ICSWX_LAZY_SWITCH > +static DEFINE_PER_CPU(unsigned long, acop_reg); > +#define mtspr_acop(val) \ > + if (__get_cpu_var(acop_reg) != val) { \ > + get_cpu_var(acop_reg) = val; \ > + mtspr(SPRN_ACOP, val); \ > + put_cpu_var(acop_reg); \ > + } Why get/put games here since you did a __get in the first place ? Doesn't make much sense. This is all inside context switch anyways so you just do __ always and don't bother with put. > +#else > +#define mtspr_acop(val) mtspr(SPRN_ACOP, val) > +#endif /* CONFIG_ICSWX_LAZY_SWITCH */ I'm not sure I totally get the point of having an ifdef here. Can't you make it unconditional ? Or do you expect distros to turn that off in which case what's the point ? > +#define COP_PID_NONE 0 > +#define COP_PID_MIN (COP_PID_NONE + 1) > +#define COP_PID_MAX (0xFFFF) > + > +static DEFINE_SPINLOCK(mmu_context_acop_lock); > +static DEFINE_IDA(cop_ida); > + > +void switch_cop(struct mm_struct *next) > +{ > + mtspr(SPRN_PID, next->context.cop_pid); > + mtspr_acop(next->context.acop); > +} s/mtspr_acop/set_cpu_acop() instead. > +static int new_cop_pid(struct ida *ida, int min_id, int max_id, > + spinlock_t *lock) > +{ > + int index; > + int err; > + > +again: > + if (!ida_pre_get(ida, GFP_KERNEL)) > + return -ENOMEM; > + > + spin_lock(lock); > + err = ida_get_new_above(ida, min_id, &index); > + spin_unlock(lock); > + > + if (err == -EAGAIN) > + goto again; > + else if (err) > + return err; > + > + if (index > max_id) { > + spin_lock(lock); > + ida_remove(ida, index); > + spin_unlock(lock); > + return -ENOMEM; > + } > + > + return index; > +} > + > +/* > + * Start using a coprocessor. > + * @acop: mask of coprocessor to be used. > + * @mm: The mm the coprocessor to associate with. Most likely current mm. > + * > + * Return a positive PID if successful. Negative errno otherwise. > + * The returned PID will be fed to the coprocessor to determine if an > + * icswx transaction is authenticated. > + */ > +int use_cop(unsigned long acop, struct mm_struct *mm) > +{ > + int cop_pid; > + > + if (!cpu_has_feature(CPU_FTR_ICSWX)) > + return -ENODEV; > + > + if (!mm || !acop) > + return -EINVAL; > + > + spin_lock(mm->context.cop_lockp); > + if (mm->context.cop_pid == COP_PID_NONE) { new_cop_pid goes GFP_KERNEL allocs no ? It even goes at great length to drop the mmu_context_acop_lock while doing ide_pre_get() ... but you call the whole thing with the mm->context.cop_lockp held. So that's all wrong. You need to drop the lock, allocate a PID, take the lock again, check if somebody came in and gave you a PID already, if yes, release the PID you allocated. Another option is to use a mutex since none of that is in the context switch path right ? Also do you ever call use_cop for a non-current mm ? > + cop_pid = new_cop_pid(&cop_ida, COP_PID_MIN, COP_PID_MAX, > + &mmu_context_acop_lock); > + if (cop_pid < 0) { > + spin_unlock(mm->context.cop_lockp); > + return cop_pid; > + } > + mm->context.cop_pid = cop_pid; > + if (mm == current->active_mm) > + mtspr(SPRN_PID, mm->context.cop_pid); > + } > + mm->context.acop |= acop; > + if (mm == current->active_mm) > + mtspr_acop(mm->context.acop); > + spin_unlock(mm->context.cop_lockp); > + return mm->context.cop_pid; > +} > +EXPORT_SYMBOL_GPL(use_cop); > + > +/* > + * Stop using a coprocessor. > + * @acop: mask of coprocessor to be stopped. > + * @mm: The mm the coprocessor associated with. > + */ > +void drop_cop(unsigned long acop, struct mm_struct *mm) > +{ > + if (WARN_ON(!mm)) > + return; > + > + spin_lock(mm->context.cop_lockp); > + mm->context.acop &= ~acop; > + if (mm == current->active_mm) > + mtspr_acop(mm->context.acop); > + if ((!mm->context.acop) && (mm->context.cop_pid != COP_PID_NONE)) { > + spin_lock(&mmu_context_acop_lock); > + ida_remove(&cop_ida, mm->context.cop_pid); > + spin_unlock(&mmu_context_acop_lock); > + mm->context.cop_pid = COP_PID_NONE; > + if (mm == current->active_mm) > + mtspr(SPRN_PID, mm->context.cop_pid); > + } > + spin_unlock(mm->context.cop_lockp); > +} > +EXPORT_SYMBOL_GPL(drop_cop); > + > +#endif /* CONFIG_ICSWX */ > + > static DEFINE_SPINLOCK(mmu_context_lock); > static DEFINE_IDA(mmu_context_ida); > > @@ -79,6 +245,12 @@ int init_new_context(struct task_struct *tsk, struct > mm_struct *mm) > slice_set_user_psize(mm, mmu_virtual_psize); > subpage_prot_init_new_context(mm); > mm->context.id = index; > +#ifdef CONFIG_ICSWX > + mm->context.cop_lockp = kmalloc(sizeof(spinlock_t), GFP_KERNEL); > + if (! mm->context.cop_lockp) > + return -ENOMEM; > + spin_lock_init(mm->context.cop_lockp); > +#endif /* CONFIG_ICSWX */ > > return 0; > } No, do that on the first time a process tries to use it. No need to add overhead to every task in the system. > @@ -93,6 +265,11 @@ EXPORT_SYMBOL_GPL(__destroy_context); > > void destroy_context(struct mm_struct *mm) > { > +#ifdef CONFIG_ICSWX > + drop_cop(mm->context.acop, mm); > + kfree(mm->context.cop_lockp); > + mm->context.cop_lockp = NULL; > +#endif /* CONFIG_ICSWX */ > __destroy_context(mm->context.id); > subpage_prot_free(mm); > mm->context.id = NO_CONTEXT; > diff --git a/arch/powerpc/platforms/Kconfig.cputype > b/arch/powerpc/platforms/Kconfig.cputype > index 111138c..0007b66 100644 > --- a/arch/powerpc/platforms/Kconfig.cputype > +++ b/arch/powerpc/platforms/Kconfig.cputype > @@ -226,6 +226,49 @@ config VSX > > If in doubt, say Y here. > > +config ICSWX > + bool "Support for PowerPC icswx coprocessor instruction" > + depends on POWER4 > + default n > + ---help--- > + > + Enabling this option to turn on the PowerPC Initiate Coprocessor > + Store Word (icswx) coprocessor instruction support for POWER7 > + or newer processors. > + > + This option is only useful if you have a processor that supports > + icswx coprocessor instruction. It does not have any effect on > + processors without icswx coprocessor instruction. > + > + This option slightly increases kernel memory usage. > + > + Say N if you do not have a PowerPC processor supporting icswx > + instruction and a PowerPC coprocessor. > + > +config ICSWX_LAZY_SWITCH > + bool "Lazy switching coprocessor type registers at context switching" > + depends on ICSWX > + default y > + ---help--- > + > + Coprocessor type register is part of process context. It needs > + to be switched at context switching. > + > + As most machines have a very small number (most likely <= 1) > + of coprocessors, there is a good chance that the value of the > + coprocessor type register is the same between many processes. > + We do not need to change coprocessor type register at context > + switching unless the task to switch to has a different value. > + > + Accessing CPU special purpose registers is far more expensive > + than accessing memory. This option uses a per-CPU variable > + to track the value of coprocessor type register. The variable > + is checked before updating coprocessor type register. The saving > + for one access is small but could turn big with the high > + frequency of context switching. > + > + Say Y unless you have multiple coprocessors. > + > config SPE > bool "SPE Support" > depends on E200 || (E500 && !PPC_E500MC) Ben. _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@lists.ozlabs.org https://lists.ozlabs.org/listinfo/linuxppc-dev