X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=include%2Fasm-ppc64%2Fmmu_context.h;h=505ad13b358ef05562b56aa380dfc8c9ed1ef495;hb=5e3b93f248c98873cc843e83092bb8da92ac85a2;hp=03561b5895d73f51a8fe8016daf6e22b50268536;hpb=a91482bdcc2e0f6035702e46f1b99043a0893346;p=linux-2.6.git diff --git a/include/asm-ppc64/mmu_context.h b/include/asm-ppc64/mmu_context.h index 03561b589..505ad13b3 100644 --- a/include/asm-ppc64/mmu_context.h +++ b/include/asm-ppc64/mmu_context.h @@ -136,7 +136,7 @@ destroy_context(struct mm_struct *mm) } extern void flush_stab(struct task_struct *tsk, struct mm_struct *mm); -extern void switch_slb(struct task_struct *tsk, struct mm_struct *mm); +extern void flush_slb(struct task_struct *tsk, struct mm_struct *mm); /* * switch_mm is the entry point called from the architecture independent @@ -161,7 +161,7 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, return; if (cur_cpu_spec->cpu_features & CPU_FTR_SLB) - switch_slb(tsk, next); + flush_slb(tsk, next); else flush_stab(tsk, next); } @@ -181,6 +181,10 @@ static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) local_irq_restore(flags); } +#define VSID_RANDOMIZER 42470972311UL +#define VSID_MASK 0xfffffffffUL + + /* This is only valid for kernel (including vmalloc, imalloc and bolted) EA's */ static inline unsigned long