}
extern void flush_stab(struct task_struct *tsk, struct mm_struct *mm);
-extern void switch_slb(struct task_struct *tsk, struct mm_struct *mm);
+extern void flush_slb(struct task_struct *tsk, struct mm_struct *mm);
/*
* switch_mm is the entry point called from the architecture independent
return;
if (cur_cpu_spec->cpu_features & CPU_FTR_SLB)
- switch_slb(tsk, next);
+ flush_slb(tsk, next);
else
flush_stab(tsk, next);
}
local_irq_restore(flags);
}
+#define VSID_RANDOMIZER 42470972311UL
+#define VSID_MASK 0xfffffffffUL
+
+
/* This is only valid for kernel (including vmalloc, imalloc and bolted) EA's
*/
static inline unsigned long