powerpc/slb: Rename all the 'slot' occurrences to 'entry'
authorAnshuman Khandual <khandual@linux.vnet.ibm.com>
Wed, 29 Jul 2015 07:09:59 +0000 (12:39 +0530)
committerMichael Ellerman <mpe@ellerman.id.au>
Wed, 12 Aug 2015 04:50:12 +0000 (14:50 +1000)
The SLB code uses 'slot' and 'entry' interchangeably, change it to always
use 'entry'.

Signed-off-by: Anshuman Khandual <khandual@linux.vnet.ibm.com>
[mpe: Rewrite change log]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
arch/powerpc/mm/slb.c

index 62fafb300c2c62a2b7dac2c24658676552732365..faf9f0c4e823688daddf8a9085fa880e0a6a3c90 100644 (file)
@@ -41,9 +41,9 @@ static void slb_allocate(unsigned long ea)
        (((ssize) == MMU_SEGSIZE_256M)? ESID_MASK: ESID_MASK_1T)
 
 static inline unsigned long mk_esid_data(unsigned long ea, int ssize,
-                                        unsigned long slot)
+                                        unsigned long entry)
 {
-       return (ea & slb_esid_mask(ssize)) | SLB_ESID_V | slot;
+       return (ea & slb_esid_mask(ssize)) | SLB_ESID_V | entry;
 }
 
 static inline unsigned long mk_vsid_data(unsigned long ea, int ssize,
@@ -308,12 +308,11 @@ void slb_initialize(void)
        lflags = SLB_VSID_KERNEL | linear_llp;
        vflags = SLB_VSID_KERNEL | vmalloc_llp;
 
-       /* Invalidate the entire SLB (even slot 0) & all the ERATS */
+       /* Invalidate the entire SLB (even entry 0) & all the ERATS */
        asm volatile("isync":::"memory");
        asm volatile("slbmte  %0,%0"::"r" (0) : "memory");
        asm volatile("isync; slbia; isync":::"memory");
        create_shadowed_slbe(PAGE_OFFSET, mmu_kernel_ssize, lflags, 0);
-
        create_shadowed_slbe(VMALLOC_START, mmu_kernel_ssize, vflags, 1);
 
        /* For the boot cpu, we're running on the stack in init_thread_union,