/* * PowerPC64 port by Mike Corrigan and Dave Engebretsen * {mikejc|engebret}@us.ibm.com * * Copyright (c) 2000 Mike Corrigan * * SMP scalability work: * Copyright (C) 2001 Anton Blanchard , IBM * * Module name: htab.c * * Description: * PowerPC Hashed Page Table functions * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation; either version * 2 of the License, or (at your option) any later version. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* * Note: pte --> Linux PTE * HPTE --> PowerPC Hashed Page Table Entry * * Execution context: * htab_initialize is called with the MMU off (of course), but * the kernel has been copied down to zero so it can directly * reference global data. At this point it is very difficult * to print debug info. * */ HTAB htab_data = {NULL, 0, 0, 0, 0}; extern unsigned long _SDR1; extern unsigned long klimit; void make_pte(HPTE *htab, unsigned long va, unsigned long pa, int mode, unsigned long hash_mask, int large); long plpar_pte_enter(unsigned long flags, unsigned long ptex, unsigned long new_pteh, unsigned long new_ptel, unsigned long *old_pteh_ret, unsigned long *old_ptel_ret); static long hpte_remove(unsigned long hpte_group); static long rpa_lpar_hpte_remove(unsigned long hpte_group); static long iSeries_hpte_remove(unsigned long hpte_group); inline unsigned long get_lock_slot(unsigned long vpn); static spinlock_t pSeries_tlbie_lock = SPIN_LOCK_UNLOCKED; static spinlock_t pSeries_lpar_tlbie_lock = SPIN_LOCK_UNLOCKED; #define LOCK_SPLIT #ifdef LOCK_SPLIT hash_table_lock_t hash_table_lock[128] __cacheline_aligned_in_smp = { [0 ... 31] = {SPIN_LOCK_UNLOCKED}}; #else hash_table_lock_t hash_table_lock[1] __cacheline_aligned_in_smp = { [0] = {SPIN_LOCK_UNLOCKED}}; #endif #define KB (1024) #define MB (1024*KB) static inline void loop_forever(void) { volatile unsigned long x = 1; for(;x;x|=1) ; } static inline void create_pte_mapping(unsigned long start, unsigned long end, unsigned long mode, unsigned long mask, int large) { unsigned long addr; HPTE *htab = (HPTE *)__v2a(htab_data.htab); unsigned int step; if (large) step = 16*MB; else step = 4*KB; for (addr = start; addr < end; addr += step) { unsigned long vsid = get_kernel_vsid(addr); unsigned long va = (vsid << 28) | (addr & 0xfffffff); make_pte(htab, va, (unsigned long)__v2a(addr), mode, mask, large); } } void htab_initialize(void) { unsigned long table, htab_size_bytes; unsigned long pteg_count; unsigned long mode_rw, mask, lock_shift; #if 0 /* Can't really do the call below since it calls the normal RTAS * entry point and we're still relocate off at the moment. * Temporarily diabling until it can call through the relocate off * RTAS entry point. -Peter */ ppc64_boot_msg(0x05, "htab init"); #endif /* * Calculate the required size of the htab. We want the number of * PTEGs to equal one half the number of real pages. */ htab_size_bytes = 1UL << naca->pftSize; pteg_count = htab_size_bytes >> 7; /* For debug, make the HTAB 1/8 as big as it normally would be. */ ifppcdebug(PPCDBG_HTABSIZE) { pteg_count >>= 3; htab_size_bytes = pteg_count << 7; } htab_data.htab_num_ptegs = pteg_count; htab_data.htab_hash_mask = pteg_count - 1; /* * Calculate the number of bits to shift the pteg selector such that we * use the high order 8 bits to select a page table lock. */ asm ("cntlzd %0,%1" : "=r" (lock_shift) : "r" (htab_data.htab_hash_mask)); htab_data.htab_lock_shift = (64 - lock_shift) - 8; if(systemcfg->platform == PLATFORM_PSERIES) { /* Find storage for the HPT. Must be contiguous in * the absolute address space. */ table = lmb_alloc(htab_size_bytes, htab_size_bytes); if ( !table ) { ppc64_terminate_msg(0x20, "hpt space"); loop_forever(); } htab_data.htab = (HPTE *)__a2v(table); /* htab absolute addr + encoded htabsize */ _SDR1 = table + __ilog2(pteg_count) - 11; /* Initialize the HPT with no entries */ memset((void *)table, 0, htab_size_bytes); } else { /* Using a hypervisor which owns the htab */ htab_data.htab = NULL; _SDR1 = 0; } mode_rw = _PAGE_ACCESSED | _PAGE_COHERENT | PP_RWXX; mask = pteg_count-1; /* XXX we currently map kernel text rw, should fix this */ if ((systemcfg->platform & PLATFORM_PSERIES) && (cur_cpu_spec->cpu_features & CPU_FTR_16M_PAGE) && (systemcfg->physicalMemorySize > 256*MB)) { create_pte_mapping((unsigned long)KERNELBASE, KERNELBASE + 256*MB, mode_rw, mask, 0); create_pte_mapping((unsigned long)KERNELBASE + 256*MB, KERNELBASE + (systemcfg->physicalMemorySize), mode_rw, mask, 1); } else { create_pte_mapping((unsigned long)KERNELBASE, KERNELBASE+(systemcfg->physicalMemorySize), mode_rw, mask, 0); } #if 0 /* Can't really do the call below since it calls the normal RTAS * entry point and we're still relocate off at the moment. * Temporarily diabling until it can call through the relocate off * RTAS entry point. -Peter */ ppc64_boot_msg(0x06, "htab done"); #endif } #undef KB #undef MB /* * Create a pte. Used during initialization only. * We assume the PTE will fit in the primary PTEG. */ void make_pte(HPTE *htab, unsigned long va, unsigned long pa, int mode, unsigned long hash_mask, int large) { HPTE *hptep, local_hpte, rhpte; unsigned long hash, vpn, flags, lpar_rc; unsigned long i, dummy1, dummy2; long slot; if (large) vpn = va >> LARGE_PAGE_SHIFT; else vpn = va >> PAGE_SHIFT; hash = hpt_hash(vpn, large); local_hpte.dw1.dword1 = pa | mode; local_hpte.dw0.dword0 = 0; local_hpte.dw0.dw0.avpn = va >> 23; local_hpte.dw0.dw0.bolted = 1; /* bolted */ if (large) { local_hpte.dw0.dw0.l = 1; /* large page */ local_hpte.dw0.dw0.avpn &= ~0x1UL; } local_hpte.dw0.dw0.v = 1; if (systemcfg->platform == PLATFORM_PSERIES) { hptep = htab + ((hash & hash_mask)*HPTES_PER_GROUP); for (i = 0; i < 8; ++i, ++hptep) { if (hptep->dw0.dw0.v == 0) { /* !valid */ *hptep = local_hpte; return; } } } else if (systemcfg->platform == PLATFORM_PSERIES_LPAR) { slot = ((hash & hash_mask)*HPTES_PER_GROUP); /* Set CEC cookie to 0 */ /* Zero page = 0 */ /* I-cache Invalidate = 0 */ /* I-cache synchronize = 0 */ /* Exact = 0 - modify any entry in group */ flags = 0; lpar_rc = plpar_pte_enter(flags, slot, local_hpte.dw0.dword0, local_hpte.dw1.dword1, &dummy1, &dummy2); return; } else if (systemcfg->platform == PLATFORM_ISERIES_LPAR) { slot = HvCallHpt_findValid(&rhpte, vpn); if (slot < 0) { /* Must find space in primary group */ panic("hash_page: hpte already exists\n"); } HvCallHpt_addValidate(slot, 0, (HPTE *)&local_hpte ); return; } /* We should _never_ get here and too early to call xmon. */ ppc64_terminate_msg(0x22, "hpte platform"); loop_forever(); } /* * find_linux_pte returns the address of a linux pte for a given * effective address and directory. If not found, it returns zero. */ pte_t *find_linux_pte(pgd_t *pgdir, unsigned long ea) { pgd_t *pg; pmd_t *pm; pte_t *pt = NULL; pte_t pte; pg = pgdir + pgd_index(ea); if (!pgd_none(*pg)) { pm = pmd_offset(pg, ea); if (!pmd_none(*pm)) { pt = pte_offset(pm, ea); pte = *pt; if (!pte_present(pte)) pt = NULL; } } return pt; } static inline unsigned long computeHptePP(unsigned long pte) { return (pte & _PAGE_USER) | (((pte & _PAGE_USER) >> 1) & ((~((pte >> 2) & /* _PAGE_RW */ (pte >> 7))) & /* _PAGE_DIRTY */ 1)); } /* * Handle a fault by adding an HPTE. If the address can't be determined * to be valid via Linux page tables, return 1. If handled return 0 */ int __hash_page(unsigned long ea, unsigned long access, unsigned long vsid, pte_t *ptep) { unsigned long va, vpn; unsigned long newpp, prpn; unsigned long hpteflags, lock_slot; long slot; pte_t old_pte, new_pte; /* Search the Linux page table for a match with va */ va = (vsid << 28) | (ea & 0x0fffffff); vpn = va >> PAGE_SHIFT; lock_slot = get_lock_slot(vpn); /* Acquire the hash table lock to guarantee that the linux * pte we fetch will not change */ spin_lock(&hash_table_lock[lock_slot].lock); /* * Check the user's access rights to the page. If access should be * prevented then send the problem up to do_page_fault. */ #ifdef CONFIG_SHARED_MEMORY_ADDRESSING access |= _PAGE_PRESENT; if (unlikely(access & ~(pte_val(*ptep)))) { if(!(((ea >> SMALLOC_EA_SHIFT) == (SMALLOC_START >> SMALLOC_EA_SHIFT)) && ((current->thread.flags) & PPC_FLAG_SHARED))) { spin_unlock(&hash_table_lock[lock_slot].lock); return 1; } } #else access |= _PAGE_PRESENT; if (unlikely(access & ~(pte_val(*ptep)))) { spin_unlock(&hash_table_lock[lock_slot].lock); return 1; } #endif /* * We have found a pte (which was present). * The spinlocks prevent this status from changing * The hash_table_lock prevents the _PAGE_HASHPTE status * from changing (RPN, DIRTY and ACCESSED too) * The page_table_lock prevents the pte from being * invalidated or modified */ /* * At this point, we have a pte (old_pte) which can be used to build * or update an HPTE. There are 2 cases: * * 1. There is a valid (present) pte with no associated HPTE (this is * the most common case) * 2. There is a valid (present) pte with an associated HPTE. The * current values of the pp bits in the HPTE prevent access * because we are doing software DIRTY bit management and the * page is currently not DIRTY. */ old_pte = *ptep; new_pte = old_pte; /* If the attempted access was a store */ if (access & _PAGE_RW) pte_val(new_pte) |= _PAGE_ACCESSED | _PAGE_DIRTY; else pte_val(new_pte) |= _PAGE_ACCESSED; newpp = computeHptePP(pte_val(new_pte)); /* Check if pte already has an hpte (case 2) */ if (unlikely(pte_val(old_pte) & _PAGE_HASHPTE)) { /* There MIGHT be an HPTE for this pte */ unsigned long hash, slot, secondary; /* XXX fix large pte flag */ hash = hpt_hash(vpn, 0); secondary = (pte_val(old_pte) & _PAGE_SECONDARY) >> 15; if (secondary) hash = ~hash; slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP; slot += (pte_val(old_pte) & _PAGE_GROUP_IX) >> 12; /* XXX fix large pte flag */ if (ppc_md.hpte_updatepp(slot, secondary, newpp, va, 0) == -1) { pte_val(old_pte) &= ~_PAGE_HPTEFLAGS; } else { if (!pte_same(old_pte, new_pte)) { *ptep = new_pte; } } } if (likely(!(pte_val(old_pte) & _PAGE_HASHPTE))) { /* Update the linux pte with the HPTE slot */ pte_val(new_pte) &= ~_PAGE_HPTEFLAGS; pte_val(new_pte) |= _PAGE_HASHPTE; prpn = pte_val(old_pte) >> PTE_SHIFT; /* copy appropriate flags from linux pte */ hpteflags = (pte_val(new_pte) & 0x1f8) | newpp; slot = ppc_md.hpte_insert(vpn, prpn, hpteflags, 0, 0); pte_val(new_pte) |= ((slot<<12) & (_PAGE_GROUP_IX | _PAGE_SECONDARY)); *ptep = new_pte; } spin_unlock(&hash_table_lock[lock_slot].lock); return 0; } /* * Handle a fault by adding an HPTE. If the address can't be determined * to be valid via Linux page tables, return 1. If handled return 0 */ int hash_page(unsigned long ea, unsigned long access) { void *pgdir; unsigned long vsid; struct mm_struct *mm; pte_t *ptep; int ret; /* Check for invalid addresses. */ if (!IS_VALID_EA(ea)) return 1; switch (REGION_ID(ea)) { case USER_REGION_ID: mm = current->mm; if (mm == NULL) return 1; vsid = get_vsid(mm->context, ea); break; case IO_REGION_ID: mm = &ioremap_mm; vsid = get_kernel_vsid(ea); break; case VMALLOC_REGION_ID: mm = &init_mm; vsid = get_kernel_vsid(ea); #ifdef CONFIG_SHARED_MEMORY_ADDRESSING /* * Check if this is a user task with shared access to kernel * data & we got a protection fault. If it is, the kernel * must have faulted in the segment and the protection flags * on the segment are kernel access only. Just flush the * segment table & fault in the segment with the right flags. */ if(((current->thread.flags) & PPC_FLAG_SHARED) && (access & _PAGE_USER)) { flush_stab(); } #endif break; case EEH_REGION_ID: /* * Should only be hit if there is an access to MMIO space * which is protected by EEH. * Send the problem up to do_page_fault */ case KERNEL_REGION_ID: /* * Should never get here - entire 0xC0... region is bolted. * Send the problem up to do_page_fault */ default: /* Not a valid range * Send the problem up to do_page_fault */ return 1; break; } pgdir = mm->pgd; if (pgdir == NULL) return 1; /* * Lock the Linux page table to prevent mmap and kswapd * from modifying entries while we search and update */ spin_lock(&mm->page_table_lock); ptep = find_linux_pte(pgdir, ea); /* * If no pte found or not present, send the problem up to * do_page_fault */ if (ptep && pte_present(*ptep)) { ret = __hash_page(ea, access, vsid, ptep); } else { /* If no pte, send the problem up to do_page_fault */ ret = 1; } spin_unlock(&mm->page_table_lock); return ret; } void flush_hash_page(unsigned long context, unsigned long ea, pte_t *ptep) { unsigned long vsid, vpn, va, hash, secondary, slot, flags, lock_slot; unsigned long large = 0, local = 0; pte_t pte; if ((ea >= USER_START) && (ea <= USER_END)) vsid = get_vsid(context, ea); else vsid = get_kernel_vsid(ea); va = (vsid << 28) | (ea & 0x0fffffff); if (large) vpn = va >> LARGE_PAGE_SHIFT; else vpn = va >> PAGE_SHIFT; lock_slot = get_lock_slot(vpn); hash = hpt_hash(vpn, large); spin_lock_irqsave(&hash_table_lock[lock_slot].lock, flags); pte = __pte(pte_update(ptep, _PAGE_HPTEFLAGS, 0)); secondary = (pte_val(pte) & _PAGE_SECONDARY) >> 15; if (secondary) hash = ~hash; slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP; slot += (pte_val(pte) & _PAGE_GROUP_IX) >> 12; if (pte_val(pte) & _PAGE_HASHPTE) { ppc_md.hpte_invalidate(slot, secondary, va, large, local); } spin_unlock_irqrestore(&hash_table_lock[lock_slot].lock, flags); } long plpar_pte_enter(unsigned long flags, unsigned long ptex, unsigned long new_pteh, unsigned long new_ptel, unsigned long *old_pteh_ret, unsigned long *old_ptel_ret) { unsigned long dummy, ret; ret = plpar_hcall(H_ENTER, flags, ptex, new_pteh, new_ptel, old_pteh_ret, old_ptel_ret, &dummy); return(ret); } long plpar_pte_remove(unsigned long flags, unsigned long ptex, unsigned long avpn, unsigned long *old_pteh_ret, unsigned long *old_ptel_ret) { unsigned long dummy; return plpar_hcall(H_REMOVE, flags, ptex, avpn, 0, old_pteh_ret, old_ptel_ret, &dummy); } long plpar_pte_read(unsigned long flags, unsigned long ptex, unsigned long *old_pteh_ret, unsigned long *old_ptel_ret) { unsigned long dummy; return plpar_hcall(H_READ, flags, ptex, 0, 0, old_pteh_ret, old_ptel_ret, &dummy); } long plpar_pte_protect(unsigned long flags, unsigned long ptex, unsigned long avpn) { return plpar_hcall_norets(H_PROTECT, flags, ptex, avpn); } static __inline__ void set_pp_bit(unsigned long pp, HPTE *addr) { unsigned long old; unsigned long *p = &addr->dw1.dword1; __asm__ __volatile__( "1: ldarx %0,0,%3\n\ rldimi %0,%2,0,62\n\ stdcx. %0,0,%3\n\ bne 1b" : "=&r" (old), "=m" (*p) : "r" (pp), "r" (p), "m" (*p) : "cc"); } /* * Calculate which hash_table_lock to use, based on the pteg being used. * * Given a VPN, use the high order 8 bits to select one of 2^7 locks. The * highest order bit is used to indicate primary vs. secondary group. If the * secondary is selected, complement the lock select bits. This results in * both the primary and secondary groups being covered under the same lock. */ inline unsigned long get_lock_slot(unsigned long vpn) { unsigned long lock_slot; #ifdef LOCK_SPLIT lock_slot = (hpt_hash(vpn,0) >> htab_data.htab_lock_shift) & 0xff; if(lock_slot & 0x80) lock_slot = (~lock_slot) & 0x7f; #else lock_slot = 0; #endif return(lock_slot); } /* * Functions used to retrieve word 0 of a given page table entry. * * Input : slot : PTE index within the page table of the entry to retrieve * Output: Contents of word 0 of the specified entry */ static unsigned long rpa_lpar_hpte_getword0(unsigned long slot) { unsigned long dword0; unsigned long lpar_rc; unsigned long dummy_word1; unsigned long flags; /* Read 1 pte at a time */ /* Do not need RPN to logical page translation */ /* No cross CEC PFT access */ flags = 0; lpar_rc = plpar_pte_read(flags, slot, &dword0, &dummy_word1); if (lpar_rc != H_Success) panic("Error on pte read in get_hpte0 rc = %lx\n", lpar_rc); return dword0; } unsigned long iSeries_hpte_getword0(unsigned long slot) { unsigned long dword0; HPTE hpte; HvCallHpt_get(&hpte, slot); dword0 = hpte.dw0.dword0; return dword0; } /* * Functions used to find the PTE for a particular virtual address. * Only used during boot when bolting pages. * * Input : vpn : virtual page number * Output: PTE index within the page table of the entry * -1 on failure */ static long hpte_find(unsigned long vpn) { HPTE *hptep; unsigned long hash; unsigned long i, j; long slot; Hpte_dword0 dw0; hash = hpt_hash(vpn, 0); for (j = 0; j < 2; j++) { slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP; for (i = 0; i < HPTES_PER_GROUP; i++) { hptep = htab_data.htab + slot; dw0 = hptep->dw0.dw0; if ((dw0.avpn == (vpn >> 11)) && dw0.v && (dw0.h == j)) { /* HPTE matches */ if (j) slot = -slot; return slot; } ++slot; } hash = ~hash; } return -1; } static long rpa_lpar_hpte_find(unsigned long vpn) { unsigned long hash; unsigned long i, j; long slot; union { unsigned long dword0; Hpte_dword0 dw0; } hpte_dw0; Hpte_dword0 dw0; hash = hpt_hash(vpn, 0); for (j = 0; j < 2; j++) { slot = (hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP; for (i = 0; i < HPTES_PER_GROUP; i++) { hpte_dw0.dword0 = rpa_lpar_hpte_getword0(slot); dw0 = hpte_dw0.dw0; if ((dw0.avpn == (vpn >> 11)) && dw0.v && (dw0.h == j)) { /* HPTE matches */ if (j) slot = -slot; return slot; } ++slot; } hash = ~hash; } return -1; } static long iSeries_hpte_find(unsigned long vpn) { HPTE hpte; long slot; /* * The HvCallHpt_findValid interface is as follows: * 0xffffffffffffffff : No entry found. * 0x00000000xxxxxxxx : Entry found in primary group, slot x * 0x80000000xxxxxxxx : Entry found in secondary group, slot x */ slot = HvCallHpt_findValid(&hpte, vpn); if (hpte.dw0.dw0.v) { if (slot < 0) { slot &= 0x7fffffffffffffff; slot = -slot; } } else { slot = -1; } return slot; } /* * Functions used to invalidate a page table entry from the page table * and tlb. * * Input : slot : PTE index within the page table of the entry to invalidated * va : Virtual address of the entry being invalidated * large : 1 = large page (16M) * local : 1 = Use tlbiel to only invalidate the local tlb */ static void hpte_invalidate(unsigned long slot, unsigned long secondary, unsigned long va, int large, int local) { HPTE *hptep = htab_data.htab + slot; Hpte_dword0 dw0; unsigned long vpn, avpn; unsigned long flags; if (large) vpn = va >> LARGE_PAGE_SHIFT; else vpn = va >> PAGE_SHIFT; avpn = vpn >> 11; dw0 = hptep->dw0.dw0; /* * Do not remove bolted entries. Alternatively, we could check * the AVPN, hash group, and valid bits. By doing it this way, * it is common with the pSeries LPAR optimal path. */ if (dw0.bolted) return; /* Invalidate the hpte. */ hptep->dw0.dword0 = 0; /* Invalidate the tlb */ spin_lock_irqsave(&pSeries_tlbie_lock, flags); _tlbie(va, large); spin_unlock_irqrestore(&pSeries_tlbie_lock, flags); } static void rpa_lpar_hpte_invalidate(unsigned long slot, unsigned long secondary, unsigned long va, int large, int local) { unsigned long lpar_rc; unsigned long dummy1, dummy2; /* * Don't remove a bolted entry. This case can occur when we bolt * pages dynamically after initial boot. */ lpar_rc = plpar_pte_remove(H_ANDCOND, slot, (0x1UL << 4), &dummy1, &dummy2); if (lpar_rc != H_Success) panic("Bad return code from invalidate rc = %lx\n", lpar_rc); } static void iSeries_hpte_invalidate(unsigned long slot, unsigned long secondary, unsigned long va, int large, int local) { HPTE lhpte; unsigned long vpn, avpn; if (large) vpn = va >> LARGE_PAGE_SHIFT; else vpn = va >> PAGE_SHIFT; avpn = vpn >> 11; lhpte.dw0.dword0 = iSeries_hpte_getword0(slot); if ((lhpte.dw0.dw0.avpn == avpn) && (lhpte.dw0.dw0.v) && (lhpte.dw0.dw0.h == secondary)) { HvCallHpt_invalidateSetSwBitsGet(slot, 0, 0); } } /* * Functions used to update page protection bits. * * Input : slot : PTE index within the page table of the entry to update * newpp : new page protection bits * va : Virtual address of the entry being updated * large : 1 = large page (16M) * Output: 0 on success, -1 on failure */ static long hpte_updatepp(unsigned long slot, unsigned long secondary, unsigned long newpp, unsigned long va, int large) { HPTE *hptep = htab_data.htab + slot; Hpte_dword0 dw0; Hpte_dword1 dw1; unsigned long vpn, avpn; unsigned long flags; if (large) vpn = va >> LARGE_PAGE_SHIFT; else vpn = va >> PAGE_SHIFT; avpn = vpn >> 11; dw0 = hptep->dw0.dw0; if ((dw0.avpn == avpn) && (dw0.v) && (dw0.h == secondary)) { /* Turn off valid bit in HPTE */ dw0.v = 0; hptep->dw0.dw0 = dw0; /* Ensure it is out of the tlb too */ spin_lock_irqsave(&pSeries_tlbie_lock, flags); _tlbie(va, large); spin_unlock_irqrestore(&pSeries_tlbie_lock, flags); /* Insert the new pp bits into the HPTE */ dw1 = hptep->dw1.dw1; dw1.pp = newpp; hptep->dw1.dw1 = dw1; /* Ensure it is visible before validating */ __asm__ __volatile__ ("eieio" : : : "memory"); /* Turn the valid bit back on in HPTE */ dw0.v = 1; hptep->dw0.dw0 = dw0; __asm__ __volatile__ ("ptesync" : : : "memory"); return 0; } return -1; } static long rpa_lpar_hpte_updatepp(unsigned long slot, unsigned long secondary, unsigned long newpp, unsigned long va, int large) { unsigned long lpar_rc; unsigned long flags = (newpp & 7); unsigned long avpn = va >> 23; HPTE hpte; lpar_rc = plpar_pte_read(0, slot, &hpte.dw0.dword0, &hpte.dw1.dword1); if ((hpte.dw0.dw0.avpn == avpn) && (hpte.dw0.dw0.v) && (hpte.dw0.dw0.h == secondary)) { lpar_rc = plpar_pte_protect(flags, slot, 0); if (lpar_rc != H_Success) panic("bad return code from pte protect rc = %lx\n", lpar_rc); return 0; } return -1; } static long iSeries_hpte_updatepp(unsigned long slot, unsigned long secondary, unsigned long newpp, unsigned long va, int large) { unsigned long vpn, avpn; HPTE hpte; if (large) vpn = va >> LARGE_PAGE_SHIFT; else vpn = va >> PAGE_SHIFT; avpn = vpn >> 11; HvCallHpt_get(&hpte, slot); if ((hpte.dw0.dw0.avpn == avpn) && (hpte.dw0.dw0.v) && (hpte.dw0.dw0.h == secondary)) { HvCallHpt_setPp(slot, newpp); return 0; } return -1; } /* * Functions used to update the page protection bits. Intended to be used * to create guard pages for kernel data structures on pages which are bolted * in the HPT. Assumes pages being operated on will not be stolen. * Does not work on large pages. No need to lock here because we are the * only user. * * Input : newpp : page protection flags * ea : effective kernel address to bolt. */ static void hpte_updateboltedpp(unsigned long newpp, unsigned long ea) { unsigned long vsid, va, vpn, flags; long slot; HPTE *hptep; vsid = get_kernel_vsid(ea); va = (vsid << 28) | (ea & 0x0fffffff); vpn = va >> PAGE_SHIFT; slot = hpte_find(vpn); if (slot == -1) panic("could not find page to bolt\n"); hptep = htab_data.htab + slot; set_pp_bit(newpp, hptep); /* Ensure it is out of the tlb too */ spin_lock_irqsave(&pSeries_tlbie_lock, flags); _tlbie(va, 0); spin_unlock_irqrestore(&pSeries_tlbie_lock, flags); } static void rpa_lpar_hpte_updateboltedpp(unsigned long newpp, unsigned long ea) { unsigned long lpar_rc; unsigned long vsid, va, vpn, flags; long slot; vsid = get_kernel_vsid(ea); va = (vsid << 28) | (ea & 0x0fffffff); vpn = va >> PAGE_SHIFT; slot = rpa_lpar_hpte_find(vpn); if (slot == -1) panic("updateboltedpp: Could not find page to bolt\n"); flags = newpp & 3; lpar_rc = plpar_pte_protect(flags, slot, 0); if (lpar_rc != H_Success) panic("Bad return code from pte bolted protect rc = %lx\n", lpar_rc); } void iSeries_hpte_updateboltedpp(unsigned long newpp, unsigned long ea) { unsigned long vsid,va,vpn; long slot; vsid = get_kernel_vsid( ea ); va = ( vsid << 28 ) | ( ea & 0x0fffffff ); vpn = va >> PAGE_SHIFT; slot = iSeries_hpte_find(vpn); if (slot == -1) panic("updateboltedpp: Could not find page to bolt\n"); HvCallHpt_setPp(slot, newpp); } /* * Functions used to insert new hardware page table entries. * Will castout non-bolted entries as necessary using a random * algorithm. * * Input : vpn : virtual page number * prpn : real page number in absolute space * hpteflags: page protection flags * bolted : 1 = bolt the page * large : 1 = large page (16M) * Output: hsss, where h = hash group, sss = slot within that group */ static long hpte_insert(unsigned long vpn, unsigned long prpn, unsigned long hpteflags, int bolted, int large) { HPTE *hptep; Hpte_dword0 dw0; HPTE lhpte; int i, secondary; unsigned long hash = hpt_hash(vpn, 0); unsigned long avpn = vpn >> 11; unsigned long arpn = physRpn_to_absRpn(prpn); unsigned long hpte_group; repeat: secondary = 0; hpte_group = ((hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP) & ~0x7UL; hptep = htab_data.htab + hpte_group; for (i = 0; i < HPTES_PER_GROUP; i++) { dw0 = hptep->dw0.dw0; if (!dw0.v) { /* retry with lock held */ dw0 = hptep->dw0.dw0; if (!dw0.v) break; } hptep++; } if (i == HPTES_PER_GROUP) { secondary = 1; hpte_group = ((~hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP) & ~0x7UL; hptep = htab_data.htab + hpte_group; for (i = 0; i < HPTES_PER_GROUP; i++) { dw0 = hptep->dw0.dw0; if (!dw0.v) { /* retry with lock held */ dw0 = hptep->dw0.dw0; if (!dw0.v) break; } hptep++; } if (i == HPTES_PER_GROUP) { if (mftb() & 0x1) hpte_group=((hash & htab_data.htab_hash_mask)* HPTES_PER_GROUP) & ~0x7UL; hpte_remove(hpte_group); goto repeat; } } lhpte.dw1.dword1 = 0; lhpte.dw1.dw1.rpn = arpn; lhpte.dw1.flags.flags = hpteflags; lhpte.dw0.dword0 = 0; lhpte.dw0.dw0.avpn = avpn; lhpte.dw0.dw0.h = secondary; lhpte.dw0.dw0.bolted = bolted; lhpte.dw0.dw0.v = 1; if (large) lhpte.dw0.dw0.l = 1; hptep->dw1.dword1 = lhpte.dw1.dword1; /* Guarantee the second dword is visible before the valid bit */ __asm__ __volatile__ ("eieio" : : : "memory"); /* * Now set the first dword including the valid bit * NOTE: this also unlocks the hpte */ hptep->dw0.dword0 = lhpte.dw0.dword0; __asm__ __volatile__ ("ptesync" : : : "memory"); return ((secondary << 3) | i); } static long rpa_lpar_hpte_insert(unsigned long vpn, unsigned long prpn, unsigned long hpteflags, int bolted, int large) { /* XXX fix for large page */ unsigned long lpar_rc; unsigned long flags; unsigned long slot; HPTE lhpte; int secondary; unsigned long hash = hpt_hash(vpn, 0); unsigned long avpn = vpn >> 11; unsigned long arpn = physRpn_to_absRpn(prpn); unsigned long hpte_group; /* Fill in the local HPTE with absolute rpn, avpn and flags */ lhpte.dw1.dword1 = 0; lhpte.dw1.dw1.rpn = arpn; lhpte.dw1.flags.flags = hpteflags; lhpte.dw0.dword0 = 0; lhpte.dw0.dw0.avpn = avpn; lhpte.dw0.dw0.bolted = bolted; lhpte.dw0.dw0.v = 1; if (large) lhpte.dw0.dw0.l = 1; /* Now fill in the actual HPTE */ /* Set CEC cookie to 0 */ /* Large page = 0 */ /* Zero page = 0 */ /* I-cache Invalidate = 0 */ /* I-cache synchronize = 0 */ /* Exact = 0 */ flags = 0; /* XXX why is this here? - Anton */ /* -- Because at one point we hit a case where non cachable * pages where marked coherent & this is rejected by the HV. * Perhaps it is no longer an issue ... DRENG. */ if (hpteflags & (_PAGE_GUARDED|_PAGE_NO_CACHE)) lhpte.dw1.flags.flags &= ~_PAGE_COHERENT; repeat: secondary = 0; lhpte.dw0.dw0.h = secondary; hpte_group = ((hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP) & ~0x7UL; __asm__ __volatile__ ( H_ENTER_r3 "mr 4, %2\n" "mr 5, %3\n" "mr 6, %4\n" "mr 7, %5\n" HVSC "mr %0, 3\n" "mr %1, 4\n" : "=r" (lpar_rc), "=r" (slot) : "r" (flags), "r" (hpte_group), "r" (lhpte.dw0.dword0), "r" (lhpte.dw1.dword1) : "r0", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "cc"); if (lpar_rc == H_PTEG_Full) { secondary = 1; lhpte.dw0.dw0.h = secondary; hpte_group = ((~hash & htab_data.htab_hash_mask) * HPTES_PER_GROUP) & ~0x7UL; __asm__ __volatile__ ( H_ENTER_r3 "mr 4, %2\n" "mr 5, %3\n" "mr 6, %4\n" "mr 7, %5\n" HVSC "mr %0, 3\n" "mr %1, 4\n" : "=r" (lpar_rc), "=r" (slot) : "r" (flags), "r" (hpte_group), "r" (lhpte.dw0.dword0), "r" (lhpte.dw1.dword1) : "r0", "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "r11", "r12", "cc"); if (lpar_rc == H_PTEG_Full) { if (mftb() & 0x1) hpte_group=((hash & htab_data.htab_hash_mask)* HPTES_PER_GROUP) & ~0x7UL; rpa_lpar_hpte_remove(hpte_group); goto repeat; } } if (lpar_rc != H_Success) panic("Bad return code from pte enter rc = %lx\n", lpar_rc); return ((secondary << 3) | (slot & 0x7)); } static long iSeries_hpte_insert(unsigned long vpn, unsigned long prpn, unsigned long hpteflags, int bolted, int large) { HPTE lhpte; unsigned long hash, hpte_group; unsigned long avpn = vpn >> 11; unsigned long arpn = physRpn_to_absRpn( prpn ); int secondary = 0; long slot; hash = hpt_hash(vpn, 0); repeat: slot = HvCallHpt_findValid(&lhpte, vpn); if (lhpte.dw0.dw0.v) { panic("select_hpte_slot found entry already valid\n"); } if (slot == -1) { /* No available entry found in either group */ if (mftb() & 0x1) { hpte_group=((hash & htab_data.htab_hash_mask)* HPTES_PER_GROUP) & ~0x7UL; } else { hpte_group=((~hash & htab_data.htab_hash_mask)* HPTES_PER_GROUP) & ~0x7UL; } hash = hpt_hash(vpn, 0); iSeries_hpte_remove(hpte_group); goto repeat; } else if (slot < 0) { slot &= 0x7fffffffffffffff; secondary = 1; } /* Create the HPTE */ lhpte.dw1.dword1 = 0; lhpte.dw1.dw1.rpn = arpn; lhpte.dw1.flags.flags = hpteflags; lhpte.dw0.dword0 = 0; lhpte.dw0.dw0.avpn = avpn; lhpte.dw0.dw0.h = secondary; lhpte.dw0.dw0.bolted = bolted; lhpte.dw0.dw0.v = 1; /* Now fill in the actual HPTE */ HvCallHpt_addValidate(slot, secondary, (HPTE *)&lhpte); return ((secondary << 3) | (slot & 0x7)); } /* * Functions used to remove hardware page table entries. * * Input : hpte_group: PTE index of the first entry in a group * Output: offset within the group of the entry removed or * -1 on failure */ static long hpte_remove(unsigned long hpte_group) { HPTE *hptep; Hpte_dword0 dw0; int i; int slot_offset; unsigned long vsid, group, pi, pi_high; unsigned long slot; unsigned long flags; int large; unsigned long va; /* pick a random slot to start at */ slot_offset = mftb() & 0x7; for (i = 0; i < HPTES_PER_GROUP; i++) { hptep = htab_data.htab + hpte_group + slot_offset; dw0 = hptep->dw0.dw0; if (dw0.v && !dw0.bolted) { /* retry with lock held */ dw0 = hptep->dw0.dw0; if (dw0.v && !dw0.bolted) break; } slot_offset++; slot_offset &= 0x7; } if (i == HPTES_PER_GROUP) return -1; large = dw0.l; /* Invalidate the hpte. NOTE: this also unlocks it */ hptep->dw0.dword0 = 0; /* Invalidate the tlb */ vsid = dw0.avpn >> 5; slot = hptep - htab_data.htab; group = slot >> 3; if (dw0.h) group = ~group; pi = (vsid ^ group) & 0x7ff; pi_high = (dw0.avpn & 0x1f) << 11; pi |= pi_high; if (large) va = pi << LARGE_PAGE_SHIFT; else va = pi << PAGE_SHIFT; spin_lock_irqsave(&pSeries_tlbie_lock, flags); _tlbie(va, large); spin_unlock_irqrestore(&pSeries_tlbie_lock, flags); return i; } static long rpa_lpar_hpte_remove(unsigned long hpte_group) { unsigned long slot_offset; unsigned long lpar_rc; int i; unsigned long dummy1, dummy2; /* pick a random slot to start at */ slot_offset = mftb() & 0x7; for (i = 0; i < HPTES_PER_GROUP; i++) { /* Don't remove a bolted entry */ lpar_rc = plpar_pte_remove(H_ANDCOND, hpte_group + slot_offset, (0x1UL << 4), &dummy1, &dummy2); if (lpar_rc == H_Success) return i; if (lpar_rc != H_Not_Found) panic("Bad return code from pte remove rc = %lx\n", lpar_rc); slot_offset++; slot_offset &= 0x7; } return -1; } static long iSeries_hpte_remove(unsigned long hpte_group) { unsigned long slot_offset; int i; HPTE lhpte; /* Pick a random slot to start at */ slot_offset = mftb() & 0x7; for (i = 0; i < HPTES_PER_GROUP; i++) { lhpte.dw0.dword0 = iSeries_hpte_getword0(hpte_group + slot_offset); if (!lhpte.dw0.dw0.bolted) { HvCallHpt_invalidateSetSwBitsGet(hpte_group + slot_offset, 0, 0); return i; } slot_offset++; slot_offset &= 0x7; } return -1; } void hpte_init_pSeries(void) { ppc_md.hpte_invalidate = hpte_invalidate; ppc_md.hpte_updatepp = hpte_updatepp; ppc_md.hpte_updateboltedpp = hpte_updateboltedpp; ppc_md.hpte_insert = hpte_insert; ppc_md.hpte_remove = hpte_remove; } void pSeries_lpar_mm_init(void) { ppc_md.hpte_invalidate = rpa_lpar_hpte_invalidate; ppc_md.hpte_updatepp = rpa_lpar_hpte_updatepp; ppc_md.hpte_updateboltedpp = rpa_lpar_hpte_updateboltedpp; ppc_md.hpte_insert = rpa_lpar_hpte_insert; ppc_md.hpte_remove = rpa_lpar_hpte_remove; } void hpte_init_iSeries(void) { ppc_md.hpte_invalidate = iSeries_hpte_invalidate; ppc_md.hpte_updatepp = iSeries_hpte_updatepp; ppc_md.hpte_updateboltedpp = iSeries_hpte_updateboltedpp; ppc_md.hpte_insert = iSeries_hpte_insert; ppc_md.hpte_remove = iSeries_hpte_remove; }