fc2acab31b
zap_pte_range has been counting the pages it frees in tlb->freed, then tlb_finish_mmu has used that to update the mm's rss. That got stranger when I added anon_rss, yet updated it by a different route; and stranger when rss and anon_rss became mm_counters with special access macros. And it would no longer be viable if we're relying on page_table_lock to stabilize the mm_counter, but calling tlb_finish_mmu outside that lock. Remove the mmu_gather's freed field, let tlb_finish_mmu stick to its own business, just decrement the rss mm_counter in zap_pte_range (yes, there was some point to batching the update, and a subsequent patch restores that). And forget the anal paranoia of first reading the counter to avoid going negative - if rss does go negative, just fix that bug. Remove the mmu_gather's flushes and avoided_flushes from arm and arm26: no use was being made of them. But arm26 alone was actually using the freed, in the way some others use need_flush: give it a need_flush. arm26 seems to prefer spaces to tabs here: respect that. Signed-off-by: Hugh Dickins <hugh@veritas.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
150 lines
3.1 KiB
C
150 lines
3.1 KiB
C
/* arch/sparc64/mm/tlb.c
|
|
*
|
|
* Copyright (C) 2004 David S. Miller <davem@redhat.com>
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/init.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/swap.h>
|
|
|
|
#include <asm/pgtable.h>
|
|
#include <asm/pgalloc.h>
|
|
#include <asm/tlbflush.h>
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/tlb.h>
|
|
|
|
/* Heavily inspired by the ppc64 code. */
|
|
|
|
DEFINE_PER_CPU(struct mmu_gather, mmu_gathers) = { 0, };
|
|
|
|
void flush_tlb_pending(void)
|
|
{
|
|
struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);
|
|
|
|
if (mp->tlb_nr) {
|
|
if (CTX_VALID(mp->mm->context)) {
|
|
#ifdef CONFIG_SMP
|
|
smp_flush_tlb_pending(mp->mm, mp->tlb_nr,
|
|
&mp->vaddrs[0]);
|
|
#else
|
|
__flush_tlb_pending(CTX_HWBITS(mp->mm->context),
|
|
mp->tlb_nr, &mp->vaddrs[0]);
|
|
#endif
|
|
}
|
|
mp->tlb_nr = 0;
|
|
}
|
|
}
|
|
|
|
void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig)
|
|
{
|
|
struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);
|
|
unsigned long nr;
|
|
|
|
vaddr &= PAGE_MASK;
|
|
if (pte_exec(orig))
|
|
vaddr |= 0x1UL;
|
|
|
|
if (pte_dirty(orig)) {
|
|
unsigned long paddr, pfn = pte_pfn(orig);
|
|
struct address_space *mapping;
|
|
struct page *page;
|
|
|
|
if (!pfn_valid(pfn))
|
|
goto no_cache_flush;
|
|
|
|
page = pfn_to_page(pfn);
|
|
if (PageReserved(page))
|
|
goto no_cache_flush;
|
|
|
|
/* A real file page? */
|
|
mapping = page_mapping(page);
|
|
if (!mapping)
|
|
goto no_cache_flush;
|
|
|
|
paddr = (unsigned long) page_address(page);
|
|
if ((paddr ^ vaddr) & (1 << 13))
|
|
flush_dcache_page_all(mm, page);
|
|
}
|
|
|
|
no_cache_flush:
|
|
|
|
if (mp->fullmm)
|
|
return;
|
|
|
|
nr = mp->tlb_nr;
|
|
|
|
if (unlikely(nr != 0 && mm != mp->mm)) {
|
|
flush_tlb_pending();
|
|
nr = 0;
|
|
}
|
|
|
|
if (nr == 0)
|
|
mp->mm = mm;
|
|
|
|
mp->vaddrs[nr] = vaddr;
|
|
mp->tlb_nr = ++nr;
|
|
if (nr >= TLB_BATCH_NR)
|
|
flush_tlb_pending();
|
|
}
|
|
|
|
void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end)
|
|
{
|
|
struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);
|
|
unsigned long nr = mp->tlb_nr;
|
|
long s = start, e = end, vpte_base;
|
|
|
|
if (mp->fullmm)
|
|
return;
|
|
|
|
/* If start is greater than end, that is a real problem. */
|
|
BUG_ON(start > end);
|
|
|
|
/* However, straddling the VA space hole is quite normal. */
|
|
s &= PMD_MASK;
|
|
e = (e + PMD_SIZE - 1) & PMD_MASK;
|
|
|
|
vpte_base = (tlb_type == spitfire ?
|
|
VPTE_BASE_SPITFIRE :
|
|
VPTE_BASE_CHEETAH);
|
|
|
|
if (unlikely(nr != 0 && mm != mp->mm)) {
|
|
flush_tlb_pending();
|
|
nr = 0;
|
|
}
|
|
|
|
if (nr == 0)
|
|
mp->mm = mm;
|
|
|
|
start = vpte_base + (s >> (PAGE_SHIFT - 3));
|
|
end = vpte_base + (e >> (PAGE_SHIFT - 3));
|
|
|
|
/* If the request straddles the VA space hole, we
|
|
* need to swap start and end. The reason this
|
|
* occurs is that "vpte_base" is the center of
|
|
* the linear page table mapping area. Thus,
|
|
* high addresses with the sign bit set map to
|
|
* addresses below vpte_base and non-sign bit
|
|
* addresses map to addresses above vpte_base.
|
|
*/
|
|
if (end < start) {
|
|
unsigned long tmp = start;
|
|
|
|
start = end;
|
|
end = tmp;
|
|
}
|
|
|
|
while (start < end) {
|
|
mp->vaddrs[nr] = start;
|
|
mp->tlb_nr = ++nr;
|
|
if (nr >= TLB_BATCH_NR) {
|
|
flush_tlb_pending();
|
|
nr = 0;
|
|
}
|
|
start += PAGE_SIZE;
|
|
}
|
|
if (nr)
|
|
flush_tlb_pending();
|
|
}
|