diff options
author | Nitin Gupta <nitin.m.gupta@oracle.com> | 2016-03-30 11:17:13 -0700 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2016-05-20 18:44:27 -0700 |
commit | 24e49ee3d76b70853a96520e46b8837e5eae65b2 (patch) | |
tree | 4893dd3011dc7a4db11082ce6953d1e1d2c13803 /arch/sparc/include/asm/tlbflush_64.h | |
parent | b1ac6b7b4061f6c92bacf6938f94fb61b2fbf7f3 (diff) |
sparc64: Reduce TLB flushes during hugepte changes
During hugepage map/unmap, TSB and TLB flushes are currently
issued at every PAGE_SIZE'd boundary which is unnecessary.
We now issue the flush at REAL_HPAGE_SIZE boundaries only.
Without this patch workloads which unmap a large hugepage
backed VMA region get CPU lockups due to excessive TLB
flush calls.
Orabug: 22365539, 22643230, 22995196
Signed-off-by: Nitin Gupta <nitin.m.gupta@oracle.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'arch/sparc/include/asm/tlbflush_64.h')
-rw-r--r-- | arch/sparc/include/asm/tlbflush_64.h | 3 |
1 files changed, 2 insertions, 1 deletions
diff --git a/arch/sparc/include/asm/tlbflush_64.h b/arch/sparc/include/asm/tlbflush_64.h index dea1cfa2122b..a8e192e90700 100644 --- a/arch/sparc/include/asm/tlbflush_64.h +++ b/arch/sparc/include/asm/tlbflush_64.h @@ -8,6 +8,7 @@ #define TLB_BATCH_NR 192 struct tlb_batch { + bool huge; struct mm_struct *mm; unsigned long tlb_nr; unsigned long active; @@ -16,7 +17,7 @@ struct tlb_batch { void flush_tsb_kernel_range(unsigned long start, unsigned long end); void flush_tsb_user(struct tlb_batch *tb); -void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr); +void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr, bool huge); /* TLB flush operations. */ |