Merge branch 'for-next' of git://git.kernel.org/pub/scm/linux/kernel/git/hch/hfsplus
[sfrench/cifs-2.6.git] / arch / sh / mm / tlbflush_32.c
1 /*
2  * TLB flushing operations for SH with an MMU.
3  *
4  *  Copyright (C) 1999  Niibe Yutaka
5  *  Copyright (C) 2003  Paul Mundt
6  *
7  * This file is subject to the terms and conditions of the GNU General Public
8  * License.  See the file "COPYING" in the main directory of this archive
9  * for more details.
10  */
11 #include <linux/mm.h>
12 #include <asm/mmu_context.h>
13 #include <asm/tlbflush.h>
14
15 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page)
16 {
17         unsigned int cpu = smp_processor_id();
18
19         if (vma->vm_mm && cpu_context(cpu, vma->vm_mm) != NO_CONTEXT) {
20                 unsigned long flags;
21                 unsigned long asid;
22                 unsigned long saved_asid = MMU_NO_ASID;
23
24                 asid = cpu_asid(cpu, vma->vm_mm);
25                 page &= PAGE_MASK;
26
27                 local_irq_save(flags);
28                 if (vma->vm_mm != current->mm) {
29                         saved_asid = get_asid();
30                         set_asid(asid);
31                 }
32                 local_flush_tlb_one(asid, page);
33                 if (saved_asid != MMU_NO_ASID)
34                         set_asid(saved_asid);
35                 local_irq_restore(flags);
36         }
37 }
38
39 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
40                            unsigned long end)
41 {
42         struct mm_struct *mm = vma->vm_mm;
43         unsigned int cpu = smp_processor_id();
44
45         if (cpu_context(cpu, mm) != NO_CONTEXT) {
46                 unsigned long flags;
47                 int size;
48
49                 local_irq_save(flags);
50                 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
51                 if (size > (MMU_NTLB_ENTRIES/4)) { /* Too many TLB to flush */
52                         cpu_context(cpu, mm) = NO_CONTEXT;
53                         if (mm == current->mm)
54                                 activate_context(mm, cpu);
55                 } else {
56                         unsigned long asid;
57                         unsigned long saved_asid = MMU_NO_ASID;
58
59                         asid = cpu_asid(cpu, mm);
60                         start &= PAGE_MASK;
61                         end += (PAGE_SIZE - 1);
62                         end &= PAGE_MASK;
63                         if (mm != current->mm) {
64                                 saved_asid = get_asid();
65                                 set_asid(asid);
66                         }
67                         while (start < end) {
68                                 local_flush_tlb_one(asid, start);
69                                 start += PAGE_SIZE;
70                         }
71                         if (saved_asid != MMU_NO_ASID)
72                                 set_asid(saved_asid);
73                 }
74                 local_irq_restore(flags);
75         }
76 }
77
78 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end)
79 {
80         unsigned int cpu = smp_processor_id();
81         unsigned long flags;
82         int size;
83
84         local_irq_save(flags);
85         size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
86         if (size > (MMU_NTLB_ENTRIES/4)) { /* Too many TLB to flush */
87                 local_flush_tlb_all();
88         } else {
89                 unsigned long asid;
90                 unsigned long saved_asid = get_asid();
91
92                 asid = cpu_asid(cpu, &init_mm);
93                 start &= PAGE_MASK;
94                 end += (PAGE_SIZE - 1);
95                 end &= PAGE_MASK;
96                 set_asid(asid);
97                 while (start < end) {
98                         local_flush_tlb_one(asid, start);
99                         start += PAGE_SIZE;
100                 }
101                 set_asid(saved_asid);
102         }
103         local_irq_restore(flags);
104 }
105
106 void local_flush_tlb_mm(struct mm_struct *mm)
107 {
108         unsigned int cpu = smp_processor_id();
109
110         /* Invalidate all TLB of this process. */
111         /* Instead of invalidating each TLB, we get new MMU context. */
112         if (cpu_context(cpu, mm) != NO_CONTEXT) {
113                 unsigned long flags;
114
115                 local_irq_save(flags);
116                 cpu_context(cpu, mm) = NO_CONTEXT;
117                 if (mm == current->mm)
118                         activate_context(mm, cpu);
119                 local_irq_restore(flags);
120         }
121 }
122
123 void __flush_tlb_global(void)
124 {
125         unsigned long flags;
126
127         local_irq_save(flags);
128
129         /*
130          * This is the most destructive of the TLB flushing options,
131          * and will tear down all of the UTLB/ITLB mappings, including
132          * wired entries.
133          */
134         __raw_writel(__raw_readl(MMUCR) | MMUCR_TI, MMUCR);
135
136         local_irq_restore(flags);
137 }