GNU Linux-libre 5.4.274-gnu1
[releases.git] / arch / csky / mm / tlb.c
1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd.
3
4 #include <linux/init.h>
5 #include <linux/mm.h>
6 #include <linux/module.h>
7 #include <linux/sched.h>
8
9 #include <asm/mmu_context.h>
10 #include <asm/pgtable.h>
11 #include <asm/setup.h>
12
13 /*
14  * One C-SKY MMU TLB entry contain two PFN/page entry, ie:
15  * 1VPN -> 2PFN
16  */
17 #define TLB_ENTRY_SIZE          (PAGE_SIZE * 2)
18 #define TLB_ENTRY_SIZE_MASK     (PAGE_MASK << 1)
19
20 void flush_tlb_all(void)
21 {
22         tlb_invalid_all();
23 }
24
25 void flush_tlb_mm(struct mm_struct *mm)
26 {
27 #ifdef CONFIG_CPU_HAS_TLBI
28         asm volatile("tlbi.asids %0"::"r"(cpu_asid(mm)));
29 #else
30         tlb_invalid_all();
31 #endif
32 }
33
34 /*
35  * MMU operation regs only could invalid tlb entry in jtlb and we
36  * need change asid field to invalid I-utlb & D-utlb.
37  */
38 #ifndef CONFIG_CPU_HAS_TLBI
39 #define restore_asid_inv_utlb(oldpid, newpid) \
40 do { \
41         if (oldpid == newpid) \
42                 write_mmu_entryhi(oldpid + 1); \
43         write_mmu_entryhi(oldpid); \
44 } while (0)
45 #endif
46
47 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
48                         unsigned long end)
49 {
50         unsigned long newpid = cpu_asid(vma->vm_mm);
51
52         start &= TLB_ENTRY_SIZE_MASK;
53         end   += TLB_ENTRY_SIZE - 1;
54         end   &= TLB_ENTRY_SIZE_MASK;
55
56 #ifdef CONFIG_CPU_HAS_TLBI
57         while (start < end) {
58                 asm volatile("tlbi.vas %0"::"r"(start | newpid));
59                 start += 2*PAGE_SIZE;
60         }
61         sync_is();
62 #else
63         {
64         unsigned long flags, oldpid;
65
66         local_irq_save(flags);
67         oldpid = read_mmu_entryhi() & ASID_MASK;
68         while (start < end) {
69                 int idx;
70
71                 write_mmu_entryhi(start | newpid);
72                 start += 2*PAGE_SIZE;
73                 tlb_probe();
74                 idx = read_mmu_index();
75                 if (idx >= 0)
76                         tlb_invalid_indexed();
77         }
78         restore_asid_inv_utlb(oldpid, newpid);
79         local_irq_restore(flags);
80         }
81 #endif
82 }
83
84 void flush_tlb_kernel_range(unsigned long start, unsigned long end)
85 {
86         start &= TLB_ENTRY_SIZE_MASK;
87         end   += TLB_ENTRY_SIZE - 1;
88         end   &= TLB_ENTRY_SIZE_MASK;
89
90 #ifdef CONFIG_CPU_HAS_TLBI
91         while (start < end) {
92                 asm volatile("tlbi.vaas %0"::"r"(start));
93                 start += 2*PAGE_SIZE;
94         }
95         sync_is();
96 #else
97         {
98         unsigned long flags, oldpid;
99
100         local_irq_save(flags);
101         oldpid = read_mmu_entryhi() & ASID_MASK;
102         while (start < end) {
103                 int idx;
104
105                 write_mmu_entryhi(start | oldpid);
106                 start += 2*PAGE_SIZE;
107                 tlb_probe();
108                 idx = read_mmu_index();
109                 if (idx >= 0)
110                         tlb_invalid_indexed();
111         }
112         restore_asid_inv_utlb(oldpid, oldpid);
113         local_irq_restore(flags);
114         }
115 #endif
116 }
117
118 void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr)
119 {
120         int newpid = cpu_asid(vma->vm_mm);
121
122         addr &= TLB_ENTRY_SIZE_MASK;
123
124 #ifdef CONFIG_CPU_HAS_TLBI
125         asm volatile("tlbi.vas %0"::"r"(addr | newpid));
126         sync_is();
127 #else
128         {
129         int oldpid, idx;
130         unsigned long flags;
131
132         local_irq_save(flags);
133         oldpid = read_mmu_entryhi() & ASID_MASK;
134         write_mmu_entryhi(addr | newpid);
135         tlb_probe();
136         idx = read_mmu_index();
137         if (idx >= 0)
138                 tlb_invalid_indexed();
139
140         restore_asid_inv_utlb(oldpid, newpid);
141         local_irq_restore(flags);
142         }
143 #endif
144 }
145
146 void flush_tlb_one(unsigned long addr)
147 {
148         addr &= TLB_ENTRY_SIZE_MASK;
149
150 #ifdef CONFIG_CPU_HAS_TLBI
151         asm volatile("tlbi.vaas %0"::"r"(addr));
152         sync_is();
153 #else
154         {
155         int oldpid, idx;
156         unsigned long flags;
157
158         local_irq_save(flags);
159         oldpid = read_mmu_entryhi() & ASID_MASK;
160         write_mmu_entryhi(addr | oldpid);
161         tlb_probe();
162         idx = read_mmu_index();
163         if (idx >= 0)
164                 tlb_invalid_indexed();
165
166         restore_asid_inv_utlb(oldpid, oldpid);
167         local_irq_restore(flags);
168         }
169 #endif
170 }
171 EXPORT_SYMBOL(flush_tlb_one);