parisc: fixes and cleanups in page cache flushing (4/4)
[linux-2.6/btrfs-unstable.git] / arch / parisc / kernel / cache.c
blob1c61b8245650ba8ef031b3500cbdd251f1d84a2a
1 /*
2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
4 * for more details.
6 * Copyright (C) 1999-2006 Helge Deller <deller@gmx.de> (07-13-1999)
7 * Copyright (C) 1999 SuSE GmbH Nuernberg
8 * Copyright (C) 2000 Philipp Rumpf (prumpf@tux.org)
10 * Cache and TLB management
14 #include <linux/init.h>
15 #include <linux/kernel.h>
16 #include <linux/mm.h>
17 #include <linux/module.h>
18 #include <linux/seq_file.h>
19 #include <linux/pagemap.h>
20 #include <linux/sched.h>
21 #include <asm/pdc.h>
22 #include <asm/cache.h>
23 #include <asm/cacheflush.h>
24 #include <asm/tlbflush.h>
25 #include <asm/page.h>
26 #include <asm/pgalloc.h>
27 #include <asm/processor.h>
28 #include <asm/sections.h>
29 #include <asm/shmparam.h>
31 int split_tlb __read_mostly;
32 int dcache_stride __read_mostly;
33 int icache_stride __read_mostly;
34 EXPORT_SYMBOL(dcache_stride);
36 void flush_dcache_page_asm(unsigned long phys_addr, unsigned long vaddr);
37 EXPORT_SYMBOL(flush_dcache_page_asm);
38 void flush_icache_page_asm(unsigned long phys_addr, unsigned long vaddr);
41 /* On some machines (e.g. ones with the Merced bus), there can be
42 * only a single PxTLB broadcast at a time; this must be guaranteed
43 * by software. We put a spinlock around all TLB flushes to
44 * ensure this.
46 DEFINE_SPINLOCK(pa_tlb_lock);
48 struct pdc_cache_info cache_info __read_mostly;
49 #ifndef CONFIG_PA20
50 static struct pdc_btlb_info btlb_info __read_mostly;
51 #endif
53 #ifdef CONFIG_SMP
54 void
55 flush_data_cache(void)
57 on_each_cpu(flush_data_cache_local, NULL, 1);
59 void
60 flush_instruction_cache(void)
62 on_each_cpu(flush_instruction_cache_local, NULL, 1);
64 #endif
66 void
67 flush_cache_all_local(void)
69 flush_instruction_cache_local(NULL);
70 flush_data_cache_local(NULL);
72 EXPORT_SYMBOL(flush_cache_all_local);
74 void
75 update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep)
77 struct page *page = pte_page(*ptep);
79 if (pfn_valid(page_to_pfn(page)) && page_mapping(page) &&
80 test_bit(PG_dcache_dirty, &page->flags)) {
82 flush_kernel_dcache_page(page);
83 clear_bit(PG_dcache_dirty, &page->flags);
84 } else if (parisc_requires_coherency())
85 flush_kernel_dcache_page(page);
88 void
89 show_cache_info(struct seq_file *m)
91 char buf[32];
93 seq_printf(m, "I-cache\t\t: %ld KB\n",
94 cache_info.ic_size/1024 );
95 if (cache_info.dc_loop != 1)
96 snprintf(buf, 32, "%lu-way associative", cache_info.dc_loop);
97 seq_printf(m, "D-cache\t\t: %ld KB (%s%s, %s)\n",
98 cache_info.dc_size/1024,
99 (cache_info.dc_conf.cc_wt ? "WT":"WB"),
100 (cache_info.dc_conf.cc_sh ? ", shared I/D":""),
101 ((cache_info.dc_loop == 1) ? "direct mapped" : buf));
102 seq_printf(m, "ITLB entries\t: %ld\n" "DTLB entries\t: %ld%s\n",
103 cache_info.it_size,
104 cache_info.dt_size,
105 cache_info.dt_conf.tc_sh ? " - shared with ITLB":""
108 #ifndef CONFIG_PA20
109 /* BTLB - Block TLB */
110 if (btlb_info.max_size==0) {
111 seq_printf(m, "BTLB\t\t: not supported\n" );
112 } else {
113 seq_printf(m,
114 "BTLB fixed\t: max. %d pages, pagesize=%d (%dMB)\n"
115 "BTLB fix-entr.\t: %d instruction, %d data (%d combined)\n"
116 "BTLB var-entr.\t: %d instruction, %d data (%d combined)\n",
117 btlb_info.max_size, (int)4096,
118 btlb_info.max_size>>8,
119 btlb_info.fixed_range_info.num_i,
120 btlb_info.fixed_range_info.num_d,
121 btlb_info.fixed_range_info.num_comb,
122 btlb_info.variable_range_info.num_i,
123 btlb_info.variable_range_info.num_d,
124 btlb_info.variable_range_info.num_comb
127 #endif
130 void __init
131 parisc_cache_init(void)
133 if (pdc_cache_info(&cache_info) < 0)
134 panic("parisc_cache_init: pdc_cache_info failed");
136 #if 0
137 printk("ic_size %lx dc_size %lx it_size %lx\n",
138 cache_info.ic_size,
139 cache_info.dc_size,
140 cache_info.it_size);
142 printk("DC base 0x%lx stride 0x%lx count 0x%lx loop 0x%lx\n",
143 cache_info.dc_base,
144 cache_info.dc_stride,
145 cache_info.dc_count,
146 cache_info.dc_loop);
148 printk("dc_conf = 0x%lx alias %d blk %d line %d shift %d\n",
149 *(unsigned long *) (&cache_info.dc_conf),
150 cache_info.dc_conf.cc_alias,
151 cache_info.dc_conf.cc_block,
152 cache_info.dc_conf.cc_line,
153 cache_info.dc_conf.cc_shift);
154 printk(" wt %d sh %d cst %d hv %d\n",
155 cache_info.dc_conf.cc_wt,
156 cache_info.dc_conf.cc_sh,
157 cache_info.dc_conf.cc_cst,
158 cache_info.dc_conf.cc_hv);
160 printk("IC base 0x%lx stride 0x%lx count 0x%lx loop 0x%lx\n",
161 cache_info.ic_base,
162 cache_info.ic_stride,
163 cache_info.ic_count,
164 cache_info.ic_loop);
166 printk("ic_conf = 0x%lx alias %d blk %d line %d shift %d\n",
167 *(unsigned long *) (&cache_info.ic_conf),
168 cache_info.ic_conf.cc_alias,
169 cache_info.ic_conf.cc_block,
170 cache_info.ic_conf.cc_line,
171 cache_info.ic_conf.cc_shift);
172 printk(" wt %d sh %d cst %d hv %d\n",
173 cache_info.ic_conf.cc_wt,
174 cache_info.ic_conf.cc_sh,
175 cache_info.ic_conf.cc_cst,
176 cache_info.ic_conf.cc_hv);
178 printk("D-TLB conf: sh %d page %d cst %d aid %d pad1 %d\n",
179 cache_info.dt_conf.tc_sh,
180 cache_info.dt_conf.tc_page,
181 cache_info.dt_conf.tc_cst,
182 cache_info.dt_conf.tc_aid,
183 cache_info.dt_conf.tc_pad1);
185 printk("I-TLB conf: sh %d page %d cst %d aid %d pad1 %d\n",
186 cache_info.it_conf.tc_sh,
187 cache_info.it_conf.tc_page,
188 cache_info.it_conf.tc_cst,
189 cache_info.it_conf.tc_aid,
190 cache_info.it_conf.tc_pad1);
191 #endif
193 split_tlb = 0;
194 if (cache_info.dt_conf.tc_sh == 0 || cache_info.dt_conf.tc_sh == 2) {
195 if (cache_info.dt_conf.tc_sh == 2)
196 printk(KERN_WARNING "Unexpected TLB configuration. "
197 "Will flush I/D separately (could be optimized).\n");
199 split_tlb = 1;
202 /* "New and Improved" version from Jim Hull
203 * (1 << (cc_block-1)) * (cc_line << (4 + cnf.cc_shift))
204 * The following CAFL_STRIDE is an optimized version, see
205 * http://lists.parisc-linux.org/pipermail/parisc-linux/2004-June/023625.html
206 * http://lists.parisc-linux.org/pipermail/parisc-linux/2004-June/023671.html
208 #define CAFL_STRIDE(cnf) (cnf.cc_line << (3 + cnf.cc_block + cnf.cc_shift))
209 dcache_stride = CAFL_STRIDE(cache_info.dc_conf);
210 icache_stride = CAFL_STRIDE(cache_info.ic_conf);
211 #undef CAFL_STRIDE
213 #ifndef CONFIG_PA20
214 if (pdc_btlb_info(&btlb_info) < 0) {
215 memset(&btlb_info, 0, sizeof btlb_info);
217 #endif
219 if ((boot_cpu_data.pdc.capabilities & PDC_MODEL_NVA_MASK) ==
220 PDC_MODEL_NVA_UNSUPPORTED) {
221 printk(KERN_WARNING "parisc_cache_init: Only equivalent aliasing supported!\n");
222 #if 0
223 panic("SMP kernel required to avoid non-equivalent aliasing");
224 #endif
228 void disable_sr_hashing(void)
230 int srhash_type, retval;
231 unsigned long space_bits;
233 switch (boot_cpu_data.cpu_type) {
234 case pcx: /* We shouldn't get this far. setup.c should prevent it. */
235 BUG();
236 return;
238 case pcxs:
239 case pcxt:
240 case pcxt_:
241 srhash_type = SRHASH_PCXST;
242 break;
244 case pcxl:
245 srhash_type = SRHASH_PCXL;
246 break;
248 case pcxl2: /* pcxl2 doesn't support space register hashing */
249 return;
251 default: /* Currently all PA2.0 machines use the same ins. sequence */
252 srhash_type = SRHASH_PA20;
253 break;
256 disable_sr_hashing_asm(srhash_type);
258 retval = pdc_spaceid_bits(&space_bits);
259 /* If this procedure isn't implemented, don't panic. */
260 if (retval < 0 && retval != PDC_BAD_OPTION)
261 panic("pdc_spaceid_bits call failed.\n");
262 if (space_bits != 0)
263 panic("SpaceID hashing is still on!\n");
266 static inline void
267 __flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr,
268 unsigned long physaddr)
270 flush_dcache_page_asm(physaddr, vmaddr);
271 if (vma->vm_flags & VM_EXEC)
272 flush_icache_page_asm(physaddr, vmaddr);
275 void flush_dcache_page(struct page *page)
277 struct address_space *mapping = page_mapping(page);
278 struct vm_area_struct *mpnt;
279 unsigned long offset;
280 unsigned long addr, old_addr = 0;
281 pgoff_t pgoff;
283 if (mapping && !mapping_mapped(mapping)) {
284 set_bit(PG_dcache_dirty, &page->flags);
285 return;
288 flush_kernel_dcache_page(page);
290 if (!mapping)
291 return;
293 pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT);
295 /* We have carefully arranged in arch_get_unmapped_area() that
296 * *any* mappings of a file are always congruently mapped (whether
297 * declared as MAP_PRIVATE or MAP_SHARED), so we only need
298 * to flush one address here for them all to become coherent */
300 flush_dcache_mmap_lock(mapping);
301 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) {
302 offset = (pgoff - mpnt->vm_pgoff) << PAGE_SHIFT;
303 addr = mpnt->vm_start + offset;
305 /* The TLB is the engine of coherence on parisc: The
306 * CPU is entitled to speculate any page with a TLB
307 * mapping, so here we kill the mapping then flush the
308 * page along a special flush only alias mapping.
309 * This guarantees that the page is no-longer in the
310 * cache for any process and nor may it be
311 * speculatively read in (until the user or kernel
312 * specifically accesses it, of course) */
314 flush_tlb_page(mpnt, addr);
315 if (old_addr == 0 || (old_addr & (SHMLBA - 1)) != (addr & (SHMLBA - 1))) {
316 __flush_cache_page(mpnt, addr, page_to_phys(page));
317 if (old_addr)
318 printk(KERN_ERR "INEQUIVALENT ALIASES 0x%lx and 0x%lx in file %s\n", old_addr, addr, mpnt->vm_file ? (char *)mpnt->vm_file->f_path.dentry->d_name.name : "(null)");
319 old_addr = addr;
322 flush_dcache_mmap_unlock(mapping);
324 EXPORT_SYMBOL(flush_dcache_page);
326 /* Defined in arch/parisc/kernel/pacache.S */
327 EXPORT_SYMBOL(flush_kernel_dcache_range_asm);
328 EXPORT_SYMBOL(flush_kernel_dcache_page_asm);
329 EXPORT_SYMBOL(flush_data_cache_local);
330 EXPORT_SYMBOL(flush_kernel_icache_range_asm);
332 #define FLUSH_THRESHOLD 0x80000 /* 0.5MB */
333 int parisc_cache_flush_threshold __read_mostly = FLUSH_THRESHOLD;
335 void __init parisc_setup_cache_timing(void)
337 unsigned long rangetime, alltime;
338 unsigned long size;
340 alltime = mfctl(16);
341 flush_data_cache();
342 alltime = mfctl(16) - alltime;
344 size = (unsigned long)(_end - _text);
345 rangetime = mfctl(16);
346 flush_kernel_dcache_range((unsigned long)_text, size);
347 rangetime = mfctl(16) - rangetime;
349 printk(KERN_DEBUG "Whole cache flush %lu cycles, flushing %lu bytes %lu cycles\n",
350 alltime, size, rangetime);
352 /* Racy, but if we see an intermediate value, it's ok too... */
353 parisc_cache_flush_threshold = size * alltime / rangetime;
355 parisc_cache_flush_threshold = (parisc_cache_flush_threshold + L1_CACHE_BYTES - 1) &~ (L1_CACHE_BYTES - 1);
356 if (!parisc_cache_flush_threshold)
357 parisc_cache_flush_threshold = FLUSH_THRESHOLD;
359 if (parisc_cache_flush_threshold > cache_info.dc_size)
360 parisc_cache_flush_threshold = cache_info.dc_size;
362 printk(KERN_INFO "Setting cache flush threshold to %x (%d CPUs online)\n", parisc_cache_flush_threshold, num_online_cpus());
365 extern void purge_kernel_dcache_page_asm(unsigned long);
366 extern void clear_user_page_asm(void *, unsigned long);
367 extern void copy_user_page_asm(void *, void *, unsigned long);
369 void flush_kernel_dcache_page_addr(void *addr)
371 unsigned long flags;
373 flush_kernel_dcache_page_asm(addr);
374 purge_tlb_start(flags);
375 pdtlb_kernel(addr);
376 purge_tlb_end(flags);
378 EXPORT_SYMBOL(flush_kernel_dcache_page_addr);
380 void clear_user_page(void *vto, unsigned long vaddr, struct page *page)
382 clear_page_asm(vto);
383 if (!parisc_requires_coherency())
384 flush_kernel_dcache_page_asm(vto);
386 EXPORT_SYMBOL(clear_user_page);
388 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr,
389 struct page *pg)
391 /* Copy using kernel mapping. No coherency is needed
392 (all in kmap/kunmap) on machines that don't support
393 non-equivalent aliasing. However, the `from' page
394 needs to be flushed before it can be accessed through
395 the kernel mapping. */
396 preempt_disable();
397 flush_dcache_page_asm(__pa(vfrom), vaddr);
398 preempt_enable();
399 copy_page_asm(vto, vfrom);
400 if (!parisc_requires_coherency())
401 flush_kernel_dcache_page_asm(vto);
403 EXPORT_SYMBOL(copy_user_page);
405 #ifdef CONFIG_PA8X00
407 void kunmap_parisc(void *addr)
409 if (parisc_requires_coherency())
410 flush_kernel_dcache_page_addr(addr);
412 EXPORT_SYMBOL(kunmap_parisc);
413 #endif
415 void purge_tlb_entries(struct mm_struct *mm, unsigned long addr)
417 unsigned long flags;
419 /* Note: purge_tlb_entries can be called at startup with
420 no context. */
422 /* Disable preemption while we play with %sr1. */
423 preempt_disable();
424 mtsp(mm->context, 1);
425 purge_tlb_start(flags);
426 pdtlb(addr);
427 pitlb(addr);
428 purge_tlb_end(flags);
429 preempt_enable();
431 EXPORT_SYMBOL(purge_tlb_entries);
433 void __flush_tlb_range(unsigned long sid, unsigned long start,
434 unsigned long end)
436 unsigned long npages;
438 npages = ((end - (start & PAGE_MASK)) + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
439 if (npages >= 512) /* 2MB of space: arbitrary, should be tuned */
440 flush_tlb_all();
441 else {
442 unsigned long flags;
444 mtsp(sid, 1);
445 purge_tlb_start(flags);
446 if (split_tlb) {
447 while (npages--) {
448 pdtlb(start);
449 pitlb(start);
450 start += PAGE_SIZE;
452 } else {
453 while (npages--) {
454 pdtlb(start);
455 start += PAGE_SIZE;
458 purge_tlb_end(flags);
462 static void cacheflush_h_tmp_function(void *dummy)
464 flush_cache_all_local();
467 void flush_cache_all(void)
469 on_each_cpu(cacheflush_h_tmp_function, NULL, 1);
472 static inline unsigned long mm_total_size(struct mm_struct *mm)
474 struct vm_area_struct *vma;
475 unsigned long usize = 0;
477 for (vma = mm->mmap; vma; vma = vma->vm_next)
478 usize += vma->vm_end - vma->vm_start;
479 return usize;
482 static inline pte_t *get_ptep(pgd_t *pgd, unsigned long addr)
484 pte_t *ptep = NULL;
486 if (!pgd_none(*pgd)) {
487 pud_t *pud = pud_offset(pgd, addr);
488 if (!pud_none(*pud)) {
489 pmd_t *pmd = pmd_offset(pud, addr);
490 if (!pmd_none(*pmd))
491 ptep = pte_offset_map(pmd, addr);
494 return ptep;
497 void flush_cache_mm(struct mm_struct *mm)
499 /* Flushing the whole cache on each cpu takes forever on
500 rp3440, etc. So, avoid it if the mm isn't too big. */
501 if (mm_total_size(mm) < parisc_cache_flush_threshold) {
502 struct vm_area_struct *vma;
504 if (mm->context == mfsp(3)) {
505 for (vma = mm->mmap; vma; vma = vma->vm_next) {
506 flush_user_dcache_range_asm(vma->vm_start,
507 vma->vm_end);
508 if (vma->vm_flags & VM_EXEC)
509 flush_user_icache_range_asm(
510 vma->vm_start, vma->vm_end);
512 } else {
513 pgd_t *pgd = mm->pgd;
515 for (vma = mm->mmap; vma; vma = vma->vm_next) {
516 unsigned long addr;
518 for (addr = vma->vm_start; addr < vma->vm_end;
519 addr += PAGE_SIZE) {
520 pte_t *ptep = get_ptep(pgd, addr);
521 if (ptep != NULL) {
522 pte_t pte = *ptep;
523 __flush_cache_page(vma, addr,
524 page_to_phys(pte_page(pte)));
529 return;
532 #ifdef CONFIG_SMP
533 flush_cache_all();
534 #else
535 flush_cache_all_local();
536 #endif
539 void
540 flush_user_dcache_range(unsigned long start, unsigned long end)
542 if ((end - start) < parisc_cache_flush_threshold)
543 flush_user_dcache_range_asm(start,end);
544 else
545 flush_data_cache();
548 void
549 flush_user_icache_range(unsigned long start, unsigned long end)
551 if ((end - start) < parisc_cache_flush_threshold)
552 flush_user_icache_range_asm(start,end);
553 else
554 flush_instruction_cache();
557 void flush_cache_range(struct vm_area_struct *vma,
558 unsigned long start, unsigned long end)
560 BUG_ON(!vma->vm_mm->context);
562 if ((end - start) < parisc_cache_flush_threshold) {
563 if (vma->vm_mm->context == mfsp(3)) {
564 flush_user_dcache_range_asm(start, end);
565 if (vma->vm_flags & VM_EXEC)
566 flush_user_icache_range_asm(start, end);
567 } else {
568 unsigned long addr;
569 pgd_t *pgd = vma->vm_mm->pgd;
571 for (addr = start & PAGE_MASK; addr < end;
572 addr += PAGE_SIZE) {
573 pte_t *ptep = get_ptep(pgd, addr);
574 if (ptep != NULL) {
575 pte_t pte = *ptep;
576 flush_cache_page(vma,
577 addr, pte_pfn(pte));
581 } else {
582 #ifdef CONFIG_SMP
583 flush_cache_all();
584 #else
585 flush_cache_all_local();
586 #endif
590 void
591 flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn)
593 BUG_ON(!vma->vm_mm->context);
595 flush_tlb_page(vma, vmaddr);
596 __flush_cache_page(vma, vmaddr, page_to_phys(pfn_to_page(pfn)));
600 #ifdef CONFIG_PARISC_TMPALIAS
602 void clear_user_highpage(struct page *page, unsigned long vaddr)
604 void *vto;
605 unsigned long flags;
607 /* Clear using TMPALIAS region. The page doesn't need to
608 be flushed but the kernel mapping needs to be purged. */
610 vto = kmap_atomic(page, KM_USER0);
612 /* The PA-RISC 2.0 Architecture book states on page F-6:
613 "Before a write-capable translation is enabled, *all*
614 non-equivalently-aliased translations must be removed
615 from the page table and purged from the TLB. (Note
616 that the caches are not required to be flushed at this
617 time.) Before any non-equivalent aliased translation
618 is re-enabled, the virtual address range for the writeable
619 page (the entire page) must be flushed from the cache,
620 and the write-capable translation removed from the page
621 table and purged from the TLB." */
623 purge_kernel_dcache_page_asm((unsigned long)vto);
624 purge_tlb_start(flags);
625 pdtlb_kernel(vto);
626 purge_tlb_end(flags);
627 preempt_disable();
628 clear_user_page_asm(vto, vaddr);
629 preempt_enable();
631 pagefault_enable(); /* kunmap_atomic(addr, KM_USER0); */
634 void copy_user_highpage(struct page *to, struct page *from,
635 unsigned long vaddr, struct vm_area_struct *vma)
637 void *vfrom, *vto;
638 unsigned long flags;
640 /* Copy using TMPALIAS region. This has the advantage
641 that the `from' page doesn't need to be flushed. However,
642 the `to' page must be flushed in copy_user_page_asm since
643 it can be used to bring in executable code. */
645 vfrom = kmap_atomic(from, KM_USER0);
646 vto = kmap_atomic(to, KM_USER1);
648 purge_kernel_dcache_page_asm((unsigned long)vto);
649 purge_tlb_start(flags);
650 pdtlb_kernel(vto);
651 pdtlb_kernel(vfrom);
652 purge_tlb_end(flags);
653 preempt_disable();
654 copy_user_page_asm(vto, vfrom, vaddr);
655 flush_dcache_page_asm(__pa(vto), vaddr);
656 preempt_enable();
658 pagefault_enable(); /* kunmap_atomic(addr, KM_USER1); */
659 pagefault_enable(); /* kunmap_atomic(addr, KM_USER0); */
662 #endif /* CONFIG_PARISC_TMPALIAS */