123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254 |
- /* SPDX-License-Identifier: GPL-2.0 */
- /* clear_page.S: UltraSparc optimized copy page.
- *
- * Copyright (C) 1996, 1998, 1999, 2000, 2004 David S. Miller (davem@redhat.com)
- * Copyright (C) 1997 Jakub Jelinek (jakub@redhat.com)
- */
- #include <asm/visasm.h>
- #include <asm/thread_info.h>
- #include <asm/page.h>
- #include <asm/pgtable.h>
- #include <asm/spitfire.h>
- #include <asm/head.h>
- #include <asm/export.h>
- /* What we used to do was lock a TLB entry into a specific
- * TLB slot, clear the page with interrupts disabled, then
- * restore the original TLB entry. This was great for
- * disturbing the TLB as little as possible, but it meant
- * we had to keep interrupts disabled for a long time.
- *
- * Now, we simply use the normal TLB loading mechanism,
- * and this makes the cpu choose a slot all by itself.
- * Then we do a normal TLB flush on exit. We need only
- * disable preemption during the clear.
- */
- #define DCACHE_SIZE (PAGE_SIZE * 2)
- #if (PAGE_SHIFT == 13)
- #define PAGE_SIZE_REM 0x80
- #elif (PAGE_SHIFT == 16)
- #define PAGE_SIZE_REM 0x100
- #else
- #error Wrong PAGE_SHIFT specified
- #endif
- #define TOUCH(reg0, reg1, reg2, reg3, reg4, reg5, reg6, reg7) \
- fsrc2 %reg0, %f48; fsrc2 %reg1, %f50; \
- fsrc2 %reg2, %f52; fsrc2 %reg3, %f54; \
- fsrc2 %reg4, %f56; fsrc2 %reg5, %f58; \
- fsrc2 %reg6, %f60; fsrc2 %reg7, %f62;
- .text
- .align 32
- .globl copy_user_page
- .type copy_user_page,#function
- EXPORT_SYMBOL(copy_user_page)
- copy_user_page: /* %o0=dest, %o1=src, %o2=vaddr */
- lduw [%g6 + TI_PRE_COUNT], %o4
- sethi %hi(PAGE_OFFSET), %g2
- sethi %hi(PAGE_SIZE), %o3
- ldx [%g2 + %lo(PAGE_OFFSET)], %g2
- sethi %hi(PAGE_KERNEL_LOCKED), %g3
- ldx [%g3 + %lo(PAGE_KERNEL_LOCKED)], %g3
- sub %o0, %g2, %g1 ! dest paddr
- sub %o1, %g2, %g2 ! src paddr
- and %o2, %o3, %o0 ! vaddr D-cache alias bit
- or %g1, %g3, %g1 ! dest TTE data
- or %g2, %g3, %g2 ! src TTE data
- sethi %hi(TLBTEMP_BASE), %o3
- sethi %hi(DCACHE_SIZE), %o1
- add %o0, %o3, %o0 ! dest TTE vaddr
- add %o4, 1, %o2
- add %o0, %o1, %o1 ! src TTE vaddr
- /* Disable preemption. */
- mov TLB_TAG_ACCESS, %g3
- stw %o2, [%g6 + TI_PRE_COUNT]
- /* Load TLB entries. */
- rdpr %pstate, %o2
- wrpr %o2, PSTATE_IE, %pstate
- stxa %o0, [%g3] ASI_DMMU
- stxa %g1, [%g0] ASI_DTLB_DATA_IN
- membar #Sync
- stxa %o1, [%g3] ASI_DMMU
- stxa %g2, [%g0] ASI_DTLB_DATA_IN
- membar #Sync
- wrpr %o2, 0x0, %pstate
- cheetah_copy_page_insn:
- ba,pt %xcc, 9f
- nop
- 1:
- VISEntryHalf
- membar #StoreLoad | #StoreStore | #LoadStore
- sethi %hi((PAGE_SIZE/64)-2), %o2
- mov %o0, %g1
- prefetch [%o1 + 0x000], #one_read
- or %o2, %lo((PAGE_SIZE/64)-2), %o2
- prefetch [%o1 + 0x040], #one_read
- prefetch [%o1 + 0x080], #one_read
- prefetch [%o1 + 0x0c0], #one_read
- ldd [%o1 + 0x000], %f0
- prefetch [%o1 + 0x100], #one_read
- ldd [%o1 + 0x008], %f2
- prefetch [%o1 + 0x140], #one_read
- ldd [%o1 + 0x010], %f4
- prefetch [%o1 + 0x180], #one_read
- fsrc2 %f0, %f16
- ldd [%o1 + 0x018], %f6
- fsrc2 %f2, %f18
- ldd [%o1 + 0x020], %f8
- fsrc2 %f4, %f20
- ldd [%o1 + 0x028], %f10
- fsrc2 %f6, %f22
- ldd [%o1 + 0x030], %f12
- fsrc2 %f8, %f24
- ldd [%o1 + 0x038], %f14
- fsrc2 %f10, %f26
- ldd [%o1 + 0x040], %f0
- 1: ldd [%o1 + 0x048], %f2
- fsrc2 %f12, %f28
- ldd [%o1 + 0x050], %f4
- fsrc2 %f14, %f30
- stda %f16, [%o0] ASI_BLK_P
- ldd [%o1 + 0x058], %f6
- fsrc2 %f0, %f16
- ldd [%o1 + 0x060], %f8
- fsrc2 %f2, %f18
- ldd [%o1 + 0x068], %f10
- fsrc2 %f4, %f20
- ldd [%o1 + 0x070], %f12
- fsrc2 %f6, %f22
- ldd [%o1 + 0x078], %f14
- fsrc2 %f8, %f24
- ldd [%o1 + 0x080], %f0
- prefetch [%o1 + 0x180], #one_read
- fsrc2 %f10, %f26
- subcc %o2, 1, %o2
- add %o0, 0x40, %o0
- bne,pt %xcc, 1b
- add %o1, 0x40, %o1
- ldd [%o1 + 0x048], %f2
- fsrc2 %f12, %f28
- ldd [%o1 + 0x050], %f4
- fsrc2 %f14, %f30
- stda %f16, [%o0] ASI_BLK_P
- ldd [%o1 + 0x058], %f6
- fsrc2 %f0, %f16
- ldd [%o1 + 0x060], %f8
- fsrc2 %f2, %f18
- ldd [%o1 + 0x068], %f10
- fsrc2 %f4, %f20
- ldd [%o1 + 0x070], %f12
- fsrc2 %f6, %f22
- add %o0, 0x40, %o0
- ldd [%o1 + 0x078], %f14
- fsrc2 %f8, %f24
- fsrc2 %f10, %f26
- fsrc2 %f12, %f28
- fsrc2 %f14, %f30
- stda %f16, [%o0] ASI_BLK_P
- membar #Sync
- VISExitHalf
- ba,pt %xcc, 5f
- nop
- 9:
- VISEntry
- ldub [%g6 + TI_FAULT_CODE], %g3
- mov %o0, %g1
- cmp %g3, 0
- rd %asi, %g3
- be,a,pt %icc, 1f
- wr %g0, ASI_BLK_P, %asi
- wr %g0, ASI_BLK_COMMIT_P, %asi
- 1: ldda [%o1] ASI_BLK_P, %f0
- add %o1, 0x40, %o1
- ldda [%o1] ASI_BLK_P, %f16
- add %o1, 0x40, %o1
- sethi %hi(PAGE_SIZE), %o2
- 1: TOUCH(f0, f2, f4, f6, f8, f10, f12, f14)
- ldda [%o1] ASI_BLK_P, %f32
- stda %f48, [%o0] %asi
- add %o1, 0x40, %o1
- sub %o2, 0x40, %o2
- add %o0, 0x40, %o0
- TOUCH(f16, f18, f20, f22, f24, f26, f28, f30)
- ldda [%o1] ASI_BLK_P, %f0
- stda %f48, [%o0] %asi
- add %o1, 0x40, %o1
- sub %o2, 0x40, %o2
- add %o0, 0x40, %o0
- TOUCH(f32, f34, f36, f38, f40, f42, f44, f46)
- ldda [%o1] ASI_BLK_P, %f16
- stda %f48, [%o0] %asi
- sub %o2, 0x40, %o2
- add %o1, 0x40, %o1
- cmp %o2, PAGE_SIZE_REM
- bne,pt %xcc, 1b
- add %o0, 0x40, %o0
- #if (PAGE_SHIFT == 16)
- TOUCH(f0, f2, f4, f6, f8, f10, f12, f14)
- ldda [%o1] ASI_BLK_P, %f32
- stda %f48, [%o0] %asi
- add %o1, 0x40, %o1
- sub %o2, 0x40, %o2
- add %o0, 0x40, %o0
- TOUCH(f16, f18, f20, f22, f24, f26, f28, f30)
- ldda [%o1] ASI_BLK_P, %f0
- stda %f48, [%o0] %asi
- add %o1, 0x40, %o1
- sub %o2, 0x40, %o2
- add %o0, 0x40, %o0
- membar #Sync
- stda %f32, [%o0] %asi
- add %o0, 0x40, %o0
- stda %f0, [%o0] %asi
- #else
- membar #Sync
- stda %f0, [%o0] %asi
- add %o0, 0x40, %o0
- stda %f16, [%o0] %asi
- #endif
- membar #Sync
- wr %g3, 0x0, %asi
- VISExit
- 5:
- stxa %g0, [%g1] ASI_DMMU_DEMAP
- membar #Sync
- sethi %hi(DCACHE_SIZE), %g2
- stxa %g0, [%g1 + %g2] ASI_DMMU_DEMAP
- membar #Sync
- retl
- stw %o4, [%g6 + TI_PRE_COUNT]
- .size copy_user_page, .-copy_user_page
- .globl cheetah_patch_copy_page
- cheetah_patch_copy_page:
- sethi %hi(0x01000000), %o1 ! NOP
- sethi %hi(cheetah_copy_page_insn), %o0
- or %o0, %lo(cheetah_copy_page_insn), %o0
- stw %o1, [%o0]
- membar #StoreStore
- flush %o0
- retl
- nop
|