[U-Boot] [RFC PATCH] armv8: cache: Switch MMU table without turning off MMU
York Sun
york.sun at nxp.com
Fri Feb 3 22:22:48 UTC 2017
We don't have to completely turn off MMU and cache to switch to
another MMU table as far as the data is coherent before and after
the switching. This patch relaxes the procedure.
Signed-off-by: York Sun <york.sun at nxp.com>
CC: Alexander Graf <agraf at suse.de>
---
I found this issue while trying to change MMU table for a SPL boot.
The RAM version U-Boot was copied to memory when d-cache was enabled.
SPL code never flushed the cache (I will send another patch to fix that).
With existing code, U-Boot stops running as soon as the MMU is diabled.
With below propsed change, U-Boot continues to run. I have been in
contact with ARM support and got very useful information. However, this
switching TTBR method is "behavious that should work" but not well
verified by ARM. During my debugging, I found other minor issue which
convinced me this code wasn't exercised. I don't intend to use this
method in long term, but figure it may help others by fixing it.
arch/arm/cpu/armv8/cache.S | 44 +++++++++-----------------------------------
1 file changed, 9 insertions(+), 35 deletions(-)
diff --git a/arch/arm/cpu/armv8/cache.S b/arch/arm/cpu/armv8/cache.S
index f1deaa7..63fb112 100644
--- a/arch/arm/cpu/armv8/cache.S
+++ b/arch/arm/cpu/armv8/cache.S
@@ -171,35 +171,10 @@ ENDPROC(__asm_invalidate_l3_icache)
/*
* void __asm_switch_ttbr(ulong new_ttbr)
*
- * Safely switches to a new page table.
+ * Switches to a new page table. Cache coherency must be maintained
+ * before calling this function.
*/
ENTRY(__asm_switch_ttbr)
- /* x2 = SCTLR (alive throghout the function) */
- switch_el x4, 3f, 2f, 1f
-3: mrs x2, sctlr_el3
- b 0f
-2: mrs x2, sctlr_el2
- b 0f
-1: mrs x2, sctlr_el1
-0:
-
- /* Unset CR_M | CR_C | CR_I from SCTLR to disable all caches */
- movn x1, #(CR_M | CR_C | CR_I)
- and x1, x2, x1
- switch_el x4, 3f, 2f, 1f
-3: msr sctlr_el3, x1
- b 0f
-2: msr sctlr_el2, x1
- b 0f
-1: msr sctlr_el1, x1
-0: isb
-
- /* This call only clobbers x30 (lr) and x9 (unused) */
- mov x3, x30
- bl __asm_invalidate_tlb_all
-
- /* From here on we're running safely with caches disabled */
-
/* Set TTBR to our first argument */
switch_el x4, 3f, 2f, 1f
3: msr ttbr0_el3, x0
@@ -209,14 +184,13 @@ ENTRY(__asm_switch_ttbr)
1: msr ttbr0_el1, x0
0: isb
- /* Restore original SCTLR and thus enable caches again */
- switch_el x4, 3f, 2f, 1f
-3: msr sctlr_el3, x2
- b 0f
-2: msr sctlr_el2, x2
- b 0f
-1: msr sctlr_el1, x2
-0: isb
+ /* invalidate i-cache */
+ ic ialluis
+ isb sy
+
+ /* This call only clobbers x30 (lr) and x9 (unused) */
+ mov x3, x30
+ bl __asm_invalidate_tlb_all
ret x3
ENDPROC(__asm_switch_ttbr)
--
2.7.4
More information about the U-Boot
mailing list