aboutsummaryrefslogtreecommitdiffstats
path: root/package/boot/uboot-layerscape/patches/0001-armv8-Support-loading-32-bit-OS-in-AArch32-execution.patch
diff options
context:
space:
mode:
Diffstat (limited to 'package/boot/uboot-layerscape/patches/0001-armv8-Support-loading-32-bit-OS-in-AArch32-execution.patch')
-rw-r--r--package/boot/uboot-layerscape/patches/0001-armv8-Support-loading-32-bit-OS-in-AArch32-execution.patch214
1 files changed, 214 insertions, 0 deletions
diff --git a/package/boot/uboot-layerscape/patches/0001-armv8-Support-loading-32-bit-OS-in-AArch32-execution.patch b/package/boot/uboot-layerscape/patches/0001-armv8-Support-loading-32-bit-OS-in-AArch32-execution.patch
new file mode 100644
index 0000000000..ace542890a
--- /dev/null
+++ b/package/boot/uboot-layerscape/patches/0001-armv8-Support-loading-32-bit-OS-in-AArch32-execution.patch
@@ -0,0 +1,214 @@
+From e03f571b1e9564ca422134b2013d80c8373c9db2 Mon Sep 17 00:00:00 2001
+From: Alison Wang <b18965@freescale.com>
+Date: Fri, 13 May 2016 10:19:24 +0800
+Subject: [PATCH 01/93] armv8: Support loading 32-bit OS in AArch32 execution
+ state
+
+To support loading a 32-bit OS, the execution state will change from
+AArch64 to AArch32 when jumping to kernel.
+
+The architecture information will be got through checking FIT
+image, then U-Boot will load 32-bit OS or 64-bit OS automatically.
+
+Signed-off-by: Ebony Zhu <ebony.zhu@nxp.com>
+Signed-off-by: Alison Wang <alison.wang@nxp.com>
+Signed-off-by: Chenhui Zhao <chenhui.zhao@nxp.com>
+---
+ arch/arm/cpu/armv8/transition.S | 100 +++++++++++++++++++++++++++++++++++++++
+ arch/arm/include/asm/system.h | 2 +
+ arch/arm/lib/bootm.c | 20 +++++++-
+ common/image-fit.c | 12 ++++-
+ 4 files changed, 131 insertions(+), 3 deletions(-)
+
+diff --git a/arch/arm/cpu/armv8/transition.S b/arch/arm/cpu/armv8/transition.S
+index 253a39b..9d7a17a 100644
+--- a/arch/arm/cpu/armv8/transition.S
++++ b/arch/arm/cpu/armv8/transition.S
+@@ -21,3 +21,103 @@ ENTRY(armv8_switch_to_el1)
+ 0: ret
+ 1: armv8_switch_to_el1_m x0, x1
+ ENDPROC(armv8_switch_to_el1)
++
++/*
++ * x0: kernel entry point
++ * x1: machine nr
++ * x2: fdt address
++ */
++ENTRY(armv8_switch_to_el2_aarch32)
++ switch_el x3, 1f, 0f, 0f
++0: ret
++1:
++ mov x7, x0
++ mov x8, x1
++ mov x9, x2
++
++ /* 32bit EL2 | HCE | SMD | RES1 (Bits[5:4]) | Non-secure EL0/EL1 */
++ mov x1, 0x1b1
++ msr scr_el3, x1
++ msr cptr_el3, xzr /* Disable coprocessor traps to EL3 */
++ mov x1, 0x33ff
++ msr cptr_el2, x1 /* Disable coprocessor traps to EL2 */
++
++ /* Initialize Generic Timers */
++ msr cntvoff_el2, xzr
++
++ mov x1, #0x0830
++ movk x1, #0x30c5, lsl #16
++ msr sctlr_el2, x1
++
++ /* Return to AArch32 Hypervisor mode */
++ mov x1, sp
++ msr sp_el2, x1
++ mrs x1, vbar_el3
++ msr vbar_el2, x1 /* Migrate VBAR */
++ mov x1, #0x1da
++ msr spsr_el3, x1
++ msr elr_el3, x7
++
++ mov x0, #0
++ mov x1, x8
++ mov x2, x9
++
++ eret
++ENDPROC(armv8_switch_to_el2_aarch32)
++
++/*
++ * x0: kernel entry point
++ * x1: machine nr
++ * x2: fdt address
++ */
++ENTRY(armv8_switch_to_el1_aarch32)
++ switch_el x3, 0f, 1f, 0f
++0: ret
++1:
++ mov x7, x0
++ mov x8, x1
++ mov x9, x2
++
++ /* Initialize Generic Timers */
++ mrs x0, cnthctl_el2
++ orr x0, x0, #0x3 /* Enable EL1 access to timers */
++ msr cnthctl_el2, x0
++ msr cntvoff_el2, xzr
++
++ /* Initialize MPID/MPIDR registers */
++ mrs x0, midr_el1
++ mrs x1, mpidr_el1
++ msr vpidr_el2, x0
++ msr vmpidr_el2, x1
++
++ /* Disable coprocessor traps */
++ mov x0, #0x33ff
++ msr cptr_el2, x0 /* Disable coprocessor traps to EL2 */
++ msr hstr_el2, xzr /* Disable coprocessor traps to EL2 */
++ mov x0, #3 << 20
++ msr cpacr_el1, x0 /* Enable FP/SIMD at EL1 */
++
++ /* Initialize HCR_EL2 */
++ mov x0, #(0 << 31) /* 32bit EL1 */
++ orr x0, x0, #(1 << 29) /* Disable HVC */
++ msr hcr_el2, x0
++
++ mov x0, #0x0800
++ movk x0, #0x30d0, lsl #16
++ msr sctlr_el1, x0
++
++ /* Return to AArch32 Supervisor mode */
++ mov x0, sp
++ msr sp_el1, x0 /* Migrate SP */
++ mrs x0, vbar_el2
++ msr vbar_el1, x0 /* Migrate VBAR */
++ mov x0, #0x1d3
++ msr spsr_el2, x0
++ msr elr_el2, x7
++
++ mov x0, #0
++ mov x1, x8
++ mov x2, x9
++
++ eret
++ENDPROC(armv8_switch_to_el1_aarch32)
+diff --git a/arch/arm/include/asm/system.h b/arch/arm/include/asm/system.h
+index 71b3108..0ecbcf7 100644
+--- a/arch/arm/include/asm/system.h
++++ b/arch/arm/include/asm/system.h
+@@ -81,6 +81,8 @@ int __asm_flush_l3_cache(void);
+
+ void armv8_switch_to_el2(void);
+ void armv8_switch_to_el1(void);
++void armv8_switch_to_el2_aarch32(u64 entry_point, u64 mach_nr, u64 fdt_addr);
++void armv8_switch_to_el1_aarch32(u64 entry_point, u64 mach_nr, u64 fdt_addr);
+ void gic_init(void);
+ void gic_send_sgi(unsigned long sgino);
+ void wait_for_wakeup(void);
+diff --git a/arch/arm/lib/bootm.c b/arch/arm/lib/bootm.c
+index a477cae..36f2cb0 100644
+--- a/arch/arm/lib/bootm.c
++++ b/arch/arm/lib/bootm.c
+@@ -276,8 +276,24 @@ static void boot_jump_linux(bootm_headers_t *images, int flag)
+ announce_and_cleanup(fake);
+
+ if (!fake) {
+- do_nonsec_virt_switch();
+- kernel_entry(images->ft_addr, NULL, NULL, NULL);
++ if ((IH_ARCH_DEFAULT == IH_ARCH_ARM64) &&
++ (images->os.arch == IH_ARCH_ARM)) {
++ smp_kick_all_cpus();
++ dcache_disable();
++#ifdef CONFIG_ARMV8_SWITCH_TO_EL1
++ armv8_switch_to_el2();
++ armv8_switch_to_el1_aarch32((u64)images->ep,
++ (u64)gd->bd->bi_arch_number,
++ (u64)images->ft_addr);
++#else
++ armv8_switch_to_el2_aarch32((u64)images->ep,
++ (u64)gd->bd->bi_arch_number,
++ (u64)images->ft_addr);
++#endif
++ } else {
++ do_nonsec_virt_switch();
++ kernel_entry(images->ft_addr, NULL, NULL, NULL);
++ }
+ }
+ #else
+ unsigned long machid = gd->bd->bi_arch_number;
+diff --git a/common/image-fit.c b/common/image-fit.c
+index c531ee7..0d54f71 100644
+--- a/common/image-fit.c
++++ b/common/image-fit.c
+@@ -1144,7 +1144,8 @@ int fit_image_check_arch(const void *fit, int noffset, uint8_t arch)
+ if (fit_image_get_arch(fit, noffset, &image_arch))
+ return 0;
+ return (arch == image_arch) ||
+- (arch == IH_ARCH_I386 && image_arch == IH_ARCH_X86_64);
++ (arch == IH_ARCH_I386 && image_arch == IH_ARCH_X86_64) ||
++ (arch == IH_ARCH_ARM64 && image_arch == IH_ARCH_ARM);
+ }
+
+ /**
+@@ -1567,6 +1568,9 @@ int fit_image_load(bootm_headers_t *images, ulong addr,
+ int type_ok, os_ok;
+ ulong load, data, len;
+ uint8_t os;
++#ifndef USE_HOSTCC
++ uint8_t os_arch;
++#endif
+ const char *prop_name;
+ int ret;
+
+@@ -1650,6 +1654,12 @@ int fit_image_load(bootm_headers_t *images, ulong addr,
+ return -ENOEXEC;
+ }
+ #endif
++
++#ifndef USE_HOSTCC
++ fit_image_get_arch(fit, noffset, &os_arch);
++ images->os.arch = os_arch;
++#endif
++
+ if (image_type == IH_TYPE_FLATDT &&
+ !fit_image_check_comp(fit, noffset, IH_COMP_NONE)) {
+ puts("FDT image is compressed");
+--
+1.7.9.5
+