mirror of
https://github.com/AsahiLinux/u-boot
synced 2024-12-15 15:53:02 +00:00
c32248601c
Rename current assembler implementation of allow_unaligned() to v7_arch_cp15_allow_unaligned() and add it into armv7.h header, then add C wrapper of allow_unaligned(). This fixes misbehavior when linking U-Boot on ARMv7a i.MX6Q, where the CPU specific allow_unaligned() implementation was ignored and instead the __weak allow_unaligned() implementation from lib/efi_loader/efi_setup.c was used, which led to "data abort" just before booting Linux via tftp, in efi_dp_from_file() -> path_to_uefi() -> utf16_put() . The problem is triggerd byc7c0ca3767
("efi_loader: fix efi_dp_from_file()") . Adding the wrapper fixes the problem. Fixes:78f90aaeec
("arm: armv7: allow unaligned memory access") Signed-off-by: Marek Vasut <marex@denx.de>
162 lines
4.3 KiB
C
162 lines
4.3 KiB
C
/* SPDX-License-Identifier: GPL-2.0+ */
|
|
/*
|
|
* (C) Copyright 2010
|
|
* Texas Instruments, <www.ti.com>
|
|
* Aneesh V <aneesh@ti.com>
|
|
*/
|
|
#ifndef ARMV7_H
|
|
#define ARMV7_H
|
|
|
|
/* Cortex-A9 revisions */
|
|
#define MIDR_CORTEX_A9_R0P1 0x410FC091
|
|
#define MIDR_CORTEX_A9_R1P2 0x411FC092
|
|
#define MIDR_CORTEX_A9_R1P3 0x411FC093
|
|
#define MIDR_CORTEX_A9_R2P10 0x412FC09A
|
|
|
|
/* Cortex-A15 revisions */
|
|
#define MIDR_CORTEX_A15_R0P0 0x410FC0F0
|
|
#define MIDR_CORTEX_A15_R2P2 0x412FC0F2
|
|
|
|
/* Cortex-A7 revisions */
|
|
#define MIDR_CORTEX_A7_R0P0 0x410FC070
|
|
|
|
#define MIDR_PRIMARY_PART_MASK 0xFF0FFFF0
|
|
|
|
/* ID_PFR1 feature fields */
|
|
#define CPUID_ARM_SEC_SHIFT 4
|
|
#define CPUID_ARM_SEC_MASK (0xF << CPUID_ARM_SEC_SHIFT)
|
|
#define CPUID_ARM_VIRT_SHIFT 12
|
|
#define CPUID_ARM_VIRT_MASK (0xF << CPUID_ARM_VIRT_SHIFT)
|
|
#define CPUID_ARM_GENTIMER_SHIFT 16
|
|
#define CPUID_ARM_GENTIMER_MASK (0xF << CPUID_ARM_GENTIMER_SHIFT)
|
|
|
|
/* valid bits in CBAR register / PERIPHBASE value */
|
|
#define CBAR_MASK 0xFFFF8000
|
|
|
|
/* CCSIDR */
|
|
#define CCSIDR_LINE_SIZE_OFFSET 0
|
|
#define CCSIDR_LINE_SIZE_MASK 0x7
|
|
#define CCSIDR_ASSOCIATIVITY_OFFSET 3
|
|
#define CCSIDR_ASSOCIATIVITY_MASK (0x3FF << 3)
|
|
#define CCSIDR_NUM_SETS_OFFSET 13
|
|
#define CCSIDR_NUM_SETS_MASK (0x7FFF << 13)
|
|
|
|
/*
|
|
* Values for InD field in CSSELR
|
|
* Selects the type of cache
|
|
*/
|
|
#define ARMV7_CSSELR_IND_DATA_UNIFIED 0
|
|
#define ARMV7_CSSELR_IND_INSTRUCTION 1
|
|
|
|
/* Values for Ctype fields in CLIDR */
|
|
#define ARMV7_CLIDR_CTYPE_NO_CACHE 0
|
|
#define ARMV7_CLIDR_CTYPE_INSTRUCTION_ONLY 1
|
|
#define ARMV7_CLIDR_CTYPE_DATA_ONLY 2
|
|
#define ARMV7_CLIDR_CTYPE_INSTRUCTION_DATA 3
|
|
#define ARMV7_CLIDR_CTYPE_UNIFIED 4
|
|
|
|
#ifndef __ASSEMBLY__
|
|
#include <linux/types.h>
|
|
#include <asm/io.h>
|
|
#include <asm/barriers.h>
|
|
|
|
/* read L2 control register (L2CTLR) */
|
|
static inline uint32_t read_l2ctlr(void)
|
|
{
|
|
uint32_t val = 0;
|
|
|
|
asm volatile ("mrc p15, 1, %0, c9, c0, 2" : "=r" (val));
|
|
|
|
return val;
|
|
}
|
|
|
|
/* write L2 control register (L2CTLR) */
|
|
static inline void write_l2ctlr(uint32_t val)
|
|
{
|
|
/*
|
|
* Note: L2CTLR can only be written when the L2 memory system
|
|
* is idle, ie before the MMU is enabled.
|
|
*/
|
|
asm volatile("mcr p15, 1, %0, c9, c0, 2" : : "r" (val) : "memory");
|
|
isb();
|
|
}
|
|
|
|
/*
|
|
* Workaround for ARM errata # 798870
|
|
* Set L2ACTLR[7] to reissue any memory transaction in the L2 that has been
|
|
* stalled for 1024 cycles to verify that its hazard condition still exists.
|
|
*/
|
|
static inline void v7_enable_l2_hazard_detect(void)
|
|
{
|
|
uint32_t val;
|
|
|
|
/* L2ACTLR[7]: Enable hazard detect timeout */
|
|
asm volatile ("mrc p15, 1, %0, c15, c0, 0\n\t" : "=r"(val));
|
|
val |= (1 << 7);
|
|
asm volatile ("mcr p15, 1, %0, c15, c0, 0\n\t" : : "r"(val));
|
|
}
|
|
|
|
/*
|
|
* Workaround for ARM errata # 799270
|
|
* Ensure that the L2 logic has been used within the previous 256 cycles
|
|
* before modifying the ACTLR.SMP bit. This is required during boot before
|
|
* MMU has been enabled, or during a specified reset or power down sequence.
|
|
*/
|
|
static inline void v7_enable_smp(uint32_t address)
|
|
{
|
|
uint32_t temp, val;
|
|
|
|
/* Read auxiliary control register */
|
|
asm volatile ("mrc p15, 0, %0, c1, c0, 1\n\t" : "=r"(val));
|
|
|
|
/* Enable SMP */
|
|
val |= (1 << 6);
|
|
|
|
/* Dummy read to assure L2 access */
|
|
temp = readl(address);
|
|
temp &= 0;
|
|
val |= temp;
|
|
|
|
/* Write auxiliary control register */
|
|
asm volatile ("mcr p15, 0, %0, c1, c0, 1\n\t" : : "r"(val));
|
|
|
|
CP15DSB;
|
|
CP15ISB;
|
|
}
|
|
|
|
void v7_en_l2_hazard_detect(void);
|
|
void v7_outer_cache_enable(void);
|
|
void v7_outer_cache_disable(void);
|
|
void v7_outer_cache_flush_all(void);
|
|
void v7_outer_cache_inval_all(void);
|
|
void v7_outer_cache_flush_range(u32 start, u32 end);
|
|
void v7_outer_cache_inval_range(u32 start, u32 end);
|
|
|
|
#ifdef CONFIG_ARMV7_NONSEC
|
|
|
|
int armv7_init_nonsec(void);
|
|
int armv7_apply_memory_carveout(u64 *start, u64 *size);
|
|
bool armv7_boot_nonsec(void);
|
|
|
|
/* defined in assembly file */
|
|
unsigned int _nonsec_init(void);
|
|
void _do_nonsec_entry(void *target_pc, unsigned long r0,
|
|
unsigned long r1, unsigned long r2);
|
|
void _smp_pen(void);
|
|
|
|
extern char __secure_start[];
|
|
extern char __secure_end[];
|
|
extern char __secure_stack_start[];
|
|
extern char __secure_stack_end[];
|
|
|
|
#endif /* CONFIG_ARMV7_NONSEC */
|
|
|
|
void v7_arch_cp15_set_l2aux_ctrl(u32 l2auxctrl, u32 cpu_midr,
|
|
u32 cpu_rev_comb, u32 cpu_variant,
|
|
u32 cpu_rev);
|
|
void v7_arch_cp15_set_acr(u32 acr, u32 cpu_midr, u32 cpu_rev_comb,
|
|
u32 cpu_variant, u32 cpu_rev);
|
|
void v7_arch_cp15_allow_unaligned(void);
|
|
#endif /* ! __ASSEMBLY__ */
|
|
|
|
#endif
|