mirror of
https://github.com/AsahiLinux/u-boot
synced 2024-11-12 16:07:30 +00:00
e731a5385d
If CONFIG_UNIPHIER_L2CACHE_ON is undefined, the L2 cache is never enabled, so there is no need for v7_outer_cache_disable(). The weak stub avoids the compile error anyway. Signed-off-by: Masahiro Yamada <yamada.masahiro@socionext.com>
200 lines
4.8 KiB
C
200 lines
4.8 KiB
C
/*
|
|
* Copyright (C) 2012-2014 Panasonic Corporation
|
|
* Copyright (C) 2015-2016 Socionext Inc.
|
|
* Author: Masahiro Yamada <yamada.masahiro@socionext.com>
|
|
*
|
|
* SPDX-License-Identifier: GPL-2.0+
|
|
*/
|
|
|
|
#include <common.h>
|
|
#include <linux/io.h>
|
|
#include <linux/kernel.h>
|
|
#include <asm/armv7.h>
|
|
#include <asm/processor.h>
|
|
|
|
#include "cache-uniphier.h"
|
|
#include "ssc-regs.h"
|
|
|
|
#define UNIPHIER_SSCOQAD_IS_NEEDED(op) \
|
|
((op & UNIPHIER_SSCOQM_S_MASK) == UNIPHIER_SSCOQM_S_RANGE)
|
|
#define UNIPHIER_SSCOQWM_IS_NEEDED(op) \
|
|
((op & UNIPHIER_SSCOQM_TID_MASK) == UNIPHIER_SSCOQM_TID_WAY)
|
|
|
|
/* uniphier_cache_sync - perform a sync point for a particular cache level */
|
|
static void uniphier_cache_sync(void)
|
|
{
|
|
/* drain internal buffers */
|
|
writel(UNIPHIER_SSCOPE_CM_SYNC, UNIPHIER_SSCOPE);
|
|
/* need a read back to confirm */
|
|
readl(UNIPHIER_SSCOPE);
|
|
}
|
|
|
|
/**
|
|
* uniphier_cache_maint_common - run a queue operation
|
|
*
|
|
* @start: start address of range operation (don't care for "all" operation)
|
|
* @size: data size of range operation (don't care for "all" operation)
|
|
* @ways: target ways (don't care for operations other than pre-fetch, touch
|
|
* @operation: flags to specify the desired cache operation
|
|
*/
|
|
static void uniphier_cache_maint_common(u32 start, u32 size, u32 ways,
|
|
u32 operation)
|
|
{
|
|
/* clear the complete notification flag */
|
|
writel(UNIPHIER_SSCOLPQS_EF, UNIPHIER_SSCOLPQS);
|
|
|
|
do {
|
|
/* set cache operation */
|
|
writel(UNIPHIER_SSCOQM_CE | operation, UNIPHIER_SSCOQM);
|
|
|
|
/* set address range if needed */
|
|
if (likely(UNIPHIER_SSCOQAD_IS_NEEDED(operation))) {
|
|
writel(start, UNIPHIER_SSCOQAD);
|
|
writel(size, UNIPHIER_SSCOQSZ);
|
|
}
|
|
|
|
/* set target ways if needed */
|
|
if (unlikely(UNIPHIER_SSCOQWM_IS_NEEDED(operation)))
|
|
writel(ways, UNIPHIER_SSCOQWN);
|
|
} while (unlikely(readl(UNIPHIER_SSCOPPQSEF) &
|
|
(UNIPHIER_SSCOPPQSEF_FE | UNIPHIER_SSCOPPQSEF_OE)));
|
|
|
|
/* wait until the operation is completed */
|
|
while (likely(readl(UNIPHIER_SSCOLPQS) != UNIPHIER_SSCOLPQS_EF))
|
|
cpu_relax();
|
|
}
|
|
|
|
static void uniphier_cache_maint_all(u32 operation)
|
|
{
|
|
uniphier_cache_maint_common(0, 0, 0, UNIPHIER_SSCOQM_S_ALL | operation);
|
|
|
|
uniphier_cache_sync();
|
|
}
|
|
|
|
static void uniphier_cache_maint_range(u32 start, u32 end, u32 ways,
|
|
u32 operation)
|
|
{
|
|
u32 size;
|
|
|
|
/*
|
|
* If the start address is not aligned,
|
|
* perform a cache operation for the first cache-line
|
|
*/
|
|
start = start & ~(UNIPHIER_SSC_LINE_SIZE - 1);
|
|
|
|
size = end - start;
|
|
|
|
if (unlikely(size >= (u32)(-UNIPHIER_SSC_LINE_SIZE))) {
|
|
/* this means cache operation for all range */
|
|
uniphier_cache_maint_all(operation);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* If the end address is not aligned,
|
|
* perform a cache operation for the last cache-line
|
|
*/
|
|
size = ALIGN(size, UNIPHIER_SSC_LINE_SIZE);
|
|
|
|
while (size) {
|
|
u32 chunk_size = min_t(u32, size, UNIPHIER_SSC_RANGE_OP_MAX_SIZE);
|
|
|
|
uniphier_cache_maint_common(start, chunk_size, ways,
|
|
UNIPHIER_SSCOQM_S_RANGE | operation);
|
|
|
|
start += chunk_size;
|
|
size -= chunk_size;
|
|
}
|
|
|
|
uniphier_cache_sync();
|
|
}
|
|
|
|
void uniphier_cache_prefetch_range(u32 start, u32 end, u32 ways)
|
|
{
|
|
uniphier_cache_maint_range(start, end, ways,
|
|
UNIPHIER_SSCOQM_TID_WAY |
|
|
UNIPHIER_SSCOQM_CM_PREFETCH);
|
|
}
|
|
|
|
void uniphier_cache_touch_range(u32 start, u32 end, u32 ways)
|
|
{
|
|
uniphier_cache_maint_range(start, end, ways,
|
|
UNIPHIER_SSCOQM_TID_WAY |
|
|
UNIPHIER_SSCOQM_CM_TOUCH);
|
|
}
|
|
|
|
void uniphier_cache_touch_zero_range(u32 start, u32 end, u32 ways)
|
|
{
|
|
uniphier_cache_maint_range(start, end, ways,
|
|
UNIPHIER_SSCOQM_TID_WAY |
|
|
UNIPHIER_SSCOQM_CM_TOUCH_ZERO);
|
|
}
|
|
|
|
#ifdef CONFIG_UNIPHIER_L2CACHE_ON
|
|
void v7_outer_cache_flush_all(void)
|
|
{
|
|
uniphier_cache_maint_all(UNIPHIER_SSCOQM_CM_FLUSH);
|
|
}
|
|
|
|
void v7_outer_cache_inval_all(void)
|
|
{
|
|
uniphier_cache_maint_all(UNIPHIER_SSCOQM_CM_INV);
|
|
}
|
|
|
|
void v7_outer_cache_flush_range(u32 start, u32 end)
|
|
{
|
|
uniphier_cache_maint_range(start, end, 0, UNIPHIER_SSCOQM_CM_FLUSH);
|
|
}
|
|
|
|
void v7_outer_cache_inval_range(u32 start, u32 end)
|
|
{
|
|
if (start & (UNIPHIER_SSC_LINE_SIZE - 1)) {
|
|
start &= ~(UNIPHIER_SSC_LINE_SIZE - 1);
|
|
uniphier_cache_maint_range(start, UNIPHIER_SSC_LINE_SIZE, 0,
|
|
UNIPHIER_SSCOQM_CM_FLUSH);
|
|
start += UNIPHIER_SSC_LINE_SIZE;
|
|
}
|
|
|
|
if (start >= end) {
|
|
uniphier_cache_sync();
|
|
return;
|
|
}
|
|
|
|
if (end & (UNIPHIER_SSC_LINE_SIZE - 1)) {
|
|
end &= ~(UNIPHIER_SSC_LINE_SIZE - 1);
|
|
uniphier_cache_maint_range(end, UNIPHIER_SSC_LINE_SIZE, 0,
|
|
UNIPHIER_SSCOQM_CM_FLUSH);
|
|
}
|
|
|
|
if (start >= end) {
|
|
uniphier_cache_sync();
|
|
return;
|
|
}
|
|
|
|
uniphier_cache_maint_range(start, end, 0, UNIPHIER_SSCOQM_CM_INV);
|
|
}
|
|
|
|
void v7_outer_cache_enable(void)
|
|
{
|
|
u32 tmp;
|
|
|
|
writel(U32_MAX, UNIPHIER_SSCLPDAWCR); /* activate all ways */
|
|
tmp = readl(UNIPHIER_SSCC);
|
|
tmp |= UNIPHIER_SSCC_ON;
|
|
writel(tmp, UNIPHIER_SSCC);
|
|
}
|
|
|
|
void v7_outer_cache_disable(void)
|
|
{
|
|
u32 tmp;
|
|
|
|
tmp = readl(UNIPHIER_SSCC);
|
|
tmp &= ~UNIPHIER_SSCC_ON;
|
|
writel(tmp, UNIPHIER_SSCC);
|
|
}
|
|
#endif
|
|
|
|
void enable_caches(void)
|
|
{
|
|
dcache_enable();
|
|
}
|