2018-05-06 21:58:06 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0+
|
2012-08-26 15:19:06 +00:00
|
|
|
/*
|
|
|
|
* Generic bounce buffer implementation
|
|
|
|
*
|
|
|
|
* Copyright (C) 2012 Marek Vasut <marex@denx.de>
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <common.h>
|
|
|
|
#include <malloc.h>
|
|
|
|
#include <errno.h>
|
|
|
|
#include <bouncebuf.h>
|
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
static int addr_aligned(struct bounce_buffer *state)
|
2012-08-26 15:19:06 +00:00
|
|
|
{
|
|
|
|
const ulong align_mask = ARCH_DMA_MINALIGN - 1;
|
|
|
|
|
|
|
|
/* Check if start is aligned */
|
2012-11-06 11:27:29 +00:00
|
|
|
if ((ulong)state->user_buffer & align_mask) {
|
|
|
|
debug("Unaligned buffer address %p\n", state->user_buffer);
|
2012-08-26 15:19:06 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
/* Check if length is aligned */
|
|
|
|
if (state->len != state->len_aligned) {
|
2014-08-26 10:45:48 +00:00
|
|
|
debug("Unaligned buffer length %zu\n", state->len);
|
2012-08-26 15:19:06 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Aligned */
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
int bounce_buffer_start(struct bounce_buffer *state, void *data,
|
|
|
|
size_t len, unsigned int flags)
|
2012-08-26 15:19:06 +00:00
|
|
|
{
|
2012-11-06 11:27:29 +00:00
|
|
|
state->user_buffer = data;
|
|
|
|
state->bounce_buffer = data;
|
|
|
|
state->len = len;
|
|
|
|
state->len_aligned = roundup(len, ARCH_DMA_MINALIGN);
|
|
|
|
state->flags = flags;
|
|
|
|
|
|
|
|
if (!addr_aligned(state)) {
|
|
|
|
state->bounce_buffer = memalign(ARCH_DMA_MINALIGN,
|
|
|
|
state->len_aligned);
|
|
|
|
if (!state->bounce_buffer)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
if (state->flags & GEN_BB_READ)
|
|
|
|
memcpy(state->bounce_buffer, state->user_buffer,
|
|
|
|
state->len);
|
2012-08-26 15:19:06 +00:00
|
|
|
}
|
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
/*
|
|
|
|
* Flush data to RAM so DMA reads can pick it up,
|
|
|
|
* and any CPU writebacks don't race with DMA writes
|
|
|
|
*/
|
|
|
|
flush_dcache_range((unsigned long)state->bounce_buffer,
|
|
|
|
(unsigned long)(state->bounce_buffer) +
|
|
|
|
state->len_aligned);
|
2012-08-26 15:19:06 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
int bounce_buffer_stop(struct bounce_buffer *state)
|
2012-08-26 15:19:06 +00:00
|
|
|
{
|
2012-11-06 11:27:29 +00:00
|
|
|
if (state->flags & GEN_BB_WRITE) {
|
|
|
|
/* Invalidate cache so that CPU can see any newly DMA'd data */
|
|
|
|
invalidate_dcache_range((unsigned long)state->bounce_buffer,
|
|
|
|
(unsigned long)(state->bounce_buffer) +
|
|
|
|
state->len_aligned);
|
|
|
|
}
|
2012-08-26 15:19:06 +00:00
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
if (state->bounce_buffer == state->user_buffer)
|
2012-08-26 15:19:06 +00:00
|
|
|
return 0;
|
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
if (state->flags & GEN_BB_WRITE)
|
|
|
|
memcpy(state->user_buffer, state->bounce_buffer, state->len);
|
2012-08-26 15:19:06 +00:00
|
|
|
|
2012-11-06 11:27:29 +00:00
|
|
|
free(state->bounce_buffer);
|
2012-08-26 15:19:06 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|