2017-04-16 13:31:59 +00:00
|
|
|
int nand_spl_load_image(uint32_t offs, unsigned int size, void *dst)
|
|
|
|
{
|
|
|
|
unsigned int block, lastblock;
|
|
|
|
unsigned int page, page_offset;
|
|
|
|
|
|
|
|
/* offs has to be aligned to a page address! */
|
|
|
|
block = offs / CONFIG_SYS_NAND_BLOCK_SIZE;
|
|
|
|
lastblock = (offs + size - 1) / CONFIG_SYS_NAND_BLOCK_SIZE;
|
|
|
|
page = (offs % CONFIG_SYS_NAND_BLOCK_SIZE) / CONFIG_SYS_NAND_PAGE_SIZE;
|
|
|
|
page_offset = offs % CONFIG_SYS_NAND_PAGE_SIZE;
|
|
|
|
|
|
|
|
while (block <= lastblock) {
|
|
|
|
if (!nand_is_bad_block(block)) {
|
|
|
|
/* Skip bad blocks */
|
|
|
|
while (page < CONFIG_SYS_NAND_PAGE_COUNT) {
|
|
|
|
nand_read_page(block, page, dst);
|
|
|
|
/*
|
|
|
|
* When offs is not aligned to page address the
|
|
|
|
* extra offset is copied to dst as well. Copy
|
|
|
|
* the image such that its first byte will be
|
|
|
|
* at the dst.
|
|
|
|
*/
|
|
|
|
if (unlikely(page_offset)) {
|
|
|
|
memmove(dst, dst + page_offset,
|
|
|
|
CONFIG_SYS_NAND_PAGE_SIZE);
|
|
|
|
dst = (void *)((int)dst - page_offset);
|
|
|
|
page_offset = 0;
|
|
|
|
}
|
|
|
|
dst += CONFIG_SYS_NAND_PAGE_SIZE;
|
|
|
|
page++;
|
|
|
|
}
|
|
|
|
|
|
|
|
page = 0;
|
|
|
|
} else {
|
|
|
|
lastblock++;
|
|
|
|
}
|
|
|
|
|
|
|
|
block++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-05-27 11:56:20 +00:00
|
|
|
/**
|
|
|
|
* nand_spl_adjust_offset - Adjust offset from a starting sector
|
|
|
|
* @sector: Address of the sector
|
|
|
|
* @offs: Offset starting from @sector
|
|
|
|
*
|
|
|
|
* If one or more bad blocks are in the address space between @sector
|
|
|
|
* and @sector + @offs, @offs is increased by the NAND block size for
|
|
|
|
* each bad block found.
|
|
|
|
*/
|
|
|
|
u32 nand_spl_adjust_offset(u32 sector, u32 offs)
|
|
|
|
{
|
|
|
|
unsigned int block, lastblock;
|
|
|
|
|
|
|
|
block = sector / CONFIG_SYS_NAND_BLOCK_SIZE;
|
|
|
|
lastblock = (sector + offs) / CONFIG_SYS_NAND_BLOCK_SIZE;
|
|
|
|
|
|
|
|
while (block <= lastblock) {
|
|
|
|
if (nand_is_bad_block(block)) {
|
|
|
|
offs += CONFIG_SYS_NAND_BLOCK_SIZE;
|
|
|
|
lastblock++;
|
|
|
|
}
|
|
|
|
|
|
|
|
block++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return offs;
|
|
|
|
}
|
|
|
|
|
2017-04-16 13:31:59 +00:00
|
|
|
#ifdef CONFIG_SPL_UBI
|
|
|
|
/*
|
|
|
|
* Temporary storage for non NAND page aligned and non NAND page sized
|
|
|
|
* reads. Note: This does not support runtime detected FLASH yet, but
|
|
|
|
* that should be reasonably easy to fix by making the buffer large
|
|
|
|
* enough :)
|
|
|
|
*/
|
|
|
|
static u8 scratch_buf[CONFIG_SYS_NAND_PAGE_SIZE];
|
|
|
|
|
|
|
|
/**
|
|
|
|
* nand_spl_read_block - Read data from physical eraseblock into a buffer
|
|
|
|
* @block: Number of the physical eraseblock
|
|
|
|
* @offset: Data offset from the start of @peb
|
|
|
|
* @len: Data size to read
|
|
|
|
* @dst: Address of the destination buffer
|
|
|
|
*
|
|
|
|
* This could be further optimized if we'd have a subpage read
|
|
|
|
* function in the simple code. On NAND which allows subpage reads
|
|
|
|
* this would spare quite some time to readout e.g. the VID header of
|
|
|
|
* UBI.
|
|
|
|
*
|
|
|
|
* Notes:
|
|
|
|
* @offset + @len are not allowed to be larger than a physical
|
|
|
|
* erase block. No sanity check done for simplicity reasons.
|
|
|
|
*
|
|
|
|
* To support runtime detected flash this needs to be extended by
|
|
|
|
* information about the actual flash geometry, but thats beyond the
|
|
|
|
* scope of this effort and for most applications where fast boot is
|
|
|
|
* required it is not an issue anyway.
|
|
|
|
*/
|
|
|
|
int nand_spl_read_block(int block, int offset, int len, void *dst)
|
|
|
|
{
|
|
|
|
int page, read;
|
|
|
|
|
|
|
|
/* Calculate the page number */
|
|
|
|
page = offset / CONFIG_SYS_NAND_PAGE_SIZE;
|
|
|
|
|
|
|
|
/* Offset to the start of a flash page */
|
|
|
|
offset = offset % CONFIG_SYS_NAND_PAGE_SIZE;
|
|
|
|
|
|
|
|
while (len) {
|
|
|
|
/*
|
|
|
|
* Non page aligned reads go to the scratch buffer.
|
|
|
|
* Page aligned reads go directly to the destination.
|
|
|
|
*/
|
|
|
|
if (offset || len < CONFIG_SYS_NAND_PAGE_SIZE) {
|
|
|
|
nand_read_page(block, page, scratch_buf);
|
|
|
|
read = min(len, CONFIG_SYS_NAND_PAGE_SIZE - offset);
|
|
|
|
memcpy(dst, scratch_buf + offset, read);
|
|
|
|
offset = 0;
|
|
|
|
} else {
|
|
|
|
nand_read_page(block, page, dst);
|
|
|
|
read = CONFIG_SYS_NAND_PAGE_SIZE;
|
|
|
|
}
|
|
|
|
page++;
|
|
|
|
len -= read;
|
|
|
|
dst += read;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif
|