mirror of
https://github.com/AsahiLinux/u-boot
synced 2024-11-14 17:07:38 +00:00
bf776679a7
At present patman sets the python path on startup so that it can access the libraries it needs. If we convert to use absolute imports this is not necessary. Move patman to use absolute imports. This requires changes in tools which use the patman libraries (which is most of them). Signed-off-by: Simon Glass <sjg@chromium.org>
887 lines
32 KiB
Python
887 lines
32 KiB
Python
# SPDX-License-Identifier: GPL-2.0+
|
|
# Copyright 2019 Google LLC
|
|
# Written by Simon Glass <sjg@chromium.org>
|
|
|
|
"""Support for coreboot's CBFS format
|
|
|
|
CBFS supports a header followed by a number of files, generally targeted at SPI
|
|
flash.
|
|
|
|
The format is somewhat defined by documentation in the coreboot tree although
|
|
it is necessary to rely on the C structures and source code (mostly cbfstool)
|
|
to fully understand it.
|
|
|
|
Currently supported: raw and stage types with compression, padding empty areas
|
|
with empty files, fixed-offset files
|
|
"""
|
|
|
|
from collections import OrderedDict
|
|
import io
|
|
import struct
|
|
import sys
|
|
|
|
from binman import elf
|
|
from patman import command
|
|
from patman import tools
|
|
|
|
# Set to True to enable printing output while working
|
|
DEBUG = False
|
|
|
|
# Set to True to enable output from running cbfstool for debugging
|
|
VERBOSE = False
|
|
|
|
# The master header, at the start of the CBFS
|
|
HEADER_FORMAT = '>IIIIIIII'
|
|
HEADER_LEN = 0x20
|
|
HEADER_MAGIC = 0x4f524243
|
|
HEADER_VERSION1 = 0x31313131
|
|
HEADER_VERSION2 = 0x31313132
|
|
|
|
# The file header, at the start of each file in the CBFS
|
|
FILE_HEADER_FORMAT = b'>8sIIII'
|
|
FILE_HEADER_LEN = 0x18
|
|
FILE_MAGIC = b'LARCHIVE'
|
|
FILENAME_ALIGN = 16 # Filename lengths are aligned to this
|
|
|
|
# A stage header containing information about 'stage' files
|
|
# Yes this is correct: this header is in litte-endian format
|
|
STAGE_FORMAT = '<IQQII'
|
|
STAGE_LEN = 0x1c
|
|
|
|
# An attribute describring the compression used in a file
|
|
ATTR_COMPRESSION_FORMAT = '>IIII'
|
|
ATTR_COMPRESSION_LEN = 0x10
|
|
|
|
# Attribute tags
|
|
# Depending on how the header was initialised, it may be backed with 0x00 or
|
|
# 0xff. Support both.
|
|
FILE_ATTR_TAG_UNUSED = 0
|
|
FILE_ATTR_TAG_UNUSED2 = 0xffffffff
|
|
FILE_ATTR_TAG_COMPRESSION = 0x42435a4c
|
|
FILE_ATTR_TAG_HASH = 0x68736148
|
|
FILE_ATTR_TAG_POSITION = 0x42435350 # PSCB
|
|
FILE_ATTR_TAG_ALIGNMENT = 0x42434c41 # ALCB
|
|
FILE_ATTR_TAG_PADDING = 0x47444150 # PDNG
|
|
|
|
# This is 'the size of bootblock reserved in firmware image (cbfs.txt)'
|
|
# Not much more info is available, but we set it to 4, due to this comment in
|
|
# cbfstool.c:
|
|
# This causes 4 bytes to be left out at the end of the image, for two reasons:
|
|
# 1. The cbfs master header pointer resides there
|
|
# 2. Ssme cbfs implementations assume that an image that resides below 4GB has
|
|
# a bootblock and get confused when the end of the image is at 4GB == 0.
|
|
MIN_BOOTBLOCK_SIZE = 4
|
|
|
|
# Files start aligned to this boundary in the CBFS
|
|
ENTRY_ALIGN = 0x40
|
|
|
|
# CBFSs must declare an architecture since much of the logic is designed with
|
|
# x86 in mind. The effect of setting this value is not well documented, but in
|
|
# general x86 is used and this makes use of a boot block and an image that ends
|
|
# at the end of 32-bit address space.
|
|
ARCHITECTURE_UNKNOWN = 0xffffffff
|
|
ARCHITECTURE_X86 = 0x00000001
|
|
ARCHITECTURE_ARM = 0x00000010
|
|
ARCHITECTURE_AARCH64 = 0x0000aa64
|
|
ARCHITECTURE_MIPS = 0x00000100
|
|
ARCHITECTURE_RISCV = 0xc001d0de
|
|
ARCHITECTURE_PPC64 = 0x407570ff
|
|
|
|
ARCH_NAMES = {
|
|
ARCHITECTURE_UNKNOWN : 'unknown',
|
|
ARCHITECTURE_X86 : 'x86',
|
|
ARCHITECTURE_ARM : 'arm',
|
|
ARCHITECTURE_AARCH64 : 'arm64',
|
|
ARCHITECTURE_MIPS : 'mips',
|
|
ARCHITECTURE_RISCV : 'riscv',
|
|
ARCHITECTURE_PPC64 : 'ppc64',
|
|
}
|
|
|
|
# File types. Only supported ones are included here
|
|
TYPE_CBFSHEADER = 0x02 # Master header, HEADER_FORMAT
|
|
TYPE_STAGE = 0x10 # Stage, holding an executable, see STAGE_FORMAT
|
|
TYPE_RAW = 0x50 # Raw file, possibly compressed
|
|
TYPE_EMPTY = 0xffffffff # Empty data
|
|
|
|
# Compression types
|
|
COMPRESS_NONE, COMPRESS_LZMA, COMPRESS_LZ4 = range(3)
|
|
|
|
COMPRESS_NAMES = {
|
|
COMPRESS_NONE : 'none',
|
|
COMPRESS_LZMA : 'lzma',
|
|
COMPRESS_LZ4 : 'lz4',
|
|
}
|
|
|
|
def find_arch(find_name):
|
|
"""Look up an architecture name
|
|
|
|
Args:
|
|
find_name: Architecture name to find
|
|
|
|
Returns:
|
|
ARCHITECTURE_... value or None if not found
|
|
"""
|
|
for arch, name in ARCH_NAMES.items():
|
|
if name == find_name:
|
|
return arch
|
|
return None
|
|
|
|
def find_compress(find_name):
|
|
"""Look up a compression algorithm name
|
|
|
|
Args:
|
|
find_name: Compression algorithm name to find
|
|
|
|
Returns:
|
|
COMPRESS_... value or None if not found
|
|
"""
|
|
for compress, name in COMPRESS_NAMES.items():
|
|
if name == find_name:
|
|
return compress
|
|
return None
|
|
|
|
def compress_name(compress):
|
|
"""Look up the name of a compression algorithm
|
|
|
|
Args:
|
|
compress: Compression algorithm number to find (COMPRESS_...)
|
|
|
|
Returns:
|
|
Compression algorithm name (string)
|
|
|
|
Raises:
|
|
KeyError if the algorithm number is invalid
|
|
"""
|
|
return COMPRESS_NAMES[compress]
|
|
|
|
def align_int(val, align):
|
|
"""Align a value up to the given alignment
|
|
|
|
Args:
|
|
val: Integer value to align
|
|
align: Integer alignment value (e.g. 4 to align to 4-byte boundary)
|
|
|
|
Returns:
|
|
integer value aligned to the required boundary, rounding up if necessary
|
|
"""
|
|
return int((val + align - 1) / align) * align
|
|
|
|
def align_int_down(val, align):
|
|
"""Align a value down to the given alignment
|
|
|
|
Args:
|
|
val: Integer value to align
|
|
align: Integer alignment value (e.g. 4 to align to 4-byte boundary)
|
|
|
|
Returns:
|
|
integer value aligned to the required boundary, rounding down if
|
|
necessary
|
|
"""
|
|
return int(val / align) * align
|
|
|
|
def _pack_string(instr):
|
|
"""Pack a string to the required aligned size by adding padding
|
|
|
|
Args:
|
|
instr: String to process
|
|
|
|
Returns:
|
|
String with required padding (at least one 0x00 byte) at the end
|
|
"""
|
|
val = tools.ToBytes(instr)
|
|
pad_len = align_int(len(val) + 1, FILENAME_ALIGN)
|
|
return val + tools.GetBytes(0, pad_len - len(val))
|
|
|
|
|
|
class CbfsFile(object):
|
|
"""Class to represent a single CBFS file
|
|
|
|
This is used to hold the information about a file, including its contents.
|
|
Use the get_data_and_offset() method to obtain the raw output for writing to
|
|
CBFS.
|
|
|
|
Properties:
|
|
name: Name of file
|
|
offset: Offset of file data from start of file header
|
|
cbfs_offset: Offset of file data in bytes from start of CBFS, or None to
|
|
place this file anyway
|
|
data: Contents of file, uncompressed
|
|
orig_data: Original data added to the file, possibly compressed
|
|
data_len: Length of (possibly compressed) data in bytes
|
|
ftype: File type (TYPE_...)
|
|
compression: Compression type (COMPRESS_...)
|
|
memlen: Length of data in memory, i.e. the uncompressed length, None if
|
|
no compression algortihm is selected
|
|
load: Load address in memory if known, else None
|
|
entry: Entry address in memory if known, else None. This is where
|
|
execution starts after the file is loaded
|
|
base_address: Base address to use for 'stage' files
|
|
erase_byte: Erase byte to use for padding between the file header and
|
|
contents (used for empty files)
|
|
size: Size of the file in bytes (used for empty files)
|
|
"""
|
|
def __init__(self, name, ftype, data, cbfs_offset, compress=COMPRESS_NONE):
|
|
self.name = name
|
|
self.offset = None
|
|
self.cbfs_offset = cbfs_offset
|
|
self.data = data
|
|
self.orig_data = data
|
|
self.ftype = ftype
|
|
self.compress = compress
|
|
self.memlen = None
|
|
self.load = None
|
|
self.entry = None
|
|
self.base_address = None
|
|
self.data_len = len(data)
|
|
self.erase_byte = None
|
|
self.size = None
|
|
|
|
def decompress(self):
|
|
"""Handle decompressing data if necessary"""
|
|
indata = self.data
|
|
if self.compress == COMPRESS_LZ4:
|
|
data = tools.Decompress(indata, 'lz4', with_header=False)
|
|
elif self.compress == COMPRESS_LZMA:
|
|
data = tools.Decompress(indata, 'lzma', with_header=False)
|
|
else:
|
|
data = indata
|
|
self.memlen = len(data)
|
|
self.data = data
|
|
self.data_len = len(indata)
|
|
|
|
@classmethod
|
|
def stage(cls, base_address, name, data, cbfs_offset):
|
|
"""Create a new stage file
|
|
|
|
Args:
|
|
base_address: Int base address for memory-mapping of ELF file
|
|
name: String file name to put in CBFS (does not need to correspond
|
|
to the name that the file originally came from)
|
|
data: Contents of file
|
|
cbfs_offset: Offset of file data in bytes from start of CBFS, or
|
|
None to place this file anyway
|
|
|
|
Returns:
|
|
CbfsFile object containing the file information
|
|
"""
|
|
cfile = CbfsFile(name, TYPE_STAGE, data, cbfs_offset)
|
|
cfile.base_address = base_address
|
|
return cfile
|
|
|
|
@classmethod
|
|
def raw(cls, name, data, cbfs_offset, compress):
|
|
"""Create a new raw file
|
|
|
|
Args:
|
|
name: String file name to put in CBFS (does not need to correspond
|
|
to the name that the file originally came from)
|
|
data: Contents of file
|
|
cbfs_offset: Offset of file data in bytes from start of CBFS, or
|
|
None to place this file anyway
|
|
compress: Compression algorithm to use (COMPRESS_...)
|
|
|
|
Returns:
|
|
CbfsFile object containing the file information
|
|
"""
|
|
return CbfsFile(name, TYPE_RAW, data, cbfs_offset, compress)
|
|
|
|
@classmethod
|
|
def empty(cls, space_to_use, erase_byte):
|
|
"""Create a new empty file of a given size
|
|
|
|
Args:
|
|
space_to_use:: Size of available space, which must be at least as
|
|
large as the alignment size for this CBFS
|
|
erase_byte: Byte to use for contents of file (repeated through the
|
|
whole file)
|
|
|
|
Returns:
|
|
CbfsFile object containing the file information
|
|
"""
|
|
cfile = CbfsFile('', TYPE_EMPTY, b'', None)
|
|
cfile.size = space_to_use - FILE_HEADER_LEN - FILENAME_ALIGN
|
|
cfile.erase_byte = erase_byte
|
|
return cfile
|
|
|
|
def calc_start_offset(self):
|
|
"""Check if this file needs to start at a particular offset in CBFS
|
|
|
|
Returns:
|
|
None if the file can be placed anywhere, or
|
|
the largest offset where the file could start (integer)
|
|
"""
|
|
if self.cbfs_offset is None:
|
|
return None
|
|
return self.cbfs_offset - self.get_header_len()
|
|
|
|
def get_header_len(self):
|
|
"""Get the length of headers required for a file
|
|
|
|
This is the minimum length required before the actual data for this file
|
|
could start. It might start later if there is padding.
|
|
|
|
Returns:
|
|
Total length of all non-data fields, in bytes
|
|
"""
|
|
name = _pack_string(self.name)
|
|
hdr_len = len(name) + FILE_HEADER_LEN
|
|
if self.ftype == TYPE_STAGE:
|
|
pass
|
|
elif self.ftype == TYPE_RAW:
|
|
hdr_len += ATTR_COMPRESSION_LEN
|
|
elif self.ftype == TYPE_EMPTY:
|
|
pass
|
|
else:
|
|
raise ValueError('Unknown file type %#x\n' % self.ftype)
|
|
return hdr_len
|
|
|
|
def get_data_and_offset(self, offset=None, pad_byte=None):
|
|
"""Obtain the contents of the file, in CBFS format and the offset of
|
|
the data within the file
|
|
|
|
Returns:
|
|
tuple:
|
|
bytes representing the contents of this file, packed and aligned
|
|
for directly inserting into the final CBFS output
|
|
offset to the file data from the start of the returned data.
|
|
"""
|
|
name = _pack_string(self.name)
|
|
hdr_len = len(name) + FILE_HEADER_LEN
|
|
attr_pos = 0
|
|
content = b''
|
|
attr = b''
|
|
pad = b''
|
|
data = self.data
|
|
if self.ftype == TYPE_STAGE:
|
|
elf_data = elf.DecodeElf(data, self.base_address)
|
|
content = struct.pack(STAGE_FORMAT, self.compress,
|
|
elf_data.entry, elf_data.load,
|
|
len(elf_data.data), elf_data.memsize)
|
|
data = elf_data.data
|
|
elif self.ftype == TYPE_RAW:
|
|
orig_data = data
|
|
if self.compress == COMPRESS_LZ4:
|
|
data = tools.Compress(orig_data, 'lz4', with_header=False)
|
|
elif self.compress == COMPRESS_LZMA:
|
|
data = tools.Compress(orig_data, 'lzma', with_header=False)
|
|
self.memlen = len(orig_data)
|
|
self.data_len = len(data)
|
|
attr = struct.pack(ATTR_COMPRESSION_FORMAT,
|
|
FILE_ATTR_TAG_COMPRESSION, ATTR_COMPRESSION_LEN,
|
|
self.compress, self.memlen)
|
|
elif self.ftype == TYPE_EMPTY:
|
|
data = tools.GetBytes(self.erase_byte, self.size)
|
|
else:
|
|
raise ValueError('Unknown type %#x when writing\n' % self.ftype)
|
|
if attr:
|
|
attr_pos = hdr_len
|
|
hdr_len += len(attr)
|
|
if self.cbfs_offset is not None:
|
|
pad_len = self.cbfs_offset - offset - hdr_len
|
|
if pad_len < 0: # pragma: no cover
|
|
# Test coverage of this is not available since this should never
|
|
# happen. It indicates that get_header_len() provided an
|
|
# incorrect value (too small) so that we decided that we could
|
|
# put this file at the requested place, but in fact a previous
|
|
# file extends far enough into the CBFS that this is not
|
|
# possible.
|
|
raise ValueError("Internal error: CBFS file '%s': Requested offset %#x but current output position is %#x" %
|
|
(self.name, self.cbfs_offset, offset))
|
|
pad = tools.GetBytes(pad_byte, pad_len)
|
|
hdr_len += pad_len
|
|
|
|
# This is the offset of the start of the file's data,
|
|
size = len(content) + len(data)
|
|
hdr = struct.pack(FILE_HEADER_FORMAT, FILE_MAGIC, size,
|
|
self.ftype, attr_pos, hdr_len)
|
|
|
|
# Do a sanity check of the get_header_len() function, to ensure that it
|
|
# stays in lockstep with this function
|
|
expected_len = self.get_header_len()
|
|
actual_len = len(hdr + name + attr)
|
|
if expected_len != actual_len: # pragma: no cover
|
|
# Test coverage of this is not available since this should never
|
|
# happen. It probably indicates that get_header_len() is broken.
|
|
raise ValueError("Internal error: CBFS file '%s': Expected headers of %#x bytes, got %#d" %
|
|
(self.name, expected_len, actual_len))
|
|
return hdr + name + attr + pad + content + data, hdr_len
|
|
|
|
|
|
class CbfsWriter(object):
|
|
"""Class to handle writing a Coreboot File System (CBFS)
|
|
|
|
Usage is something like:
|
|
|
|
cbw = CbfsWriter(size)
|
|
cbw.add_file_raw('u-boot', tools.ReadFile('u-boot.bin'))
|
|
...
|
|
data, cbfs_offset = cbw.get_data_and_offset()
|
|
|
|
Attributes:
|
|
_master_name: Name of the file containing the master header
|
|
_size: Size of the filesystem, in bytes
|
|
_files: Ordered list of files in the CBFS, each a CbfsFile
|
|
_arch: Architecture of the CBFS (ARCHITECTURE_...)
|
|
_bootblock_size: Size of the bootblock, typically at the end of the CBFS
|
|
_erase_byte: Byte to use for empty space in the CBFS
|
|
_align: Alignment to use for files, typically ENTRY_ALIGN
|
|
_base_address: Boot block offset in bytes from the start of CBFS.
|
|
Typically this is located at top of the CBFS. It is 0 when there is
|
|
no boot block
|
|
_header_offset: Offset of master header in bytes from start of CBFS
|
|
_contents_offset: Offset of first file header
|
|
_hdr_at_start: True if the master header is at the start of the CBFS,
|
|
instead of the end as normal for x86
|
|
_add_fileheader: True to add a fileheader around the master header
|
|
"""
|
|
def __init__(self, size, arch=ARCHITECTURE_X86):
|
|
"""Set up a new CBFS
|
|
|
|
This sets up all properties to default values. Files can be added using
|
|
add_file_raw(), etc.
|
|
|
|
Args:
|
|
size: Size of CBFS in bytes
|
|
arch: Architecture to declare for CBFS
|
|
"""
|
|
self._master_name = 'cbfs master header'
|
|
self._size = size
|
|
self._files = OrderedDict()
|
|
self._arch = arch
|
|
self._bootblock_size = 0
|
|
self._erase_byte = 0xff
|
|
self._align = ENTRY_ALIGN
|
|
self._add_fileheader = False
|
|
if self._arch == ARCHITECTURE_X86:
|
|
# Allow 4 bytes for the header pointer. That holds the
|
|
# twos-compliment negative offset of the master header in bytes
|
|
# measured from one byte past the end of the CBFS
|
|
self._base_address = self._size - max(self._bootblock_size,
|
|
MIN_BOOTBLOCK_SIZE)
|
|
self._header_offset = self._base_address - HEADER_LEN
|
|
self._contents_offset = 0
|
|
self._hdr_at_start = False
|
|
else:
|
|
# For non-x86, different rules apply
|
|
self._base_address = 0
|
|
self._header_offset = align_int(self._base_address +
|
|
self._bootblock_size, 4)
|
|
self._contents_offset = align_int(self._header_offset +
|
|
FILE_HEADER_LEN +
|
|
self._bootblock_size, self._align)
|
|
self._hdr_at_start = True
|
|
|
|
def _skip_to(self, fd, offset):
|
|
"""Write out pad bytes until a given offset
|
|
|
|
Args:
|
|
fd: File objext to write to
|
|
offset: Offset to write to
|
|
"""
|
|
if fd.tell() > offset:
|
|
raise ValueError('No space for data before offset %#x (current offset %#x)' %
|
|
(offset, fd.tell()))
|
|
fd.write(tools.GetBytes(self._erase_byte, offset - fd.tell()))
|
|
|
|
def _pad_to(self, fd, offset):
|
|
"""Write out pad bytes and/or an empty file until a given offset
|
|
|
|
Args:
|
|
fd: File objext to write to
|
|
offset: Offset to write to
|
|
"""
|
|
self._align_to(fd, self._align)
|
|
upto = fd.tell()
|
|
if upto > offset:
|
|
raise ValueError('No space for data before pad offset %#x (current offset %#x)' %
|
|
(offset, upto))
|
|
todo = align_int_down(offset - upto, self._align)
|
|
if todo:
|
|
cbf = CbfsFile.empty(todo, self._erase_byte)
|
|
fd.write(cbf.get_data_and_offset()[0])
|
|
self._skip_to(fd, offset)
|
|
|
|
def _align_to(self, fd, align):
|
|
"""Write out pad bytes until a given alignment is reached
|
|
|
|
This only aligns if the resulting output would not reach the end of the
|
|
CBFS, since we want to leave the last 4 bytes for the master-header
|
|
pointer.
|
|
|
|
Args:
|
|
fd: File objext to write to
|
|
align: Alignment to require (e.g. 4 means pad to next 4-byte
|
|
boundary)
|
|
"""
|
|
offset = align_int(fd.tell(), align)
|
|
if offset < self._size:
|
|
self._skip_to(fd, offset)
|
|
|
|
def add_file_stage(self, name, data, cbfs_offset=None):
|
|
"""Add a new stage file to the CBFS
|
|
|
|
Args:
|
|
name: String file name to put in CBFS (does not need to correspond
|
|
to the name that the file originally came from)
|
|
data: Contents of file
|
|
cbfs_offset: Offset of this file's data within the CBFS, in bytes,
|
|
or None to place this file anywhere
|
|
|
|
Returns:
|
|
CbfsFile object created
|
|
"""
|
|
cfile = CbfsFile.stage(self._base_address, name, data, cbfs_offset)
|
|
self._files[name] = cfile
|
|
return cfile
|
|
|
|
def add_file_raw(self, name, data, cbfs_offset=None,
|
|
compress=COMPRESS_NONE):
|
|
"""Create a new raw file
|
|
|
|
Args:
|
|
name: String file name to put in CBFS (does not need to correspond
|
|
to the name that the file originally came from)
|
|
data: Contents of file
|
|
cbfs_offset: Offset of this file's data within the CBFS, in bytes,
|
|
or None to place this file anywhere
|
|
compress: Compression algorithm to use (COMPRESS_...)
|
|
|
|
Returns:
|
|
CbfsFile object created
|
|
"""
|
|
cfile = CbfsFile.raw(name, data, cbfs_offset, compress)
|
|
self._files[name] = cfile
|
|
return cfile
|
|
|
|
def _write_header(self, fd, add_fileheader):
|
|
"""Write out the master header to a CBFS
|
|
|
|
Args:
|
|
fd: File object
|
|
add_fileheader: True to place the master header in a file header
|
|
record
|
|
"""
|
|
if fd.tell() > self._header_offset:
|
|
raise ValueError('No space for header at offset %#x (current offset %#x)' %
|
|
(self._header_offset, fd.tell()))
|
|
if not add_fileheader:
|
|
self._pad_to(fd, self._header_offset)
|
|
hdr = struct.pack(HEADER_FORMAT, HEADER_MAGIC, HEADER_VERSION2,
|
|
self._size, self._bootblock_size, self._align,
|
|
self._contents_offset, self._arch, 0xffffffff)
|
|
if add_fileheader:
|
|
name = _pack_string(self._master_name)
|
|
fd.write(struct.pack(FILE_HEADER_FORMAT, FILE_MAGIC, len(hdr),
|
|
TYPE_CBFSHEADER, 0,
|
|
FILE_HEADER_LEN + len(name)))
|
|
fd.write(name)
|
|
self._header_offset = fd.tell()
|
|
fd.write(hdr)
|
|
self._align_to(fd, self._align)
|
|
else:
|
|
fd.write(hdr)
|
|
|
|
def get_data(self):
|
|
"""Obtain the full contents of the CBFS
|
|
|
|
Thhis builds the CBFS with headers and all required files.
|
|
|
|
Returns:
|
|
'bytes' type containing the data
|
|
"""
|
|
fd = io.BytesIO()
|
|
|
|
# THe header can go at the start in some cases
|
|
if self._hdr_at_start:
|
|
self._write_header(fd, add_fileheader=self._add_fileheader)
|
|
self._skip_to(fd, self._contents_offset)
|
|
|
|
# Write out each file
|
|
for cbf in self._files.values():
|
|
# Place the file at its requested place, if any
|
|
offset = cbf.calc_start_offset()
|
|
if offset is not None:
|
|
self._pad_to(fd, align_int_down(offset, self._align))
|
|
pos = fd.tell()
|
|
data, data_offset = cbf.get_data_and_offset(pos, self._erase_byte)
|
|
fd.write(data)
|
|
self._align_to(fd, self._align)
|
|
cbf.calced_cbfs_offset = pos + data_offset
|
|
if not self._hdr_at_start:
|
|
self._write_header(fd, add_fileheader=self._add_fileheader)
|
|
|
|
# Pad to the end and write a pointer to the CBFS master header
|
|
self._pad_to(fd, self._base_address or self._size - 4)
|
|
rel_offset = self._header_offset - self._size
|
|
fd.write(struct.pack('<I', rel_offset & 0xffffffff))
|
|
|
|
return fd.getvalue()
|
|
|
|
|
|
class CbfsReader(object):
|
|
"""Class to handle reading a Coreboot File System (CBFS)
|
|
|
|
Usage is something like:
|
|
cbfs = cbfs_util.CbfsReader(data)
|
|
cfile = cbfs.files['u-boot']
|
|
self.WriteFile('u-boot.bin', cfile.data)
|
|
|
|
Attributes:
|
|
files: Ordered list of CbfsFile objects
|
|
align: Alignment to use for files, typically ENTRT_ALIGN
|
|
stage_base_address: Base address to use when mapping ELF files into the
|
|
CBFS for TYPE_STAGE files. If this is larger than the code address
|
|
of the ELF file, then data at the start of the ELF file will not
|
|
appear in the CBFS. Currently there are no tests for behaviour as
|
|
documentation is sparse
|
|
magic: Integer magic number from master header (HEADER_MAGIC)
|
|
version: Version number of CBFS (HEADER_VERSION2)
|
|
rom_size: Size of CBFS
|
|
boot_block_size: Size of boot block
|
|
cbfs_offset: Offset of the first file in bytes from start of CBFS
|
|
arch: Architecture of CBFS file (ARCHITECTURE_...)
|
|
"""
|
|
def __init__(self, data, read=True):
|
|
self.align = ENTRY_ALIGN
|
|
self.arch = None
|
|
self.boot_block_size = None
|
|
self.cbfs_offset = None
|
|
self.files = OrderedDict()
|
|
self.magic = None
|
|
self.rom_size = None
|
|
self.stage_base_address = 0
|
|
self.version = None
|
|
self.data = data
|
|
if read:
|
|
self.read()
|
|
|
|
def read(self):
|
|
"""Read all the files in the CBFS and add them to self.files"""
|
|
with io.BytesIO(self.data) as fd:
|
|
# First, get the master header
|
|
if not self._find_and_read_header(fd, len(self.data)):
|
|
raise ValueError('Cannot find master header')
|
|
fd.seek(self.cbfs_offset)
|
|
|
|
# Now read in the files one at a time
|
|
while True:
|
|
cfile = self._read_next_file(fd)
|
|
if cfile:
|
|
self.files[cfile.name] = cfile
|
|
elif cfile is False:
|
|
break
|
|
|
|
def _find_and_read_header(self, fd, size):
|
|
"""Find and read the master header in the CBFS
|
|
|
|
This looks at the pointer word at the very end of the CBFS. This is an
|
|
offset to the header relative to the size of the CBFS, which is assumed
|
|
to be known. Note that the offset is in *little endian* format.
|
|
|
|
Args:
|
|
fd: File to read from
|
|
size: Size of file
|
|
|
|
Returns:
|
|
True if header was found, False if not
|
|
"""
|
|
orig_pos = fd.tell()
|
|
fd.seek(size - 4)
|
|
rel_offset, = struct.unpack('<I', fd.read(4))
|
|
pos = (size + rel_offset) & 0xffffffff
|
|
fd.seek(pos)
|
|
found = self._read_header(fd)
|
|
if not found:
|
|
print('Relative offset seems wrong, scanning whole image')
|
|
for pos in range(0, size - HEADER_LEN, 4):
|
|
fd.seek(pos)
|
|
found = self._read_header(fd)
|
|
if found:
|
|
break
|
|
fd.seek(orig_pos)
|
|
return found
|
|
|
|
def _read_next_file(self, fd):
|
|
"""Read the next file from a CBFS
|
|
|
|
Args:
|
|
fd: File to read from
|
|
|
|
Returns:
|
|
CbfsFile object, if found
|
|
None if no object found, but data was parsed (e.g. TYPE_CBFSHEADER)
|
|
False if at end of CBFS and reading should stop
|
|
"""
|
|
file_pos = fd.tell()
|
|
data = fd.read(FILE_HEADER_LEN)
|
|
if len(data) < FILE_HEADER_LEN:
|
|
print('File header at %#x ran out of data' % file_pos)
|
|
return False
|
|
magic, size, ftype, attr, offset = struct.unpack(FILE_HEADER_FORMAT,
|
|
data)
|
|
if magic != FILE_MAGIC:
|
|
return False
|
|
pos = fd.tell()
|
|
name = self._read_string(fd)
|
|
if name is None:
|
|
print('String at %#x ran out of data' % pos)
|
|
return False
|
|
|
|
if DEBUG:
|
|
print('name', name)
|
|
|
|
# If there are attribute headers present, read those
|
|
compress = self._read_attr(fd, file_pos, attr, offset)
|
|
if compress is None:
|
|
return False
|
|
|
|
# Create the correct CbfsFile object depending on the type
|
|
cfile = None
|
|
cbfs_offset = file_pos + offset
|
|
fd.seek(cbfs_offset, io.SEEK_SET)
|
|
if ftype == TYPE_CBFSHEADER:
|
|
self._read_header(fd)
|
|
elif ftype == TYPE_STAGE:
|
|
data = fd.read(STAGE_LEN)
|
|
cfile = CbfsFile.stage(self.stage_base_address, name, b'',
|
|
cbfs_offset)
|
|
(cfile.compress, cfile.entry, cfile.load, cfile.data_len,
|
|
cfile.memlen) = struct.unpack(STAGE_FORMAT, data)
|
|
cfile.data = fd.read(cfile.data_len)
|
|
elif ftype == TYPE_RAW:
|
|
data = fd.read(size)
|
|
cfile = CbfsFile.raw(name, data, cbfs_offset, compress)
|
|
cfile.decompress()
|
|
if DEBUG:
|
|
print('data', data)
|
|
elif ftype == TYPE_EMPTY:
|
|
# Just read the data and discard it, since it is only padding
|
|
fd.read(size)
|
|
cfile = CbfsFile('', TYPE_EMPTY, b'', cbfs_offset)
|
|
else:
|
|
raise ValueError('Unknown type %#x when reading\n' % ftype)
|
|
if cfile:
|
|
cfile.offset = offset
|
|
|
|
# Move past the padding to the start of a possible next file. If we are
|
|
# already at an alignment boundary, then there is no padding.
|
|
pad = (self.align - fd.tell() % self.align) % self.align
|
|
fd.seek(pad, io.SEEK_CUR)
|
|
return cfile
|
|
|
|
@classmethod
|
|
def _read_attr(cls, fd, file_pos, attr, offset):
|
|
"""Read attributes from the file
|
|
|
|
CBFS files can have attributes which are things that cannot fit into the
|
|
header. The only attributes currently supported are compression and the
|
|
unused tag.
|
|
|
|
Args:
|
|
fd: File to read from
|
|
file_pos: Position of file in fd
|
|
attr: Offset of attributes, 0 if none
|
|
offset: Offset of file data (used to indicate the end of the
|
|
attributes)
|
|
|
|
Returns:
|
|
Compression to use for the file (COMPRESS_...)
|
|
"""
|
|
compress = COMPRESS_NONE
|
|
if not attr:
|
|
return compress
|
|
attr_size = offset - attr
|
|
fd.seek(file_pos + attr, io.SEEK_SET)
|
|
while attr_size:
|
|
pos = fd.tell()
|
|
hdr = fd.read(8)
|
|
if len(hdr) < 8:
|
|
print('Attribute tag at %x ran out of data' % pos)
|
|
return None
|
|
atag, alen = struct.unpack(">II", hdr)
|
|
data = hdr + fd.read(alen - 8)
|
|
if atag == FILE_ATTR_TAG_COMPRESSION:
|
|
# We don't currently use this information
|
|
atag, alen, compress, _decomp_size = struct.unpack(
|
|
ATTR_COMPRESSION_FORMAT, data)
|
|
elif atag == FILE_ATTR_TAG_UNUSED2:
|
|
break
|
|
else:
|
|
print('Unknown attribute tag %x' % atag)
|
|
attr_size -= len(data)
|
|
return compress
|
|
|
|
def _read_header(self, fd):
|
|
"""Read the master header
|
|
|
|
Reads the header and stores the information obtained into the member
|
|
variables.
|
|
|
|
Args:
|
|
fd: File to read from
|
|
|
|
Returns:
|
|
True if header was read OK, False if it is truncated or has the
|
|
wrong magic or version
|
|
"""
|
|
pos = fd.tell()
|
|
data = fd.read(HEADER_LEN)
|
|
if len(data) < HEADER_LEN:
|
|
print('Header at %x ran out of data' % pos)
|
|
return False
|
|
(self.magic, self.version, self.rom_size, self.boot_block_size,
|
|
self.align, self.cbfs_offset, self.arch, _) = struct.unpack(
|
|
HEADER_FORMAT, data)
|
|
return self.magic == HEADER_MAGIC and (
|
|
self.version == HEADER_VERSION1 or
|
|
self.version == HEADER_VERSION2)
|
|
|
|
@classmethod
|
|
def _read_string(cls, fd):
|
|
"""Read a string from a file
|
|
|
|
This reads a string and aligns the data to the next alignment boundary
|
|
|
|
Args:
|
|
fd: File to read from
|
|
|
|
Returns:
|
|
string read ('str' type) encoded to UTF-8, or None if we ran out of
|
|
data
|
|
"""
|
|
val = b''
|
|
while True:
|
|
data = fd.read(FILENAME_ALIGN)
|
|
if len(data) < FILENAME_ALIGN:
|
|
return None
|
|
pos = data.find(b'\0')
|
|
if pos == -1:
|
|
val += data
|
|
else:
|
|
val += data[:pos]
|
|
break
|
|
return val.decode('utf-8')
|
|
|
|
|
|
def cbfstool(fname, *cbfs_args, **kwargs):
|
|
"""Run cbfstool with provided arguments
|
|
|
|
If the tool fails then this function raises an exception and prints out the
|
|
output and stderr.
|
|
|
|
Args:
|
|
fname: Filename of CBFS
|
|
*cbfs_args: List of arguments to pass to cbfstool
|
|
|
|
Returns:
|
|
CommandResult object containing the results
|
|
"""
|
|
args = ['cbfstool', fname] + list(cbfs_args)
|
|
if kwargs.get('base') is not None:
|
|
args += ['-b', '%#x' % kwargs['base']]
|
|
result = command.RunPipe([args], capture=not VERBOSE,
|
|
capture_stderr=not VERBOSE, raise_on_error=False)
|
|
if result.return_code:
|
|
print(result.stderr, file=sys.stderr)
|
|
raise Exception("Failed to run (error %d): '%s'" %
|
|
(result.return_code, ' '.join(args)))
|