2018-05-06 21:58:06 +00:00
|
|
|
# SPDX-License-Identifier: GPL-2.0+
|
2016-11-26 03:15:52 +00:00
|
|
|
# Copyright (c) 2016 Google, Inc
|
|
|
|
# Written by Simon Glass <sjg@chromium.org>
|
|
|
|
#
|
|
|
|
# To run a single test, change to this directory, and:
|
|
|
|
#
|
|
|
|
# python -m unittest func_test.TestFunctional.testHelp
|
|
|
|
|
2020-07-10 00:39:45 +00:00
|
|
|
import collections
|
2020-04-18 00:09:03 +00:00
|
|
|
import gzip
|
2018-09-14 10:57:31 +00:00
|
|
|
import hashlib
|
2016-11-26 03:15:52 +00:00
|
|
|
from optparse import OptionParser
|
|
|
|
import os
|
2020-07-10 00:39:45 +00:00
|
|
|
import re
|
2016-11-26 03:15:52 +00:00
|
|
|
import shutil
|
|
|
|
import struct
|
|
|
|
import sys
|
|
|
|
import tempfile
|
|
|
|
import unittest
|
2022-01-10 03:13:57 +00:00
|
|
|
import unittest.mock
|
|
|
|
import urllib.error
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2022-01-10 03:13:50 +00:00
|
|
|
from binman import bintool
|
2020-04-18 00:09:03 +00:00
|
|
|
from binman import cbfs_util
|
|
|
|
from binman import cmdline
|
|
|
|
from binman import control
|
|
|
|
from binman import elf
|
|
|
|
from binman import elf_test
|
2021-11-24 04:08:59 +00:00
|
|
|
from binman import fip_util
|
2020-04-18 00:09:03 +00:00
|
|
|
from binman import fmap_util
|
|
|
|
from binman import state
|
|
|
|
from dtoc import fdt
|
|
|
|
from dtoc import fdt_util
|
|
|
|
from binman.etype import fdtmap
|
|
|
|
from binman.etype import image_header
|
2020-08-05 19:27:47 +00:00
|
|
|
from binman.image import Image
|
2020-04-18 00:09:04 +00:00
|
|
|
from patman import command
|
|
|
|
from patman import test_util
|
|
|
|
from patman import tools
|
|
|
|
from patman import tout
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Contents of test files, corresponding to different entry types
|
2019-05-18 04:00:46 +00:00
|
|
|
U_BOOT_DATA = b'1234'
|
|
|
|
U_BOOT_IMG_DATA = b'img'
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
U_BOOT_SPL_DATA = b'56780123456789abcdefghijklm'
|
|
|
|
U_BOOT_TPL_DATA = b'tpl9876543210fedcbazywvuts'
|
2022-02-28 14:16:54 +00:00
|
|
|
U_BOOT_VPL_DATA = b'vpl76543210fedcbazywxyz_'
|
2019-05-18 04:00:46 +00:00
|
|
|
BLOB_DATA = b'89'
|
|
|
|
ME_DATA = b'0abcd'
|
|
|
|
VGA_DATA = b'vga'
|
|
|
|
U_BOOT_DTB_DATA = b'udtb'
|
|
|
|
U_BOOT_SPL_DTB_DATA = b'spldtb'
|
|
|
|
U_BOOT_TPL_DTB_DATA = b'tpldtb'
|
2022-02-28 14:16:54 +00:00
|
|
|
U_BOOT_VPL_DTB_DATA = b'vpldtb'
|
2019-05-18 04:00:46 +00:00
|
|
|
X86_START16_DATA = b'start16'
|
|
|
|
X86_START16_SPL_DATA = b'start16spl'
|
|
|
|
X86_START16_TPL_DATA = b'start16tpl'
|
2019-08-24 13:22:48 +00:00
|
|
|
X86_RESET16_DATA = b'reset16'
|
|
|
|
X86_RESET16_SPL_DATA = b'reset16spl'
|
|
|
|
X86_RESET16_TPL_DATA = b'reset16tpl'
|
2019-05-18 04:00:46 +00:00
|
|
|
PPC_MPC85XX_BR_DATA = b'ppcmpc85xxbr'
|
|
|
|
U_BOOT_NODTB_DATA = b'nodtb with microcode pointer somewhere in here'
|
|
|
|
U_BOOT_SPL_NODTB_DATA = b'splnodtb with microcode pointer somewhere in here'
|
|
|
|
U_BOOT_TPL_NODTB_DATA = b'tplnodtb with microcode pointer somewhere in here'
|
2022-02-28 14:16:54 +00:00
|
|
|
U_BOOT_VPL_NODTB_DATA = b'vplnodtb'
|
2022-02-07 22:08:04 +00:00
|
|
|
U_BOOT_EXP_DATA = U_BOOT_NODTB_DATA + U_BOOT_DTB_DATA
|
|
|
|
U_BOOT_SPL_EXP_DATA = U_BOOT_SPL_NODTB_DATA + U_BOOT_SPL_DTB_DATA
|
|
|
|
U_BOOT_TPL_EXP_DATA = U_BOOT_TPL_NODTB_DATA + U_BOOT_TPL_DTB_DATA
|
2019-05-18 04:00:46 +00:00
|
|
|
FSP_DATA = b'fsp'
|
|
|
|
CMC_DATA = b'cmc'
|
|
|
|
VBT_DATA = b'vbt'
|
|
|
|
MRC_DATA = b'mrc'
|
2018-07-17 19:25:33 +00:00
|
|
|
TEXT_DATA = 'text'
|
|
|
|
TEXT_DATA2 = 'text2'
|
|
|
|
TEXT_DATA3 = 'text3'
|
2019-05-18 04:00:46 +00:00
|
|
|
CROS_EC_RW_DATA = b'ecrw'
|
|
|
|
GBB_DATA = b'gbbd'
|
|
|
|
BMPBLK_DATA = b'bmp'
|
|
|
|
VBLOCK_DATA = b'vblk'
|
|
|
|
FILES_DATA = (b"sorry I'm late\nOh, don't bother apologising, I'm " +
|
|
|
|
b"sorry you're alive\n")
|
2019-07-08 19:18:42 +00:00
|
|
|
COMPRESS_DATA = b'compress xxxxxxxxxxxxxxxxxxxxxx data'
|
2020-10-26 23:40:25 +00:00
|
|
|
COMPRESS_DATA_BIG = COMPRESS_DATA * 2
|
2019-05-18 04:00:46 +00:00
|
|
|
REFCODE_DATA = b'refcode'
|
2019-08-24 13:23:07 +00:00
|
|
|
FSP_M_DATA = b'fsp_m'
|
2019-10-21 03:31:35 +00:00
|
|
|
FSP_S_DATA = b'fsp_s'
|
2019-10-21 03:31:36 +00:00
|
|
|
FSP_T_DATA = b'fsp_t'
|
2020-09-01 11:13:58 +00:00
|
|
|
ATF_BL31_DATA = b'bl31'
|
2022-02-19 18:50:04 +00:00
|
|
|
TEE_OS_DATA = b'this is some tee OS data'
|
2021-11-24 04:08:59 +00:00
|
|
|
ATF_BL2U_DATA = b'bl2u'
|
2021-05-10 12:23:33 +00:00
|
|
|
OPENSBI_DATA = b'opensbi'
|
2020-10-22 02:12:15 +00:00
|
|
|
SCP_DATA = b'scp'
|
2020-09-01 11:13:59 +00:00
|
|
|
TEST_FDT1_DATA = b'fdt1'
|
|
|
|
TEST_FDT2_DATA = b'test-fdt2'
|
2020-09-06 16:35:33 +00:00
|
|
|
ENV_DATA = b'var1=1\nvar2="2"'
|
2022-03-28 20:57:04 +00:00
|
|
|
PRE_LOAD_MAGIC = b'UBSH'
|
|
|
|
PRE_LOAD_VERSION = 0x11223344.to_bytes(4, 'big')
|
|
|
|
PRE_LOAD_HDR_SIZE = 0x00001000.to_bytes(4, 'big')
|
2020-09-01 11:13:59 +00:00
|
|
|
|
|
|
|
# Subdirectory of the input dir to use to put test FDTs
|
|
|
|
TEST_FDT_SUBDIR = 'fdts'
|
2018-07-17 19:25:39 +00:00
|
|
|
|
2019-07-20 18:23:47 +00:00
|
|
|
# The expected size for the device tree in some tests
|
2019-07-08 20:25:50 +00:00
|
|
|
EXTRACT_DTB_SIZE = 0x3c9
|
|
|
|
|
2019-07-20 18:23:47 +00:00
|
|
|
# Properties expected to be in the device tree when update_dtb is used
|
|
|
|
BASE_DTB_PROPS = ['offset', 'size', 'image-pos']
|
|
|
|
|
2019-07-20 18:23:51 +00:00
|
|
|
# Extra properties expected to be in the device tree when allow-repack is used
|
|
|
|
REPACK_DTB_PROPS = ['orig-offset', 'orig-size']
|
|
|
|
|
2022-08-19 14:25:28 +00:00
|
|
|
# Supported compression bintools
|
2022-08-19 14:25:38 +00:00
|
|
|
COMP_BINTOOLS = ['bzip2', 'gzip', 'lz4', 'lzma_alone', 'lzop', 'xz', 'zstd']
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2023-01-07 21:07:14 +00:00
|
|
|
TEE_ADDR = 0x5678
|
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
class TestFunctional(unittest.TestCase):
|
|
|
|
"""Functional tests for binman
|
|
|
|
|
|
|
|
Most of these use a sample .dts file to build an image and then check
|
|
|
|
that it looks correct. The sample files are in the test/ subdirectory
|
|
|
|
and are numbered.
|
|
|
|
|
|
|
|
For each entry type a very small test file is created using fixed
|
|
|
|
string contents. This makes it easy to test that things look right, and
|
|
|
|
debug problems.
|
|
|
|
|
|
|
|
In some cases a 'real' file must be used - these are also supplied in
|
|
|
|
the test/ diurectory.
|
|
|
|
"""
|
|
|
|
@classmethod
|
2019-08-24 13:22:43 +00:00
|
|
|
def setUpClass(cls):
|
2017-11-13 04:52:20 +00:00
|
|
|
global entry
|
2020-04-18 00:09:03 +00:00
|
|
|
from binman import entry
|
2017-11-13 04:52:20 +00:00
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
# Handle the case where argv[0] is 'python'
|
2019-08-24 13:22:43 +00:00
|
|
|
cls._binman_dir = os.path.dirname(os.path.realpath(sys.argv[0]))
|
|
|
|
cls._binman_pathname = os.path.join(cls._binman_dir, 'binman')
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Create a temporary directory for input files
|
2019-08-24 13:22:43 +00:00
|
|
|
cls._indir = tempfile.mkdtemp(prefix='binmant.')
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Create some test files
|
|
|
|
TestFunctional._MakeInputFile('u-boot.bin', U_BOOT_DATA)
|
|
|
|
TestFunctional._MakeInputFile('u-boot.img', U_BOOT_IMG_DATA)
|
|
|
|
TestFunctional._MakeInputFile('spl/u-boot-spl.bin', U_BOOT_SPL_DATA)
|
2018-07-17 19:25:48 +00:00
|
|
|
TestFunctional._MakeInputFile('tpl/u-boot-tpl.bin', U_BOOT_TPL_DATA)
|
2022-02-28 14:16:54 +00:00
|
|
|
TestFunctional._MakeInputFile('vpl/u-boot-vpl.bin', U_BOOT_VPL_DATA)
|
2016-11-26 03:15:52 +00:00
|
|
|
TestFunctional._MakeInputFile('blobfile', BLOB_DATA)
|
2016-11-26 03:15:53 +00:00
|
|
|
TestFunctional._MakeInputFile('me.bin', ME_DATA)
|
|
|
|
TestFunctional._MakeInputFile('vga.bin', VGA_DATA)
|
2019-08-24 13:22:43 +00:00
|
|
|
cls._ResetDtbs()
|
2019-08-24 13:22:48 +00:00
|
|
|
|
2018-09-03 16:05:08 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot-br.bin', PPC_MPC85XX_BR_DATA)
|
2019-08-24 13:22:48 +00:00
|
|
|
|
2019-08-24 13:22:49 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot-x86-start16.bin', X86_START16_DATA)
|
|
|
|
TestFunctional._MakeInputFile('spl/u-boot-x86-start16-spl.bin',
|
2017-11-13 04:52:26 +00:00
|
|
|
X86_START16_SPL_DATA)
|
2019-08-24 13:22:49 +00:00
|
|
|
TestFunctional._MakeInputFile('tpl/u-boot-x86-start16-tpl.bin',
|
2018-09-14 10:57:10 +00:00
|
|
|
X86_START16_TPL_DATA)
|
2019-08-24 13:22:48 +00:00
|
|
|
|
|
|
|
TestFunctional._MakeInputFile('u-boot-x86-reset16.bin',
|
|
|
|
X86_RESET16_DATA)
|
|
|
|
TestFunctional._MakeInputFile('spl/u-boot-x86-reset16-spl.bin',
|
|
|
|
X86_RESET16_SPL_DATA)
|
|
|
|
TestFunctional._MakeInputFile('tpl/u-boot-x86-reset16-tpl.bin',
|
|
|
|
X86_RESET16_TPL_DATA)
|
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot-nodtb.bin', U_BOOT_NODTB_DATA)
|
2017-11-13 04:52:27 +00:00
|
|
|
TestFunctional._MakeInputFile('spl/u-boot-spl-nodtb.bin',
|
|
|
|
U_BOOT_SPL_NODTB_DATA)
|
2018-09-14 10:57:32 +00:00
|
|
|
TestFunctional._MakeInputFile('tpl/u-boot-tpl-nodtb.bin',
|
|
|
|
U_BOOT_TPL_NODTB_DATA)
|
2022-02-28 14:16:54 +00:00
|
|
|
TestFunctional._MakeInputFile('vpl/u-boot-vpl-nodtb.bin',
|
|
|
|
U_BOOT_VPL_NODTB_DATA)
|
2016-11-26 03:15:56 +00:00
|
|
|
TestFunctional._MakeInputFile('fsp.bin', FSP_DATA)
|
|
|
|
TestFunctional._MakeInputFile('cmc.bin', CMC_DATA)
|
2017-08-16 05:41:54 +00:00
|
|
|
TestFunctional._MakeInputFile('vbt.bin', VBT_DATA)
|
2017-11-13 04:52:28 +00:00
|
|
|
TestFunctional._MakeInputFile('mrc.bin', MRC_DATA)
|
2018-07-17 19:25:39 +00:00
|
|
|
TestFunctional._MakeInputFile('ecrw.bin', CROS_EC_RW_DATA)
|
2018-07-17 19:25:44 +00:00
|
|
|
TestFunctional._MakeInputDir('devkeys')
|
|
|
|
TestFunctional._MakeInputFile('bmpblk.bin', BMPBLK_DATA)
|
2018-10-01 18:22:31 +00:00
|
|
|
TestFunctional._MakeInputFile('refcode.bin', REFCODE_DATA)
|
2019-08-24 13:23:07 +00:00
|
|
|
TestFunctional._MakeInputFile('fsp_m.bin', FSP_M_DATA)
|
2019-10-21 03:31:35 +00:00
|
|
|
TestFunctional._MakeInputFile('fsp_s.bin', FSP_S_DATA)
|
2019-10-21 03:31:36 +00:00
|
|
|
TestFunctional._MakeInputFile('fsp_t.bin', FSP_T_DATA)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2019-08-24 13:22:53 +00:00
|
|
|
cls._elf_testdir = os.path.join(cls._indir, 'elftest')
|
|
|
|
elf_test.BuildElfTestFiles(cls._elf_testdir)
|
|
|
|
|
2016-11-26 03:15:53 +00:00
|
|
|
# ELF file with a '_dt_ucode_base_size' symbol
|
2019-08-24 13:22:54 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot',
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.read_file(cls.ElfTestFile('u_boot_ucode_ptr')))
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
# Intel flash descriptor file
|
2020-07-10 00:39:41 +00:00
|
|
|
cls._SetupDescriptor()
|
2016-11-26 03:15:53 +00:00
|
|
|
|
2019-08-24 13:22:43 +00:00
|
|
|
shutil.copytree(cls.TestFile('files'),
|
|
|
|
os.path.join(cls._indir, 'files'))
|
2018-09-14 10:57:28 +00:00
|
|
|
|
2018-09-14 10:57:26 +00:00
|
|
|
TestFunctional._MakeInputFile('compress', COMPRESS_DATA)
|
2020-10-26 23:40:25 +00:00
|
|
|
TestFunctional._MakeInputFile('compress_big', COMPRESS_DATA_BIG)
|
2020-09-01 11:13:58 +00:00
|
|
|
TestFunctional._MakeInputFile('bl31.bin', ATF_BL31_DATA)
|
2022-02-19 18:50:04 +00:00
|
|
|
TestFunctional._MakeInputFile('tee-pager.bin', TEE_OS_DATA)
|
2021-11-24 04:08:59 +00:00
|
|
|
TestFunctional._MakeInputFile('bl2u.bin', ATF_BL2U_DATA)
|
2021-05-10 12:23:33 +00:00
|
|
|
TestFunctional._MakeInputFile('fw_dynamic.bin', OPENSBI_DATA)
|
2020-10-22 02:12:15 +00:00
|
|
|
TestFunctional._MakeInputFile('scp.bin', SCP_DATA)
|
2018-09-14 10:57:26 +00:00
|
|
|
|
2020-09-01 11:13:59 +00:00
|
|
|
# Add a few .dtb files for testing
|
|
|
|
TestFunctional._MakeInputFile('%s/test-fdt1.dtb' % TEST_FDT_SUBDIR,
|
|
|
|
TEST_FDT1_DATA)
|
|
|
|
TestFunctional._MakeInputFile('%s/test-fdt2.dtb' % TEST_FDT_SUBDIR,
|
|
|
|
TEST_FDT2_DATA)
|
|
|
|
|
2020-09-06 16:35:33 +00:00
|
|
|
TestFunctional._MakeInputFile('env.txt', ENV_DATA)
|
|
|
|
|
2022-03-06 03:19:12 +00:00
|
|
|
# ELF file with two sections in different parts of memory, used for both
|
|
|
|
# ATF and OP_TEE
|
|
|
|
TestFunctional._MakeInputFile('bl31.elf',
|
|
|
|
tools.read_file(cls.ElfTestFile('elf_sections')))
|
|
|
|
TestFunctional._MakeInputFile('tee.elf',
|
|
|
|
tools.read_file(cls.ElfTestFile('elf_sections')))
|
|
|
|
|
2023-01-07 21:07:14 +00:00
|
|
|
# Newer OP_TEE file in v1 binary format
|
|
|
|
cls.make_tee_bin('tee.bin')
|
|
|
|
|
2022-08-19 14:25:28 +00:00
|
|
|
cls.comp_bintools = {}
|
|
|
|
for name in COMP_BINTOOLS:
|
|
|
|
cls.comp_bintools[name] = bintool.Bintool.create(name)
|
2019-07-08 19:18:53 +00:00
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
@classmethod
|
2019-08-24 13:22:43 +00:00
|
|
|
def tearDownClass(cls):
|
2016-11-26 03:15:52 +00:00
|
|
|
"""Remove the temporary input directory and its contents"""
|
2019-08-24 13:22:43 +00:00
|
|
|
if cls.preserve_indir:
|
|
|
|
print('Preserving input dir: %s' % cls._indir)
|
2019-07-08 19:18:49 +00:00
|
|
|
else:
|
2019-08-24 13:22:43 +00:00
|
|
|
if cls._indir:
|
|
|
|
shutil.rmtree(cls._indir)
|
|
|
|
cls._indir = None
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2019-07-08 19:18:49 +00:00
|
|
|
@classmethod
|
2019-07-08 19:18:50 +00:00
|
|
|
def setup_test_args(cls, preserve_indir=False, preserve_outdirs=False,
|
2019-07-08 20:25:29 +00:00
|
|
|
toolpath=None, verbosity=None):
|
2019-07-08 19:18:49 +00:00
|
|
|
"""Accept arguments controlling test execution
|
|
|
|
|
|
|
|
Args:
|
|
|
|
preserve_indir: Preserve the shared input directory used by all
|
|
|
|
tests in this class.
|
|
|
|
preserve_outdir: Preserve the output directories used by tests. Each
|
|
|
|
test has its own, so this is normally only useful when running a
|
|
|
|
single test.
|
2019-07-08 19:18:50 +00:00
|
|
|
toolpath: ist of paths to use for tools
|
2019-07-08 19:18:49 +00:00
|
|
|
"""
|
|
|
|
cls.preserve_indir = preserve_indir
|
|
|
|
cls.preserve_outdirs = preserve_outdirs
|
2019-07-08 19:18:50 +00:00
|
|
|
cls.toolpath = toolpath
|
2019-07-08 20:25:29 +00:00
|
|
|
cls.verbosity = verbosity
|
2019-07-08 19:18:49 +00:00
|
|
|
|
2022-08-19 14:25:28 +00:00
|
|
|
def _CheckBintool(self, bintool):
|
|
|
|
if not bintool.is_present():
|
|
|
|
self.skipTest('%s not available' % bintool.name)
|
|
|
|
|
2019-07-08 19:18:53 +00:00
|
|
|
def _CheckLz4(self):
|
2022-08-19 14:25:28 +00:00
|
|
|
bintool = self.comp_bintools['lz4']
|
|
|
|
self._CheckBintool(bintool)
|
2019-07-08 19:18:53 +00:00
|
|
|
|
2019-07-20 18:24:09 +00:00
|
|
|
def _CleanupOutputDir(self):
|
|
|
|
"""Remove the temporary output directory"""
|
|
|
|
if self.preserve_outdirs:
|
|
|
|
print('Preserving output dir: %s' % tools.outdir)
|
|
|
|
else:
|
2022-01-29 21:14:04 +00:00
|
|
|
tools._finalise_for_test()
|
2019-07-20 18:24:09 +00:00
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
def setUp(self):
|
|
|
|
# Enable this to turn on debugging output
|
2022-01-29 21:14:15 +00:00
|
|
|
# tout.init(tout.DEBUG)
|
2016-11-26 03:15:52 +00:00
|
|
|
command.test_result = None
|
|
|
|
|
|
|
|
def tearDown(self):
|
|
|
|
"""Remove the temporary output directory"""
|
2019-07-20 18:24:09 +00:00
|
|
|
self._CleanupOutputDir()
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2019-07-20 18:24:10 +00:00
|
|
|
def _SetupImageInTmpdir(self):
|
|
|
|
"""Set up the output image in a new temporary directory
|
|
|
|
|
|
|
|
This is used when an image has been generated in the output directory,
|
|
|
|
but we want to run binman again. This will create a new output
|
|
|
|
directory and fail to delete the original one.
|
|
|
|
|
|
|
|
This creates a new temporary directory, copies the image to it (with a
|
|
|
|
new name) and removes the old output directory.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Tuple:
|
|
|
|
Temporary directory to use
|
|
|
|
New image filename
|
|
|
|
"""
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-20 18:24:10 +00:00
|
|
|
tmpdir = tempfile.mkdtemp(prefix='binman.')
|
|
|
|
updated_fname = os.path.join(tmpdir, 'image-updated.bin')
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(updated_fname, tools.read_file(image_fname))
|
2019-07-20 18:24:10 +00:00
|
|
|
self._CleanupOutputDir()
|
|
|
|
return tmpdir, updated_fname
|
|
|
|
|
2018-07-17 19:25:48 +00:00
|
|
|
@classmethod
|
2019-08-24 13:22:43 +00:00
|
|
|
def _ResetDtbs(cls):
|
2018-07-17 19:25:48 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot.dtb', U_BOOT_DTB_DATA)
|
|
|
|
TestFunctional._MakeInputFile('spl/u-boot-spl.dtb', U_BOOT_SPL_DTB_DATA)
|
|
|
|
TestFunctional._MakeInputFile('tpl/u-boot-tpl.dtb', U_BOOT_TPL_DTB_DATA)
|
2022-02-28 14:16:54 +00:00
|
|
|
TestFunctional._MakeInputFile('vpl/u-boot-vpl.dtb', U_BOOT_VPL_DTB_DATA)
|
2018-07-17 19:25:48 +00:00
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
def _RunBinman(self, *args, **kwargs):
|
|
|
|
"""Run binman using the command line
|
|
|
|
|
|
|
|
Args:
|
|
|
|
Arguments to pass, as a list of strings
|
|
|
|
kwargs: Arguments to pass to Command.RunPipe()
|
|
|
|
"""
|
2022-01-29 21:14:05 +00:00
|
|
|
result = command.run_pipe([[self._binman_pathname] + list(args)],
|
2016-11-26 03:15:52 +00:00
|
|
|
capture=True, capture_stderr=True, raise_on_error=False)
|
|
|
|
if result.return_code and kwargs.get('raise_on_error', True):
|
|
|
|
raise Exception("Error running '%s': %s" % (' '.join(args),
|
|
|
|
result.stdout + result.stderr))
|
|
|
|
return result
|
|
|
|
|
2019-07-08 20:25:29 +00:00
|
|
|
def _DoBinman(self, *argv):
|
2016-11-26 03:15:52 +00:00
|
|
|
"""Run binman using directly (in the same process)
|
|
|
|
|
|
|
|
Args:
|
|
|
|
Arguments to pass, as a list of strings
|
|
|
|
Returns:
|
|
|
|
Return value (0 for success)
|
|
|
|
"""
|
2019-07-08 20:25:29 +00:00
|
|
|
argv = list(argv)
|
|
|
|
args = cmdline.ParseArgs(argv)
|
|
|
|
args.pager = 'binman-invalid-pager'
|
|
|
|
args.build_dir = self._indir
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# For testing, you can force an increase in verbosity here
|
2019-07-08 20:25:29 +00:00
|
|
|
# args.verbosity = tout.DEBUG
|
|
|
|
return control.Binman(args)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2018-07-17 19:25:32 +00:00
|
|
|
def _DoTestFile(self, fname, debug=False, map=False, update_dtb=False,
|
2019-04-26 03:58:34 +00:00
|
|
|
entry_args=None, images=None, use_real_dtb=False,
|
2021-03-18 07:25:05 +00:00
|
|
|
use_expanded=False, verbosity=None, allow_missing=False,
|
2022-01-06 10:49:41 +00:00
|
|
|
allow_fake_blobs=False, extra_indirs=None, threads=None,
|
2022-01-10 03:14:09 +00:00
|
|
|
test_section_timeout=False, update_fdt_in_elf=None,
|
2022-11-10 02:14:42 +00:00
|
|
|
force_missing_bintools='', ignore_missing=False):
|
2016-11-26 03:15:52 +00:00
|
|
|
"""Run binman with a given test file
|
|
|
|
|
|
|
|
Args:
|
2018-10-01 18:22:30 +00:00
|
|
|
fname: Device-tree source filename to use (e.g. 005_simple.dts)
|
2018-06-01 15:38:19 +00:00
|
|
|
debug: True to enable debugging output
|
2018-06-01 15:38:20 +00:00
|
|
|
map: True to output map files for the images
|
2018-08-01 21:22:37 +00:00
|
|
|
update_dtb: Update the offset and size of each entry in the device
|
2018-07-06 16:27:42 +00:00
|
|
|
tree before packing it into the image
|
2018-09-14 10:57:12 +00:00
|
|
|
entry_args: Dict of entry args to supply to binman
|
|
|
|
key: arg name
|
|
|
|
value: value of that arg
|
|
|
|
images: List of image names to build
|
2020-09-01 11:13:55 +00:00
|
|
|
use_real_dtb: True to use the test file as the contents of
|
|
|
|
the u-boot-dtb entry. Normally this is not needed and the
|
|
|
|
test contents (the U_BOOT_DTB_DATA string) can be used.
|
|
|
|
But in some test we need the real contents.
|
2021-03-18 07:25:05 +00:00
|
|
|
use_expanded: True to use expanded entries where available, e.g.
|
|
|
|
'u-boot-expanded' instead of 'u-boot'
|
2020-09-01 11:13:55 +00:00
|
|
|
verbosity: Verbosity level to use (0-3, None=don't set it)
|
|
|
|
allow_missing: Set the '--allow-missing' flag so that missing
|
|
|
|
external binaries just produce a warning instead of an error
|
2022-01-06 10:49:41 +00:00
|
|
|
allow_fake_blobs: Set the '--fake-ext-blobs' flag
|
2020-09-01 11:13:59 +00:00
|
|
|
extra_indirs: Extra input directories to add using -I
|
2021-07-06 16:36:37 +00:00
|
|
|
threads: Number of threads to use (None for default, 0 for
|
|
|
|
single-threaded)
|
2021-11-04 03:09:17 +00:00
|
|
|
test_section_timeout: True to force the first time to timeout, as
|
|
|
|
used in testThreadTimeout()
|
2021-11-04 03:09:18 +00:00
|
|
|
update_fdt_in_elf: Value to pass with --update-fdt-in-elf=xxx
|
2022-01-10 03:14:09 +00:00
|
|
|
force_missing_tools (str): comma-separated list of bintools to
|
|
|
|
regard as missing
|
2021-11-04 03:09:17 +00:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
int return code, 0 on success
|
2016-11-26 03:15:52 +00:00
|
|
|
"""
|
2019-07-08 20:25:29 +00:00
|
|
|
args = []
|
2017-11-14 01:55:00 +00:00
|
|
|
if debug:
|
|
|
|
args.append('-D')
|
2019-07-08 20:25:29 +00:00
|
|
|
if verbosity is not None:
|
|
|
|
args.append('-v%d' % verbosity)
|
|
|
|
elif self.verbosity:
|
|
|
|
args.append('-v%d' % self.verbosity)
|
|
|
|
if self.toolpath:
|
|
|
|
for path in self.toolpath:
|
|
|
|
args += ['--toolpath', path]
|
2021-07-06 16:36:37 +00:00
|
|
|
if threads is not None:
|
|
|
|
args.append('-T%d' % threads)
|
|
|
|
if test_section_timeout:
|
|
|
|
args.append('--test-section-timeout')
|
2019-07-08 20:25:29 +00:00
|
|
|
args += ['build', '-p', '-I', self._indir, '-d', self.TestFile(fname)]
|
2018-06-01 15:38:20 +00:00
|
|
|
if map:
|
|
|
|
args.append('-m')
|
2018-07-06 16:27:42 +00:00
|
|
|
if update_dtb:
|
2019-07-08 19:18:47 +00:00
|
|
|
args.append('-u')
|
2018-09-14 10:57:23 +00:00
|
|
|
if not use_real_dtb:
|
|
|
|
args.append('--fake-dtb')
|
2021-03-18 07:25:05 +00:00
|
|
|
if not use_expanded:
|
|
|
|
args.append('--no-expanded')
|
2018-07-17 19:25:32 +00:00
|
|
|
if entry_args:
|
2019-05-14 21:53:41 +00:00
|
|
|
for arg, value in entry_args.items():
|
2018-07-17 19:25:32 +00:00
|
|
|
args.append('-a%s=%s' % (arg, value))
|
2020-07-10 00:39:38 +00:00
|
|
|
if allow_missing:
|
|
|
|
args.append('-M')
|
2022-11-10 02:14:42 +00:00
|
|
|
if ignore_missing:
|
|
|
|
args.append('-W')
|
2022-01-06 10:49:41 +00:00
|
|
|
if allow_fake_blobs:
|
|
|
|
args.append('--fake-ext-blobs')
|
2022-01-10 03:14:09 +00:00
|
|
|
if force_missing_bintools:
|
|
|
|
args += ['--force-missing-bintools', force_missing_bintools]
|
2021-11-04 03:09:18 +00:00
|
|
|
if update_fdt_in_elf:
|
|
|
|
args += ['--update-fdt-in-elf', update_fdt_in_elf]
|
2018-09-14 10:57:12 +00:00
|
|
|
if images:
|
|
|
|
for image in images:
|
|
|
|
args += ['-i', image]
|
2020-09-01 11:13:59 +00:00
|
|
|
if extra_indirs:
|
|
|
|
for indir in extra_indirs:
|
|
|
|
args += ['-I', indir]
|
2017-11-14 01:55:00 +00:00
|
|
|
return self._DoBinman(*args)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
def _SetupDtb(self, fname, outfile='u-boot.dtb'):
|
2016-11-26 03:15:53 +00:00
|
|
|
"""Set up a new test device-tree file
|
|
|
|
|
|
|
|
The given file is compiled and set up as the device tree to be used
|
|
|
|
for ths test.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
fname: Filename of .dts file to read
|
2018-06-01 15:38:19 +00:00
|
|
|
outfile: Output filename for compiled device-tree binary
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
Returns:
|
2018-06-01 15:38:19 +00:00
|
|
|
Contents of device-tree binary
|
2016-11-26 03:15:53 +00:00
|
|
|
"""
|
2019-07-20 18:23:49 +00:00
|
|
|
tmpdir = tempfile.mkdtemp(prefix='binmant.')
|
|
|
|
dtb = fdt_util.EnsureCompiled(self.TestFile(fname), tmpdir)
|
2019-05-14 21:53:42 +00:00
|
|
|
with open(dtb, 'rb') as fd:
|
2016-11-26 03:15:52 +00:00
|
|
|
data = fd.read()
|
|
|
|
TestFunctional._MakeInputFile(outfile, data)
|
2019-07-20 18:23:49 +00:00
|
|
|
shutil.rmtree(tmpdir)
|
2018-10-02 03:12:41 +00:00
|
|
|
return data
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2022-02-28 14:16:54 +00:00
|
|
|
def _GetDtbContentsForSpls(self, dtb_data, name):
|
|
|
|
"""Create a version of the main DTB for SPL / TPL / VPL
|
2018-09-14 10:57:24 +00:00
|
|
|
|
|
|
|
For testing we don't actually have different versions of the DTB. With
|
|
|
|
U-Boot we normally run fdtgrep to remove unwanted nodes, but for tests
|
|
|
|
we don't normally have any unwanted nodes.
|
|
|
|
|
|
|
|
We still want the DTBs for SPL and TPL to be different though, since
|
|
|
|
otherwise it is confusing to know which one we are looking at. So add
|
|
|
|
an 'spl' or 'tpl' property to the top-level node.
|
2020-09-01 11:13:55 +00:00
|
|
|
|
|
|
|
Args:
|
|
|
|
dtb_data: dtb data to modify (this should be a value devicetree)
|
|
|
|
name: Name of a new property to add
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
New dtb data with the property added
|
2018-09-14 10:57:24 +00:00
|
|
|
"""
|
|
|
|
dtb = fdt.Fdt.FromData(dtb_data)
|
|
|
|
dtb.Scan()
|
|
|
|
dtb.GetNode('/binman').AddZeroProp(name)
|
|
|
|
dtb.Sync(auto_resize=True)
|
|
|
|
dtb.Pack()
|
|
|
|
return dtb.GetContents()
|
|
|
|
|
2021-03-18 07:25:05 +00:00
|
|
|
def _DoReadFileDtb(self, fname, use_real_dtb=False, use_expanded=False,
|
|
|
|
map=False, update_dtb=False, entry_args=None,
|
2021-07-06 16:36:37 +00:00
|
|
|
reset_dtbs=True, extra_indirs=None, threads=None):
|
2016-11-26 03:15:52 +00:00
|
|
|
"""Run binman and return the resulting image
|
|
|
|
|
|
|
|
This runs binman with a given test file and then reads the resulting
|
|
|
|
output file. It is a shortcut function since most tests need to do
|
|
|
|
these steps.
|
|
|
|
|
|
|
|
Raises an assertion failure if binman returns a non-zero exit code.
|
|
|
|
|
|
|
|
Args:
|
2018-10-01 18:22:30 +00:00
|
|
|
fname: Device-tree source filename to use (e.g. 005_simple.dts)
|
2016-11-26 03:15:52 +00:00
|
|
|
use_real_dtb: True to use the test file as the contents of
|
|
|
|
the u-boot-dtb entry. Normally this is not needed and the
|
|
|
|
test contents (the U_BOOT_DTB_DATA string) can be used.
|
|
|
|
But in some test we need the real contents.
|
2021-03-18 07:25:05 +00:00
|
|
|
use_expanded: True to use expanded entries where available, e.g.
|
|
|
|
'u-boot-expanded' instead of 'u-boot'
|
2018-06-01 15:38:20 +00:00
|
|
|
map: True to output map files for the images
|
2018-08-01 21:22:37 +00:00
|
|
|
update_dtb: Update the offset and size of each entry in the device
|
2018-07-06 16:27:42 +00:00
|
|
|
tree before packing it into the image
|
2020-09-01 11:13:55 +00:00
|
|
|
entry_args: Dict of entry args to supply to binman
|
|
|
|
key: arg name
|
|
|
|
value: value of that arg
|
|
|
|
reset_dtbs: With use_real_dtb the test dtb is overwritten by this
|
|
|
|
function. If reset_dtbs is True, then the original test dtb
|
|
|
|
is written back before this function finishes
|
2020-09-01 11:13:59 +00:00
|
|
|
extra_indirs: Extra input directories to add using -I
|
2021-07-06 16:36:37 +00:00
|
|
|
threads: Number of threads to use (None for default, 0 for
|
|
|
|
single-threaded)
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
Tuple:
|
|
|
|
Resulting image contents
|
|
|
|
Device tree contents
|
2018-06-01 15:38:20 +00:00
|
|
|
Map data showing contents of image (or None if none)
|
2018-07-17 19:25:27 +00:00
|
|
|
Output device tree binary filename ('u-boot.dtb' path)
|
2016-11-26 03:15:52 +00:00
|
|
|
"""
|
2016-11-26 03:15:53 +00:00
|
|
|
dtb_data = None
|
2016-11-26 03:15:52 +00:00
|
|
|
# Use the compiled test file as the u-boot-dtb input
|
|
|
|
if use_real_dtb:
|
2016-11-26 03:15:53 +00:00
|
|
|
dtb_data = self._SetupDtb(fname)
|
2018-09-14 10:57:24 +00:00
|
|
|
|
|
|
|
# For testing purposes, make a copy of the DT for SPL and TPL. Add
|
|
|
|
# a node indicating which it is, so aid verification.
|
2022-02-28 14:16:54 +00:00
|
|
|
for name in ['spl', 'tpl', 'vpl']:
|
2018-09-14 10:57:24 +00:00
|
|
|
dtb_fname = '%s/u-boot-%s.dtb' % (name, name)
|
|
|
|
outfile = os.path.join(self._indir, dtb_fname)
|
|
|
|
TestFunctional._MakeInputFile(dtb_fname,
|
2022-02-28 14:16:54 +00:00
|
|
|
self._GetDtbContentsForSpls(dtb_data, name))
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
try:
|
2018-07-17 19:25:32 +00:00
|
|
|
retcode = self._DoTestFile(fname, map=map, update_dtb=update_dtb,
|
2020-09-01 11:13:59 +00:00
|
|
|
entry_args=entry_args, use_real_dtb=use_real_dtb,
|
2021-07-06 16:36:37 +00:00
|
|
|
use_expanded=use_expanded, extra_indirs=extra_indirs,
|
|
|
|
threads=threads)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, retcode)
|
2022-01-29 21:14:04 +00:00
|
|
|
out_dtb_fname = tools.get_output_filename('u-boot.dtb.out')
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Find the (only) image, read it and return its contents
|
|
|
|
image = control.images['image']
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2018-07-06 16:27:42 +00:00
|
|
|
self.assertTrue(os.path.exists(image_fname))
|
2018-06-01 15:38:20 +00:00
|
|
|
if map:
|
2022-01-29 21:14:04 +00:00
|
|
|
map_fname = tools.get_output_filename('image.map')
|
2018-06-01 15:38:20 +00:00
|
|
|
with open(map_fname) as fd:
|
|
|
|
map_data = fd.read()
|
|
|
|
else:
|
|
|
|
map_data = None
|
2019-05-14 21:53:42 +00:00
|
|
|
with open(image_fname, 'rb') as fd:
|
2018-07-06 16:27:42 +00:00
|
|
|
return fd.read(), dtb_data, map_data, out_dtb_fname
|
2016-11-26 03:15:52 +00:00
|
|
|
finally:
|
|
|
|
# Put the test file back
|
2018-09-14 10:57:24 +00:00
|
|
|
if reset_dtbs and use_real_dtb:
|
2018-07-17 19:25:48 +00:00
|
|
|
self._ResetDtbs()
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2019-07-08 20:25:26 +00:00
|
|
|
def _DoReadFileRealDtb(self, fname):
|
|
|
|
"""Run binman with a real .dtb file and return the resulting data
|
|
|
|
|
|
|
|
Args:
|
|
|
|
fname: DT source filename to use (e.g. 082_fdt_update_all.dts)
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Resulting image contents
|
|
|
|
"""
|
|
|
|
return self._DoReadFileDtb(fname, use_real_dtb=True, update_dtb=True)[0]
|
|
|
|
|
2016-11-26 03:15:53 +00:00
|
|
|
def _DoReadFile(self, fname, use_real_dtb=False):
|
2018-06-01 15:38:19 +00:00
|
|
|
"""Helper function which discards the device-tree binary
|
|
|
|
|
|
|
|
Args:
|
2018-10-01 18:22:30 +00:00
|
|
|
fname: Device-tree source filename to use (e.g. 005_simple.dts)
|
2018-06-01 15:38:19 +00:00
|
|
|
use_real_dtb: True to use the test file as the contents of
|
|
|
|
the u-boot-dtb entry. Normally this is not needed and the
|
|
|
|
test contents (the U_BOOT_DTB_DATA string) can be used.
|
|
|
|
But in some test we need the real contents.
|
2018-07-17 19:25:27 +00:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
Resulting image contents
|
2018-06-01 15:38:19 +00:00
|
|
|
"""
|
2016-11-26 03:15:53 +00:00
|
|
|
return self._DoReadFileDtb(fname, use_real_dtb)[0]
|
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
@classmethod
|
2019-08-24 13:22:43 +00:00
|
|
|
def _MakeInputFile(cls, fname, contents):
|
2016-11-26 03:15:52 +00:00
|
|
|
"""Create a new test input file, creating directories as needed
|
|
|
|
|
|
|
|
Args:
|
2018-08-01 21:22:37 +00:00
|
|
|
fname: Filename to create
|
2016-11-26 03:15:52 +00:00
|
|
|
contents: File contents to write in to the file
|
|
|
|
Returns:
|
|
|
|
Full pathname of file created
|
|
|
|
"""
|
2019-08-24 13:22:43 +00:00
|
|
|
pathname = os.path.join(cls._indir, fname)
|
2016-11-26 03:15:52 +00:00
|
|
|
dirname = os.path.dirname(pathname)
|
|
|
|
if dirname and not os.path.exists(dirname):
|
|
|
|
os.makedirs(dirname)
|
|
|
|
with open(pathname, 'wb') as fd:
|
|
|
|
fd.write(contents)
|
|
|
|
return pathname
|
|
|
|
|
2018-07-17 19:25:44 +00:00
|
|
|
@classmethod
|
2019-08-24 13:22:43 +00:00
|
|
|
def _MakeInputDir(cls, dirname):
|
2018-07-17 19:25:44 +00:00
|
|
|
"""Create a new test input directory, creating directories as needed
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dirname: Directory name to create
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Full pathname of directory created
|
|
|
|
"""
|
2019-08-24 13:22:43 +00:00
|
|
|
pathname = os.path.join(cls._indir, dirname)
|
2018-07-17 19:25:44 +00:00
|
|
|
if not os.path.exists(pathname):
|
|
|
|
os.makedirs(pathname)
|
|
|
|
return pathname
|
|
|
|
|
2018-10-02 03:12:47 +00:00
|
|
|
@classmethod
|
2019-08-24 13:22:43 +00:00
|
|
|
def _SetupSplElf(cls, src_fname='bss_data'):
|
2018-10-02 03:12:47 +00:00
|
|
|
"""Set up an ELF file with a '_dt_ucode_base_size' symbol
|
|
|
|
|
|
|
|
Args:
|
|
|
|
Filename of ELF file to use as SPL
|
|
|
|
"""
|
2019-08-24 13:22:59 +00:00
|
|
|
TestFunctional._MakeInputFile('spl/u-boot-spl',
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.read_file(cls.ElfTestFile(src_fname)))
|
2018-10-02 03:12:47 +00:00
|
|
|
|
2019-08-24 13:23:00 +00:00
|
|
|
@classmethod
|
|
|
|
def _SetupTplElf(cls, src_fname='bss_data'):
|
|
|
|
"""Set up an ELF file with a '_dt_ucode_base_size' symbol
|
|
|
|
|
|
|
|
Args:
|
|
|
|
Filename of ELF file to use as TPL
|
|
|
|
"""
|
|
|
|
TestFunctional._MakeInputFile('tpl/u-boot-tpl',
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.read_file(cls.ElfTestFile(src_fname)))
|
2019-08-24 13:23:00 +00:00
|
|
|
|
2022-02-28 14:16:54 +00:00
|
|
|
@classmethod
|
|
|
|
def _SetupVplElf(cls, src_fname='bss_data'):
|
|
|
|
"""Set up an ELF file with a '_dt_ucode_base_size' symbol
|
|
|
|
|
|
|
|
Args:
|
|
|
|
Filename of ELF file to use as VPL
|
|
|
|
"""
|
|
|
|
TestFunctional._MakeInputFile('vpl/u-boot-vpl',
|
|
|
|
tools.read_file(cls.ElfTestFile(src_fname)))
|
|
|
|
|
2020-07-10 00:39:41 +00:00
|
|
|
@classmethod
|
|
|
|
def _SetupDescriptor(cls):
|
|
|
|
with open(cls.TestFile('descriptor.bin'), 'rb') as fd:
|
|
|
|
TestFunctional._MakeInputFile('descriptor.bin', fd.read())
|
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
@classmethod
|
2019-08-24 13:22:43 +00:00
|
|
|
def TestFile(cls, fname):
|
|
|
|
return os.path.join(cls._binman_dir, 'test', fname)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2019-08-24 13:22:53 +00:00
|
|
|
@classmethod
|
|
|
|
def ElfTestFile(cls, fname):
|
|
|
|
return os.path.join(cls._elf_testdir, fname)
|
|
|
|
|
2023-01-07 21:07:14 +00:00
|
|
|
@classmethod
|
|
|
|
def make_tee_bin(cls, fname, paged_sz=0, extra_data=b''):
|
|
|
|
init_sz, start_hi, start_lo, dummy = (len(U_BOOT_DATA), 0, TEE_ADDR, 0)
|
|
|
|
data = b'OPTE\x01xxx' + struct.pack('<5I', init_sz, start_hi, start_lo,
|
|
|
|
dummy, paged_sz) + U_BOOT_DATA
|
|
|
|
data += extra_data
|
|
|
|
TestFunctional._MakeInputFile(fname, data)
|
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
def AssertInList(self, grep_list, target):
|
|
|
|
"""Assert that at least one of a list of things is in a target
|
|
|
|
|
|
|
|
Args:
|
|
|
|
grep_list: List of strings to check
|
|
|
|
target: Target string
|
|
|
|
"""
|
|
|
|
for grep in grep_list:
|
|
|
|
if grep in target:
|
|
|
|
return
|
2019-05-18 04:00:50 +00:00
|
|
|
self.fail("Error: '%s' not found in '%s'" % (grep_list, target))
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
def CheckNoGaps(self, entries):
|
|
|
|
"""Check that all entries fit together without gaps
|
|
|
|
|
|
|
|
Args:
|
|
|
|
entries: List of entries to check
|
|
|
|
"""
|
2018-08-01 21:22:37 +00:00
|
|
|
offset = 0
|
2016-11-26 03:15:52 +00:00
|
|
|
for entry in entries.values():
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(offset, entry.offset)
|
|
|
|
offset += entry.size
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2016-11-26 03:15:53 +00:00
|
|
|
def GetFdtLen(self, dtb):
|
2018-06-01 15:38:19 +00:00
|
|
|
"""Get the totalsize field from a device-tree binary
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
Args:
|
2018-06-01 15:38:19 +00:00
|
|
|
dtb: Device-tree binary contents
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
Returns:
|
2018-06-01 15:38:19 +00:00
|
|
|
Total size of device-tree binary, from the header
|
2016-11-26 03:15:53 +00:00
|
|
|
"""
|
|
|
|
return struct.unpack('>L', dtb[4:8])[0]
|
|
|
|
|
2019-07-08 20:25:27 +00:00
|
|
|
def _GetPropTree(self, dtb, prop_names, prefix='/binman/'):
|
2018-07-06 16:27:42 +00:00
|
|
|
def AddNode(node, path):
|
|
|
|
if node.name != '/':
|
|
|
|
path += '/' + node.name
|
2019-07-08 20:25:27 +00:00
|
|
|
for prop in node.props.values():
|
|
|
|
if prop.name in prop_names:
|
|
|
|
prop_path = path + ':' + prop.name
|
|
|
|
tree[prop_path[len(prefix):]] = fdt_util.fdt32_to_cpu(
|
|
|
|
prop.value)
|
2018-07-06 16:27:42 +00:00
|
|
|
for subnode in node.subnodes:
|
|
|
|
AddNode(subnode, path)
|
|
|
|
|
|
|
|
tree = {}
|
|
|
|
AddNode(dtb.GetRoot(), '')
|
|
|
|
return tree
|
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
def testRun(self):
|
|
|
|
"""Test a basic run with valid args"""
|
|
|
|
result = self._RunBinman('-h')
|
|
|
|
|
|
|
|
def testFullHelp(self):
|
|
|
|
"""Test that the full help is displayed with -H"""
|
|
|
|
result = self._RunBinman('-H')
|
2021-03-18 07:25:13 +00:00
|
|
|
help_file = os.path.join(self._binman_dir, 'README.rst')
|
2018-01-16 20:29:50 +00:00
|
|
|
# Remove possible extraneous strings
|
|
|
|
extra = '::::::::::::::\n' + help_file + '\n::::::::::::::\n'
|
|
|
|
gothelp = result.stdout.replace(extra, '')
|
|
|
|
self.assertEqual(len(gothelp), os.path.getsize(help_file))
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, len(result.stderr))
|
|
|
|
self.assertEqual(0, result.return_code)
|
|
|
|
|
|
|
|
def testFullHelpInternal(self):
|
|
|
|
"""Test that the full help is displayed with -H"""
|
|
|
|
try:
|
|
|
|
command.test_result = command.CommandResult()
|
|
|
|
result = self._DoBinman('-H')
|
2021-03-18 07:25:13 +00:00
|
|
|
help_file = os.path.join(self._binman_dir, 'README.rst')
|
2016-11-26 03:15:52 +00:00
|
|
|
finally:
|
|
|
|
command.test_result = None
|
|
|
|
|
|
|
|
def testHelp(self):
|
|
|
|
"""Test that the basic help is displayed with -h"""
|
|
|
|
result = self._RunBinman('-h')
|
|
|
|
self.assertTrue(len(result.stdout) > 200)
|
|
|
|
self.assertEqual(0, len(result.stderr))
|
|
|
|
self.assertEqual(0, result.return_code)
|
|
|
|
|
|
|
|
def testBoard(self):
|
|
|
|
"""Test that we can run it with a specific board"""
|
2018-10-01 18:22:30 +00:00
|
|
|
self._SetupDtb('005_simple.dts', 'sandbox/u-boot.dtb')
|
2016-11-26 03:15:52 +00:00
|
|
|
TestFunctional._MakeInputFile('sandbox/u-boot.bin', U_BOOT_DATA)
|
2021-03-18 07:25:05 +00:00
|
|
|
result = self._DoBinman('build', '-n', '-b', 'sandbox')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, result)
|
|
|
|
|
|
|
|
def testNeedBoard(self):
|
|
|
|
"""Test that we get an error when no board ius supplied"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2019-07-08 20:25:29 +00:00
|
|
|
result = self._DoBinman('build')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("Must provide a board to process (use -b <board>)",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testMissingDt(self):
|
2018-06-01 15:38:19 +00:00
|
|
|
"""Test that an invalid device-tree file generates an error"""
|
2016-11-26 03:15:52 +00:00
|
|
|
with self.assertRaises(Exception) as e:
|
2019-07-08 20:25:29 +00:00
|
|
|
self._RunBinman('build', '-d', 'missing_file')
|
2016-11-26 03:15:52 +00:00
|
|
|
# We get one error from libfdt, and a different one from fdtget.
|
|
|
|
self.AssertInList(["Couldn't open blob from 'missing_file'",
|
|
|
|
'No such file or directory'], str(e.exception))
|
|
|
|
|
|
|
|
def testBrokenDt(self):
|
2018-06-01 15:38:19 +00:00
|
|
|
"""Test that an invalid device-tree source file generates an error
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
Since this is a source file it should be compiled and the error
|
|
|
|
will come from the device-tree compiler (dtc).
|
|
|
|
"""
|
|
|
|
with self.assertRaises(Exception) as e:
|
2019-07-08 20:25:29 +00:00
|
|
|
self._RunBinman('build', '-d', self.TestFile('001_invalid.dts'))
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("FATAL ERROR: Unable to parse input tree",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testMissingNode(self):
|
|
|
|
"""Test that a device tree without a 'binman' node generates an error"""
|
|
|
|
with self.assertRaises(Exception) as e:
|
2019-07-08 20:25:29 +00:00
|
|
|
self._DoBinman('build', '-d', self.TestFile('002_missing_node.dts'))
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("does not have a 'binman' node", str(e.exception))
|
|
|
|
|
|
|
|
def testEmpty(self):
|
|
|
|
"""Test that an empty binman node works OK (i.e. does nothing)"""
|
2019-07-08 20:25:29 +00:00
|
|
|
result = self._RunBinman('build', '-d', self.TestFile('003_empty.dts'))
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, len(result.stderr))
|
|
|
|
self.assertEqual(0, result.return_code)
|
|
|
|
|
|
|
|
def testInvalidEntry(self):
|
|
|
|
"""Test that an invalid entry is flagged"""
|
|
|
|
with self.assertRaises(Exception) as e:
|
2019-07-08 20:25:29 +00:00
|
|
|
result = self._RunBinman('build', '-d',
|
2018-10-01 18:22:30 +00:00
|
|
|
self.TestFile('004_invalid_entry.dts'))
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("Unknown entry type 'not-a-valid-type' in node "
|
|
|
|
"'/binman/not-a-valid-type'", str(e.exception))
|
|
|
|
|
|
|
|
def testSimple(self):
|
|
|
|
"""Test a simple binman with a single file"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('005_simple.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
2017-11-14 01:55:00 +00:00
|
|
|
def testSimpleDebug(self):
|
|
|
|
"""Test a simple binman run with debugging enabled"""
|
2019-07-08 20:25:53 +00:00
|
|
|
self._DoTestFile('005_simple.dts', debug=True)
|
2017-11-14 01:55:00 +00:00
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
def testDual(self):
|
|
|
|
"""Test that we can handle creating two images
|
|
|
|
|
|
|
|
This also tests image padding.
|
|
|
|
"""
|
2018-10-01 18:22:30 +00:00
|
|
|
retcode = self._DoTestFile('006_dual_image.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
|
|
|
|
image = control.images['image1']
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(len(U_BOOT_DATA), image.size)
|
2022-01-29 21:14:04 +00:00
|
|
|
fname = tools.get_output_filename('image1.bin')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertTrue(os.path.exists(fname))
|
2019-05-14 21:53:42 +00:00
|
|
|
with open(fname, 'rb') as fd:
|
2016-11-26 03:15:52 +00:00
|
|
|
data = fd.read()
|
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
image = control.images['image2']
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(3 + len(U_BOOT_DATA) + 5, image.size)
|
2022-01-29 21:14:04 +00:00
|
|
|
fname = tools.get_output_filename('image2.bin')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertTrue(os.path.exists(fname))
|
2019-05-14 21:53:42 +00:00
|
|
|
with open(fname, 'rb') as fd:
|
2016-11-26 03:15:52 +00:00
|
|
|
data = fd.read()
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[3:7])
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0, 3), data[:3])
|
|
|
|
self.assertEqual(tools.get_bytes(0, 5), data[7:])
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
def testBadAlign(self):
|
|
|
|
"""Test that an invalid alignment value is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('007_bad_align.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Alignment 23 must be a power "
|
|
|
|
"of two", str(e.exception))
|
|
|
|
|
|
|
|
def testPackSimple(self):
|
|
|
|
"""Test that packing works as expected"""
|
2018-10-01 18:22:30 +00:00
|
|
|
retcode = self._DoTestFile('008_pack.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
self.assertIn('image', control.images)
|
|
|
|
image = control.images['image']
|
2018-06-01 15:38:12 +00:00
|
|
|
entries = image.GetEntries()
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(5, len(entries))
|
|
|
|
|
|
|
|
# First u-boot
|
|
|
|
self.assertIn('u-boot', entries)
|
|
|
|
entry = entries['u-boot']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(0, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
|
|
|
# Second u-boot, aligned to 16-byte boundary
|
|
|
|
self.assertIn('u-boot-align', entries)
|
|
|
|
entry = entries['u-boot-align']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(16, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
|
|
|
# Third u-boot, size 23 bytes
|
|
|
|
self.assertIn('u-boot-size', entries)
|
|
|
|
entry = entries['u-boot-size']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(20, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.contents_size)
|
|
|
|
self.assertEqual(23, entry.size)
|
|
|
|
|
|
|
|
# Fourth u-boot, placed immediate after the above
|
|
|
|
self.assertIn('u-boot-next', entries)
|
|
|
|
entry = entries['u-boot-next']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(43, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
2018-08-01 21:22:37 +00:00
|
|
|
# Fifth u-boot, placed at a fixed offset
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn('u-boot-fixed', entries)
|
|
|
|
entry = entries['u-boot-fixed']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(61, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(65, image.size)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
def testPackExtra(self):
|
|
|
|
"""Test that extra packing feature works as expected"""
|
2020-10-26 23:40:10 +00:00
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb('009_pack_extra.dts',
|
|
|
|
update_dtb=True)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
self.assertIn('image', control.images)
|
|
|
|
image = control.images['image']
|
2018-06-01 15:38:12 +00:00
|
|
|
entries = image.GetEntries()
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(5, len(entries))
|
|
|
|
|
|
|
|
# First u-boot with padding before and after
|
|
|
|
self.assertIn('u-boot', entries)
|
|
|
|
entry = entries['u-boot']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(0, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(3, entry.pad_before)
|
|
|
|
self.assertEqual(3 + 5 + len(U_BOOT_DATA), entry.size)
|
2020-10-26 23:40:08 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0, 3) + U_BOOT_DATA +
|
|
|
|
tools.get_bytes(0, 5), data[:entry.size])
|
2020-10-26 23:40:08 +00:00
|
|
|
pos = entry.size
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Second u-boot has an aligned size, but it has no effect
|
|
|
|
self.assertIn('u-boot-align-size-nop', entries)
|
|
|
|
entry = entries['u-boot-align-size-nop']
|
2020-10-26 23:40:08 +00:00
|
|
|
self.assertEqual(pos, entry.offset)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[pos:pos + entry.size])
|
|
|
|
pos += entry.size
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Third u-boot has an aligned size too
|
|
|
|
self.assertIn('u-boot-align-size', entries)
|
|
|
|
entry = entries['u-boot-align-size']
|
2020-10-26 23:40:08 +00:00
|
|
|
self.assertEqual(pos, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(32, entry.size)
|
2020-10-26 23:40:08 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA + tools.get_bytes(0, 32 - len(U_BOOT_DATA)),
|
2020-10-26 23:40:08 +00:00
|
|
|
data[pos:pos + entry.size])
|
|
|
|
pos += entry.size
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Fourth u-boot has an aligned end
|
|
|
|
self.assertIn('u-boot-align-end', entries)
|
|
|
|
entry = entries['u-boot-align-end']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(48, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(16, entry.size)
|
2020-10-26 23:40:08 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data[:len(U_BOOT_DATA)])
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA + tools.get_bytes(0, 16 - len(U_BOOT_DATA)),
|
2020-10-26 23:40:08 +00:00
|
|
|
data[pos:pos + entry.size])
|
|
|
|
pos += entry.size
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
# Fifth u-boot immediately afterwards
|
|
|
|
self.assertIn('u-boot-align-both', entries)
|
|
|
|
entry = entries['u-boot-align-both']
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertEqual(64, entry.offset)
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(64, entry.size)
|
2020-10-26 23:40:08 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data[:len(U_BOOT_DATA)])
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA + tools.get_bytes(0, 64 - len(U_BOOT_DATA)),
|
2020-10-26 23:40:08 +00:00
|
|
|
data[pos:pos + entry.size])
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
self.CheckNoGaps(entries)
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(128, image.size)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2020-10-26 23:40:10 +00:00
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['size', 'offset', 'image-pos'])
|
|
|
|
expected = {
|
|
|
|
'image-pos': 0,
|
|
|
|
'offset': 0,
|
|
|
|
'size': 128,
|
|
|
|
|
|
|
|
'u-boot:image-pos': 0,
|
|
|
|
'u-boot:offset': 0,
|
|
|
|
'u-boot:size': 3 + 5 + len(U_BOOT_DATA),
|
|
|
|
|
|
|
|
'u-boot-align-size-nop:image-pos': 12,
|
|
|
|
'u-boot-align-size-nop:offset': 12,
|
|
|
|
'u-boot-align-size-nop:size': 4,
|
|
|
|
|
|
|
|
'u-boot-align-size:image-pos': 16,
|
|
|
|
'u-boot-align-size:offset': 16,
|
|
|
|
'u-boot-align-size:size': 32,
|
|
|
|
|
|
|
|
'u-boot-align-end:image-pos': 48,
|
|
|
|
'u-boot-align-end:offset': 48,
|
|
|
|
'u-boot-align-end:size': 16,
|
|
|
|
|
|
|
|
'u-boot-align-both:image-pos': 64,
|
|
|
|
'u-boot-align-both:offset': 64,
|
|
|
|
'u-boot-align-both:size': 64,
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
2016-11-26 03:15:52 +00:00
|
|
|
def testPackAlignPowerOf2(self):
|
|
|
|
"""Test that invalid entry alignment is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('010_pack_align_power2.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Alignment 5 must be a power "
|
|
|
|
"of two", str(e.exception))
|
|
|
|
|
|
|
|
def testPackAlignSizePowerOf2(self):
|
|
|
|
"""Test that invalid entry size alignment is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('011_pack_align_size_power2.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Alignment size 55 must be a "
|
|
|
|
"power of two", str(e.exception))
|
|
|
|
|
|
|
|
def testPackInvalidAlign(self):
|
2018-08-01 21:22:37 +00:00
|
|
|
"""Test detection of an offset that does not match its alignment"""
|
2016-11-26 03:15:52 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('012_pack_inv_align.dts')
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Offset 0x5 (5) does not match "
|
2016-11-26 03:15:52 +00:00
|
|
|
"align 0x4 (4)", str(e.exception))
|
|
|
|
|
|
|
|
def testPackInvalidSizeAlign(self):
|
|
|
|
"""Test that invalid entry size alignment is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('013_pack_inv_size_align.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Size 0x5 (5) does not match "
|
|
|
|
"align-size 0x4 (4)", str(e.exception))
|
|
|
|
|
|
|
|
def testPackOverlap(self):
|
|
|
|
"""Test that overlapping regions are detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('014_pack_overlap.dts')
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-align': Offset 0x3 (3) overlaps "
|
2016-11-26 03:15:52 +00:00
|
|
|
"with previous entry '/binman/u-boot' ending at 0x4 (4)",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testPackEntryOverflow(self):
|
|
|
|
"""Test that entries that overflow their size are detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('015_pack_overflow.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Entry contents size is 0x4 (4) "
|
|
|
|
"but entry size is 0x3 (3)", str(e.exception))
|
|
|
|
|
|
|
|
def testPackImageOverflow(self):
|
|
|
|
"""Test that entries which overflow the image size are detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('016_pack_image_overflow.dts')
|
2018-06-01 15:38:12 +00:00
|
|
|
self.assertIn("Section '/binman': contents size 0x4 (4) exceeds section "
|
2016-11-26 03:15:52 +00:00
|
|
|
"size 0x3 (3)", str(e.exception))
|
|
|
|
|
|
|
|
def testPackImageSize(self):
|
|
|
|
"""Test that the image size can be set"""
|
2018-10-01 18:22:30 +00:00
|
|
|
retcode = self._DoTestFile('017_pack_image_size.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
self.assertIn('image', control.images)
|
|
|
|
image = control.images['image']
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(7, image.size)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
def testPackImageSizeAlign(self):
|
|
|
|
"""Test that image size alignemnt works as expected"""
|
2018-10-01 18:22:30 +00:00
|
|
|
retcode = self._DoTestFile('018_pack_image_align.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
self.assertIn('image', control.images)
|
|
|
|
image = control.images['image']
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(16, image.size)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
def testPackInvalidImageAlign(self):
|
|
|
|
"""Test that invalid image alignment is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('019_pack_inv_image_align.dts')
|
2018-06-01 15:38:12 +00:00
|
|
|
self.assertIn("Section '/binman': Size 0x7 (7) does not match "
|
2016-11-26 03:15:52 +00:00
|
|
|
"align-size 0x8 (8)", str(e.exception))
|
|
|
|
|
2022-02-11 20:23:21 +00:00
|
|
|
def testPackAlignPowerOf2Inv(self):
|
2016-11-26 03:15:52 +00:00
|
|
|
"""Test that invalid image alignment is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('020_pack_inv_image_align_power2.dts')
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertIn("Image '/binman': Alignment size 131 must be a power of "
|
2016-11-26 03:15:52 +00:00
|
|
|
"two", str(e.exception))
|
|
|
|
|
|
|
|
def testImagePadByte(self):
|
|
|
|
"""Test that the image pad byte can be specified"""
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf()
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('021_image_pad.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(U_BOOT_SPL_DATA + tools.get_bytes(0xff, 1) +
|
2019-05-14 21:53:47 +00:00
|
|
|
U_BOOT_DATA, data)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
|
|
|
def testImageName(self):
|
|
|
|
"""Test that image files can be named"""
|
2018-10-01 18:22:30 +00:00
|
|
|
retcode = self._DoTestFile('022_image_name.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
image = control.images['image1']
|
2022-01-29 21:14:04 +00:00
|
|
|
fname = tools.get_output_filename('test-name')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertTrue(os.path.exists(fname))
|
|
|
|
|
|
|
|
image = control.images['image2']
|
2022-01-29 21:14:04 +00:00
|
|
|
fname = tools.get_output_filename('test-name.xx')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertTrue(os.path.exists(fname))
|
|
|
|
|
|
|
|
def testBlobFilename(self):
|
|
|
|
"""Test that generic blobs can be provided by filename"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('023_blob.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(BLOB_DATA, data)
|
|
|
|
|
|
|
|
def testPackSorted(self):
|
|
|
|
"""Test that entries can be sorted"""
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf()
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('024_sorted.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0, 1) + U_BOOT_SPL_DATA +
|
|
|
|
tools.get_bytes(0, 2) + U_BOOT_DATA, data)
|
2016-11-26 03:15:52 +00:00
|
|
|
|
2018-08-01 21:22:37 +00:00
|
|
|
def testPackZeroOffset(self):
|
|
|
|
"""Test that an entry at offset 0 is not given a new offset"""
|
2016-11-26 03:15:52 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('025_pack_zero_size.dts')
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-spl': Offset 0x0 (0) overlaps "
|
2016-11-26 03:15:52 +00:00
|
|
|
"with previous entry '/binman/u-boot' ending at 0x4 (4)",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testPackUbootDtb(self):
|
|
|
|
"""Test that a device tree can be added to U-Boot"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('026_pack_u_boot_dtb.dts')
|
2016-11-26 03:15:52 +00:00
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA + U_BOOT_DTB_DATA, data)
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
def testPackX86RomNoSize(self):
|
|
|
|
"""Test that the end-at-4gb property requires a size property"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('027_pack_4gb_no_size.dts')
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertIn("Image '/binman': Section size must be provided when "
|
2016-11-26 03:15:53 +00:00
|
|
|
"using end-at-4gb", str(e.exception))
|
|
|
|
|
2018-09-03 16:05:07 +00:00
|
|
|
def test4gbAndSkipAtStartTogether(self):
|
|
|
|
"""Test that the end-at-4gb and skip-at-size property can't be used
|
|
|
|
together"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2019-08-24 13:23:02 +00:00
|
|
|
self._DoTestFile('098_4gb_and_skip_at_start_together.dts')
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertIn("Image '/binman': Provide either 'end-at-4gb' or "
|
2018-09-03 16:05:07 +00:00
|
|
|
"'skip-at-start'", str(e.exception))
|
|
|
|
|
2016-11-26 03:15:53 +00:00
|
|
|
def testPackX86RomOutside(self):
|
2018-08-01 21:22:37 +00:00
|
|
|
"""Test that the end-at-4gb property checks for offset boundaries"""
|
2016-11-26 03:15:53 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('028_pack_4gb_outside.dts')
|
2020-10-26 23:40:05 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Offset 0x0 (0) size 0x4 (4) "
|
|
|
|
"is outside the section '/binman' starting at "
|
|
|
|
'0xffffffe0 (4294967264) of size 0x20 (32)',
|
2016-11-26 03:15:53 +00:00
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testPackX86Rom(self):
|
|
|
|
"""Test that a basic x86 ROM can be created"""
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf()
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('029_x86_rom.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA + tools.get_bytes(0, 3) + U_BOOT_SPL_DATA +
|
|
|
|
tools.get_bytes(0, 2), data)
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
def testPackX86RomMeNoDesc(self):
|
|
|
|
"""Test that an invalid Intel descriptor entry is detected"""
|
2020-07-10 00:39:41 +00:00
|
|
|
try:
|
2020-07-25 21:11:19 +00:00
|
|
|
TestFunctional._MakeInputFile('descriptor-empty.bin', b'')
|
2020-07-10 00:39:41 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
2020-07-25 21:11:19 +00:00
|
|
|
self._DoTestFile('163_x86_rom_me_empty.dts')
|
2020-07-10 00:39:41 +00:00
|
|
|
self.assertIn("Node '/binman/intel-descriptor': Cannot find Intel Flash Descriptor (FD) signature",
|
|
|
|
str(e.exception))
|
|
|
|
finally:
|
|
|
|
self._SetupDescriptor()
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
def testPackX86RomBadDesc(self):
|
|
|
|
"""Test that the Intel requires a descriptor entry"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2019-08-24 13:23:01 +00:00
|
|
|
self._DoTestFile('030_x86_rom_me_no_desc.dts')
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertIn("Node '/binman/intel-me': No offset set with "
|
|
|
|
"offset-unset: should another entry provide this correct "
|
|
|
|
"offset?", str(e.exception))
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
def testPackX86RomMe(self):
|
|
|
|
"""Test that an x86 ROM with an ME region can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('031_x86_rom_me.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
expected_desc = tools.read_file(self.TestFile('descriptor.bin'))
|
2019-07-08 19:18:54 +00:00
|
|
|
if data[:0x1000] != expected_desc:
|
|
|
|
self.fail('Expected descriptor binary at start of image')
|
2016-11-26 03:15:53 +00:00
|
|
|
self.assertEqual(ME_DATA, data[0x1000:0x1000 + len(ME_DATA)])
|
|
|
|
|
|
|
|
def testPackVga(self):
|
|
|
|
"""Test that an image with a VGA binary can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('032_intel_vga.dts')
|
2016-11-26 03:15:53 +00:00
|
|
|
self.assertEqual(VGA_DATA, data[:len(VGA_DATA)])
|
|
|
|
|
|
|
|
def testPackStart16(self):
|
|
|
|
"""Test that an image with an x86 start16 region can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('033_x86_start16.dts')
|
2016-11-26 03:15:53 +00:00
|
|
|
self.assertEqual(X86_START16_DATA, data[:len(X86_START16_DATA)])
|
|
|
|
|
2018-09-03 16:05:08 +00:00
|
|
|
def testPackPowerpcMpc85xxBootpgResetvec(self):
|
|
|
|
"""Test that an image with powerpc-mpc85xx-bootpg-resetvec can be
|
|
|
|
created"""
|
2019-08-24 13:23:02 +00:00
|
|
|
data = self._DoReadFile('150_powerpc_mpc85xx_bootpg_resetvec.dts')
|
2018-09-03 16:05:08 +00:00
|
|
|
self.assertEqual(PPC_MPC85XX_BR_DATA, data[:len(PPC_MPC85XX_BR_DATA)])
|
|
|
|
|
2018-07-06 16:27:17 +00:00
|
|
|
def _RunMicrocodeTest(self, dts_fname, nodtb_data, ucode_second=False):
|
2018-07-06 16:27:16 +00:00
|
|
|
"""Handle running a test for insertion of microcode
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dts_fname: Name of test .dts file
|
|
|
|
nodtb_data: Data that we expect in the first section
|
2018-07-06 16:27:17 +00:00
|
|
|
ucode_second: True if the microsecond entry is second instead of
|
|
|
|
third
|
2018-07-06 16:27:16 +00:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
Tuple:
|
|
|
|
Contents of first region (U-Boot or SPL)
|
2018-08-01 21:22:37 +00:00
|
|
|
Offset and size components of microcode pointer, as inserted
|
2018-07-06 16:27:16 +00:00
|
|
|
in the above (two 4-byte words)
|
|
|
|
"""
|
2017-11-13 04:52:27 +00:00
|
|
|
data = self._DoReadFile(dts_fname, True)
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
# Now check the device tree has no microcode
|
2018-07-06 16:27:17 +00:00
|
|
|
if ucode_second:
|
|
|
|
ucode_content = data[len(nodtb_data):]
|
|
|
|
ucode_pos = len(nodtb_data)
|
|
|
|
dtb_with_ucode = ucode_content[16:]
|
|
|
|
fdt_len = self.GetFdtLen(dtb_with_ucode)
|
|
|
|
else:
|
|
|
|
dtb_with_ucode = data[len(nodtb_data):]
|
|
|
|
fdt_len = self.GetFdtLen(dtb_with_ucode)
|
|
|
|
ucode_content = dtb_with_ucode[fdt_len:]
|
|
|
|
ucode_pos = len(nodtb_data) + fdt_len
|
2022-01-29 21:14:04 +00:00
|
|
|
fname = tools.get_output_filename('test.dtb')
|
2016-11-26 03:15:53 +00:00
|
|
|
with open(fname, 'wb') as fd:
|
2018-07-06 16:27:16 +00:00
|
|
|
fd.write(dtb_with_ucode)
|
2017-05-27 13:38:29 +00:00
|
|
|
dtb = fdt.FdtScan(fname)
|
|
|
|
ucode = dtb.GetNode('/microcode')
|
2016-11-26 03:15:53 +00:00
|
|
|
self.assertTrue(ucode)
|
|
|
|
for node in ucode.subnodes:
|
|
|
|
self.assertFalse(node.props.get('data'))
|
|
|
|
|
|
|
|
# Check that the microcode appears immediately after the Fdt
|
|
|
|
# This matches the concatenation of the data properties in
|
2017-11-13 04:52:26 +00:00
|
|
|
# the /microcode/update@xxx nodes in 34_x86_ucode.dts.
|
2016-11-26 03:15:53 +00:00
|
|
|
ucode_data = struct.pack('>4L', 0x12345678, 0x12345679, 0xabcd0000,
|
|
|
|
0x78235609)
|
2018-07-06 16:27:16 +00:00
|
|
|
self.assertEqual(ucode_data, ucode_content[:len(ucode_data)])
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
# Check that the microcode pointer was inserted. It should match the
|
2018-08-01 21:22:37 +00:00
|
|
|
# expected offset and size
|
2016-11-26 03:15:53 +00:00
|
|
|
pos_and_size = struct.pack('<2L', 0xfffffe00 + ucode_pos,
|
|
|
|
len(ucode_data))
|
2018-07-06 16:27:17 +00:00
|
|
|
u_boot = data[:len(nodtb_data)]
|
|
|
|
return u_boot, pos_and_size
|
2017-11-13 04:52:27 +00:00
|
|
|
|
|
|
|
def testPackUbootMicrocode(self):
|
|
|
|
"""Test that x86 microcode can be handled correctly
|
|
|
|
|
|
|
|
We expect to see the following in the image, in order:
|
|
|
|
u-boot-nodtb.bin with a microcode pointer inserted at the correct
|
|
|
|
place
|
|
|
|
u-boot.dtb with the microcode removed
|
|
|
|
the microcode
|
|
|
|
"""
|
2018-10-01 18:22:30 +00:00
|
|
|
first, pos_and_size = self._RunMicrocodeTest('034_x86_ucode.dts',
|
2017-11-13 04:52:27 +00:00
|
|
|
U_BOOT_NODTB_DATA)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'nodtb with microcode' + pos_and_size +
|
|
|
|
b' somewhere in here', first)
|
2016-11-26 03:15:53 +00:00
|
|
|
|
2017-05-27 13:38:26 +00:00
|
|
|
def _RunPackUbootSingleMicrocode(self):
|
2016-11-26 03:15:53 +00:00
|
|
|
"""Test that x86 microcode can be handled correctly
|
|
|
|
|
|
|
|
We expect to see the following in the image, in order:
|
|
|
|
u-boot-nodtb.bin with a microcode pointer inserted at the correct
|
|
|
|
place
|
|
|
|
u-boot.dtb with the microcode
|
|
|
|
an empty microcode region
|
|
|
|
"""
|
|
|
|
# We need the libfdt library to run this test since only that allows
|
|
|
|
# finding the offset of a property. This is required by
|
|
|
|
# Entry_u_boot_dtb_with_ucode.ObtainContents().
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('035_x86_single_ucode.dts', True)
|
2016-11-26 03:15:53 +00:00
|
|
|
|
|
|
|
second = data[len(U_BOOT_NODTB_DATA):]
|
|
|
|
|
|
|
|
fdt_len = self.GetFdtLen(second)
|
|
|
|
third = second[fdt_len:]
|
|
|
|
second = second[:fdt_len]
|
|
|
|
|
2017-05-27 13:38:26 +00:00
|
|
|
ucode_data = struct.pack('>2L', 0x12345678, 0x12345679)
|
|
|
|
self.assertIn(ucode_data, second)
|
|
|
|
ucode_pos = second.find(ucode_data) + len(U_BOOT_NODTB_DATA)
|
2016-11-26 03:15:53 +00:00
|
|
|
|
2017-05-27 13:38:26 +00:00
|
|
|
# Check that the microcode pointer was inserted. It should match the
|
2018-08-01 21:22:37 +00:00
|
|
|
# expected offset and size
|
2017-05-27 13:38:26 +00:00
|
|
|
pos_and_size = struct.pack('<2L', 0xfffffe00 + ucode_pos,
|
|
|
|
len(ucode_data))
|
|
|
|
first = data[:len(U_BOOT_NODTB_DATA)]
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'nodtb with microcode' + pos_and_size +
|
|
|
|
b' somewhere in here', first)
|
2016-11-26 03:15:54 +00:00
|
|
|
|
2016-11-26 03:15:55 +00:00
|
|
|
def testPackUbootSingleMicrocode(self):
|
|
|
|
"""Test that x86 microcode can be handled correctly with fdt_normal.
|
|
|
|
"""
|
2017-05-27 13:38:26 +00:00
|
|
|
self._RunPackUbootSingleMicrocode()
|
2016-11-26 03:15:55 +00:00
|
|
|
|
2016-11-26 03:15:54 +00:00
|
|
|
def testUBootImg(self):
|
|
|
|
"""Test that u-boot.img can be put in a file"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('036_u_boot_img.dts')
|
2016-11-26 03:15:54 +00:00
|
|
|
self.assertEqual(U_BOOT_IMG_DATA, data)
|
2016-11-26 03:15:55 +00:00
|
|
|
|
|
|
|
def testNoMicrocode(self):
|
|
|
|
"""Test that a missing microcode region is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('037_x86_no_ucode.dts', True)
|
2016-11-26 03:15:55 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-dtb-with-ucode': No /microcode "
|
|
|
|
"node found in ", str(e.exception))
|
|
|
|
|
|
|
|
def testMicrocodeWithoutNode(self):
|
|
|
|
"""Test that a missing u-boot-dtb-with-ucode node is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('038_x86_ucode_missing_node.dts', True)
|
2016-11-26 03:15:55 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-with-ucode-ptr': Cannot find "
|
|
|
|
"microcode region u-boot-dtb-with-ucode", str(e.exception))
|
|
|
|
|
|
|
|
def testMicrocodeWithoutNode2(self):
|
|
|
|
"""Test that a missing u-boot-ucode node is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('039_x86_ucode_missing_node2.dts', True)
|
2016-11-26 03:15:55 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-with-ucode-ptr': Cannot find "
|
|
|
|
"microcode region u-boot-ucode", str(e.exception))
|
|
|
|
|
|
|
|
def testMicrocodeWithoutPtrInElf(self):
|
|
|
|
"""Test that a U-Boot binary without the microcode symbol is detected"""
|
|
|
|
# ELF file without a '_dt_ucode_base_size' symbol
|
|
|
|
try:
|
2019-08-24 13:22:55 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot',
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.read_file(self.ElfTestFile('u_boot_no_ucode_ptr')))
|
2016-11-26 03:15:55 +00:00
|
|
|
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2017-05-27 13:38:26 +00:00
|
|
|
self._RunPackUbootSingleMicrocode()
|
2016-11-26 03:15:55 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-with-ucode-ptr': Cannot locate "
|
|
|
|
"_dt_ucode_base_size symbol in u-boot", str(e.exception))
|
|
|
|
|
|
|
|
finally:
|
|
|
|
# Put the original file back
|
2019-08-24 13:22:54 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot',
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.read_file(self.ElfTestFile('u_boot_ucode_ptr')))
|
2016-11-26 03:15:55 +00:00
|
|
|
|
|
|
|
def testMicrocodeNotInImage(self):
|
|
|
|
"""Test that microcode must be placed within the image"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('040_x86_ucode_not_in_image.dts', True)
|
2016-11-26 03:15:55 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-with-ucode-ptr': Microcode "
|
|
|
|
"pointer _dt_ucode_base_size at fffffe14 is outside the "
|
2018-06-01 15:38:14 +00:00
|
|
|
"section ranging from 00000000 to 0000002e", str(e.exception))
|
2016-11-26 03:15:55 +00:00
|
|
|
|
|
|
|
def testWithoutMicrocode(self):
|
|
|
|
"""Test that we can cope with an image without microcode (e.g. qemu)"""
|
2019-08-24 13:22:55 +00:00
|
|
|
TestFunctional._MakeInputFile('u-boot',
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.read_file(self.ElfTestFile('u_boot_no_ucode_ptr')))
|
2018-10-01 18:22:30 +00:00
|
|
|
data, dtb, _, _ = self._DoReadFileDtb('044_x86_optional_ucode.dts', True)
|
2016-11-26 03:15:55 +00:00
|
|
|
|
|
|
|
# Now check the device tree has no microcode
|
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA, data[:len(U_BOOT_NODTB_DATA)])
|
|
|
|
second = data[len(U_BOOT_NODTB_DATA):]
|
|
|
|
|
|
|
|
fdt_len = self.GetFdtLen(second)
|
|
|
|
self.assertEqual(dtb, second[:fdt_len])
|
|
|
|
|
|
|
|
used_len = len(U_BOOT_NODTB_DATA) + fdt_len
|
|
|
|
third = data[used_len:]
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0, 0x200 - used_len), third)
|
2016-11-26 03:15:55 +00:00
|
|
|
|
|
|
|
def testUnknownPosSize(self):
|
|
|
|
"""Test that microcode must be placed within the image"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('041_unknown_pos_size.dts', True)
|
2018-08-01 21:22:37 +00:00
|
|
|
self.assertIn("Section '/binman': Unable to set offset/size for unknown "
|
2016-11-26 03:15:55 +00:00
|
|
|
"entry 'invalid-entry'", str(e.exception))
|
2016-11-26 03:15:56 +00:00
|
|
|
|
|
|
|
def testPackFsp(self):
|
|
|
|
"""Test that an image with a FSP binary can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('042_intel_fsp.dts')
|
2016-11-26 03:15:56 +00:00
|
|
|
self.assertEqual(FSP_DATA, data[:len(FSP_DATA)])
|
|
|
|
|
|
|
|
def testPackCmc(self):
|
2017-08-16 05:41:54 +00:00
|
|
|
"""Test that an image with a CMC binary can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('043_intel_cmc.dts')
|
2016-11-26 03:15:56 +00:00
|
|
|
self.assertEqual(CMC_DATA, data[:len(CMC_DATA)])
|
2017-08-16 05:41:54 +00:00
|
|
|
|
|
|
|
def testPackVbt(self):
|
|
|
|
"""Test that an image with a VBT binary can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('046_intel_vbt.dts')
|
2017-08-16 05:41:54 +00:00
|
|
|
self.assertEqual(VBT_DATA, data[:len(VBT_DATA)])
|
2017-11-13 04:52:22 +00:00
|
|
|
|
2017-11-13 04:52:25 +00:00
|
|
|
def testSplBssPad(self):
|
|
|
|
"""Test that we can pad SPL's BSS with zeros"""
|
2017-11-13 04:52:27 +00:00
|
|
|
# ELF file with a '__bss_size' symbol
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf()
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('047_spl_bss_pad.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(U_BOOT_SPL_DATA + tools.get_bytes(0, 10) + U_BOOT_DATA,
|
2019-05-14 21:53:47 +00:00
|
|
|
data)
|
2017-11-13 04:52:25 +00:00
|
|
|
|
2018-10-02 03:12:42 +00:00
|
|
|
def testSplBssPadMissing(self):
|
|
|
|
"""Test that a missing symbol is detected"""
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf('u_boot_ucode_ptr')
|
2017-11-14 01:54:54 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('047_spl_bss_pad.dts')
|
2017-11-14 01:54:54 +00:00
|
|
|
self.assertIn('Expected __bss_size symbol in spl/u-boot-spl',
|
|
|
|
str(e.exception))
|
|
|
|
|
2017-11-13 04:52:26 +00:00
|
|
|
def testPackStart16Spl(self):
|
2018-09-14 10:57:10 +00:00
|
|
|
"""Test that an image with an x86 start16 SPL region can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('048_x86_start16_spl.dts')
|
2017-11-13 04:52:26 +00:00
|
|
|
self.assertEqual(X86_START16_SPL_DATA, data[:len(X86_START16_SPL_DATA)])
|
|
|
|
|
2018-07-06 16:27:17 +00:00
|
|
|
def _PackUbootSplMicrocode(self, dts, ucode_second=False):
|
|
|
|
"""Helper function for microcode tests
|
2017-11-13 04:52:27 +00:00
|
|
|
|
|
|
|
We expect to see the following in the image, in order:
|
|
|
|
u-boot-spl-nodtb.bin with a microcode pointer inserted at the
|
|
|
|
correct place
|
|
|
|
u-boot.dtb with the microcode removed
|
|
|
|
the microcode
|
2018-07-06 16:27:17 +00:00
|
|
|
|
|
|
|
Args:
|
|
|
|
dts: Device tree file to use for test
|
|
|
|
ucode_second: True if the microsecond entry is second instead of
|
|
|
|
third
|
2017-11-13 04:52:27 +00:00
|
|
|
"""
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf('u_boot_ucode_ptr')
|
2018-07-06 16:27:17 +00:00
|
|
|
first, pos_and_size = self._RunMicrocodeTest(dts, U_BOOT_SPL_NODTB_DATA,
|
|
|
|
ucode_second=ucode_second)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'splnodtb with microc' + pos_and_size +
|
|
|
|
b'ter somewhere in here', first)
|
2017-11-13 04:52:27 +00:00
|
|
|
|
2018-07-06 16:27:17 +00:00
|
|
|
def testPackUbootSplMicrocode(self):
|
|
|
|
"""Test that x86 microcode can be handled correctly in SPL"""
|
2018-10-01 18:22:30 +00:00
|
|
|
self._PackUbootSplMicrocode('049_x86_ucode_spl.dts')
|
2018-07-06 16:27:17 +00:00
|
|
|
|
|
|
|
def testPackUbootSplMicrocodeReorder(self):
|
|
|
|
"""Test that order doesn't matter for microcode entries
|
|
|
|
|
|
|
|
This is the same as testPackUbootSplMicrocode but when we process the
|
|
|
|
u-boot-ucode entry we have not yet seen the u-boot-dtb-with-ucode
|
|
|
|
entry, so we reply on binman to try later.
|
|
|
|
"""
|
2018-10-01 18:22:30 +00:00
|
|
|
self._PackUbootSplMicrocode('058_x86_ucode_spl_needs_retry.dts',
|
2018-07-06 16:27:17 +00:00
|
|
|
ucode_second=True)
|
|
|
|
|
2017-11-13 04:52:28 +00:00
|
|
|
def testPackMrc(self):
|
|
|
|
"""Test that an image with an MRC binary can be created"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('050_intel_mrc.dts')
|
2017-11-13 04:52:28 +00:00
|
|
|
self.assertEqual(MRC_DATA, data[:len(MRC_DATA)])
|
|
|
|
|
2017-11-14 01:54:55 +00:00
|
|
|
def testSplDtb(self):
|
|
|
|
"""Test that an image with spl/u-boot-spl.dtb can be created"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('051_u_boot_spl_dtb.dts')
|
2017-11-14 01:54:55 +00:00
|
|
|
self.assertEqual(U_BOOT_SPL_DTB_DATA, data[:len(U_BOOT_SPL_DTB_DATA)])
|
|
|
|
|
2017-11-14 01:54:56 +00:00
|
|
|
def testSplNoDtb(self):
|
|
|
|
"""Test that an image with spl/u-boot-spl-nodtb.bin can be created"""
|
2021-04-24 20:39:32 +00:00
|
|
|
self._SetupSplElf()
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('052_u_boot_spl_nodtb.dts')
|
2017-11-14 01:54:56 +00:00
|
|
|
self.assertEqual(U_BOOT_SPL_NODTB_DATA, data[:len(U_BOOT_SPL_NODTB_DATA)])
|
|
|
|
|
2021-03-21 05:24:30 +00:00
|
|
|
def checkSymbols(self, dts, base_data, u_boot_offset, entry_args=None,
|
|
|
|
use_expanded=False):
|
2021-03-18 07:24:56 +00:00
|
|
|
"""Check the image contains the expected symbol values
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dts: Device tree file to use for test
|
|
|
|
base_data: Data before and after 'u-boot' section
|
|
|
|
u_boot_offset: Offset of 'u-boot' section in image
|
2021-03-21 05:24:30 +00:00
|
|
|
entry_args: Dict of entry args to supply to binman
|
|
|
|
key: arg name
|
|
|
|
value: value of that arg
|
|
|
|
use_expanded: True to use expanded entries where available, e.g.
|
|
|
|
'u-boot-expanded' instead of 'u-boot'
|
2021-03-18 07:24:56 +00:00
|
|
|
"""
|
2019-08-24 13:22:56 +00:00
|
|
|
elf_fname = self.ElfTestFile('u_boot_binman_syms')
|
binman: Support accessing binman tables at run time
Binman construct images consisting of multiple binary files. These files
sometimes need to know (at run timme) where their peers are located. For
example, SPL may want to know where U-Boot is located in the image, so
that it can jump to U-Boot correctly on boot.
In general the positions where the binaries end up after binman has
finished packing them cannot be known at compile time. One reason for
this is that binman does not know the size of the binaries until
everything is compiled, linked and converted to binaries with objcopy.
To make this work, we add a feature to binman which checks each binary
for symbol names starting with '_binman'. These are then decoded to figure
out which entry and property they refer to. Then binman writes the value
of this symbol into the appropriate binary. With this, the symbol will
have the correct value at run time.
Macros are used to make this easier to use. As an example, this declares
a symbol that will access the 'u-boot-spl' entry to find the 'pos' value
(i.e. the position of SPL in the image):
binman_sym_declare(unsigned long, u_boot_spl, pos);
This converts to a symbol called '_binman_u_boot_spl_prop_pos' in any
binary that includes it. Binman then updates the value in that binary,
ensuring that it can be accessed at runtime with:
ulong u_boot_pos = binman_sym(ulong, u_boot_spl, pos);
This assigns the variable u_boot_pos to the position of SPL in the image.
Signed-off-by: Simon Glass <sjg@chromium.org>
2017-11-14 01:55:01 +00:00
|
|
|
syms = elf.GetSymbols(elf_fname, ['binman', 'image'])
|
|
|
|
addr = elf.GetSymbolAddress(elf_fname, '__image_copy_start')
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
self.assertEqual(syms['_binman_sym_magic'].address, addr)
|
2021-03-18 07:24:56 +00:00
|
|
|
self.assertEqual(syms['_binman_u_boot_spl_any_prop_offset'].address,
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
addr + 4)
|
binman: Support accessing binman tables at run time
Binman construct images consisting of multiple binary files. These files
sometimes need to know (at run timme) where their peers are located. For
example, SPL may want to know where U-Boot is located in the image, so
that it can jump to U-Boot correctly on boot.
In general the positions where the binaries end up after binman has
finished packing them cannot be known at compile time. One reason for
this is that binman does not know the size of the binaries until
everything is compiled, linked and converted to binaries with objcopy.
To make this work, we add a feature to binman which checks each binary
for symbol names starting with '_binman'. These are then decoded to figure
out which entry and property they refer to. Then binman writes the value
of this symbol into the appropriate binary. With this, the symbol will
have the correct value at run time.
Macros are used to make this easier to use. As an example, this declares
a symbol that will access the 'u-boot-spl' entry to find the 'pos' value
(i.e. the position of SPL in the image):
binman_sym_declare(unsigned long, u_boot_spl, pos);
This converts to a symbol called '_binman_u_boot_spl_prop_pos' in any
binary that includes it. Binman then updates the value in that binary,
ensuring that it can be accessed at runtime with:
ulong u_boot_pos = binman_sym(ulong, u_boot_spl, pos);
This assigns the variable u_boot_pos to the position of SPL in the image.
Signed-off-by: Simon Glass <sjg@chromium.org>
2017-11-14 01:55:01 +00:00
|
|
|
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf('u_boot_binman_syms')
|
2021-03-21 05:24:30 +00:00
|
|
|
data = self._DoReadFileDtb(dts, entry_args=entry_args,
|
|
|
|
use_expanded=use_expanded)[0]
|
2021-03-18 07:24:56 +00:00
|
|
|
# The image should contain the symbols from u_boot_binman_syms.c
|
|
|
|
# Note that image_pos is adjusted by the base address of the image,
|
|
|
|
# which is 0x10 in our test image
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
sym_values = struct.pack('<LLQLL', elf.BINMAN_SYM_MAGIC_VALUE,
|
|
|
|
0x00, u_boot_offset + len(U_BOOT_DATA),
|
2021-03-18 07:24:56 +00:00
|
|
|
0x10 + u_boot_offset, 0x04)
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
expected = (sym_values + base_data[24:] +
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.get_bytes(0xff, 1) + U_BOOT_DATA + sym_values +
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
base_data[24:])
|
binman: Support accessing binman tables at run time
Binman construct images consisting of multiple binary files. These files
sometimes need to know (at run timme) where their peers are located. For
example, SPL may want to know where U-Boot is located in the image, so
that it can jump to U-Boot correctly on boot.
In general the positions where the binaries end up after binman has
finished packing them cannot be known at compile time. One reason for
this is that binman does not know the size of the binaries until
everything is compiled, linked and converted to binaries with objcopy.
To make this work, we add a feature to binman which checks each binary
for symbol names starting with '_binman'. These are then decoded to figure
out which entry and property they refer to. Then binman writes the value
of this symbol into the appropriate binary. With this, the symbol will
have the correct value at run time.
Macros are used to make this easier to use. As an example, this declares
a symbol that will access the 'u-boot-spl' entry to find the 'pos' value
(i.e. the position of SPL in the image):
binman_sym_declare(unsigned long, u_boot_spl, pos);
This converts to a symbol called '_binman_u_boot_spl_prop_pos' in any
binary that includes it. Binman then updates the value in that binary,
ensuring that it can be accessed at runtime with:
ulong u_boot_pos = binman_sym(ulong, u_boot_spl, pos);
This assigns the variable u_boot_pos to the position of SPL in the image.
Signed-off-by: Simon Glass <sjg@chromium.org>
2017-11-14 01:55:01 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2021-03-18 07:24:56 +00:00
|
|
|
def testSymbols(self):
|
|
|
|
"""Test binman can assign symbols embedded in U-Boot"""
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
self.checkSymbols('053_symbols.dts', U_BOOT_SPL_DATA, 0x1c)
|
2021-03-18 07:24:56 +00:00
|
|
|
|
|
|
|
def testSymbolsNoDtb(self):
|
|
|
|
"""Test binman can assign symbols embedded in U-Boot SPL"""
|
2021-03-21 05:24:29 +00:00
|
|
|
self.checkSymbols('196_symbols_nodtb.dts',
|
2021-03-18 07:24:56 +00:00
|
|
|
U_BOOT_SPL_NODTB_DATA + U_BOOT_SPL_DTB_DATA,
|
|
|
|
0x38)
|
|
|
|
|
2018-06-01 15:38:11 +00:00
|
|
|
def testPackUnitAddress(self):
|
|
|
|
"""Test that we support multiple binaries with the same name"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('054_unit_address.dts')
|
2018-06-01 15:38:11 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA + U_BOOT_DATA, data)
|
|
|
|
|
2018-06-01 15:38:16 +00:00
|
|
|
def testSections(self):
|
|
|
|
"""Basic test of sections"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('055_sections.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (U_BOOT_DATA + tools.get_bytes(ord('!'), 12) +
|
|
|
|
U_BOOT_DATA + tools.get_bytes(ord('a'), 12) +
|
|
|
|
U_BOOT_DATA + tools.get_bytes(ord('&'), 4))
|
2018-06-01 15:38:16 +00:00
|
|
|
self.assertEqual(expected, data)
|
2017-11-13 04:52:22 +00:00
|
|
|
|
2018-06-01 15:38:20 +00:00
|
|
|
def testMap(self):
|
|
|
|
"""Tests outputting a map of the images"""
|
2018-10-01 18:22:30 +00:00
|
|
|
_, _, map_data, _ = self._DoReadFileDtb('055_sections.dts', map=True)
|
2018-07-17 19:25:49 +00:00
|
|
|
self.assertEqual('''ImagePos Offset Size Name
|
|
|
|
00000000 00000000 00000028 main-section
|
|
|
|
00000000 00000000 00000010 section@0
|
|
|
|
00000000 00000000 00000004 u-boot
|
|
|
|
00000010 00000010 00000010 section@1
|
|
|
|
00000010 00000000 00000004 u-boot
|
|
|
|
00000020 00000020 00000004 section@2
|
|
|
|
00000020 00000000 00000004 u-boot
|
2018-06-01 15:38:20 +00:00
|
|
|
''', map_data)
|
|
|
|
|
2018-06-01 15:38:21 +00:00
|
|
|
def testNamePrefix(self):
|
|
|
|
"""Tests that name prefixes are used"""
|
2018-10-01 18:22:30 +00:00
|
|
|
_, _, map_data, _ = self._DoReadFileDtb('056_name_prefix.dts', map=True)
|
2018-07-17 19:25:49 +00:00
|
|
|
self.assertEqual('''ImagePos Offset Size Name
|
|
|
|
00000000 00000000 00000028 main-section
|
|
|
|
00000000 00000000 00000010 section@0
|
|
|
|
00000000 00000000 00000004 ro-u-boot
|
|
|
|
00000010 00000010 00000010 section@1
|
|
|
|
00000010 00000000 00000004 rw-u-boot
|
2018-06-01 15:38:21 +00:00
|
|
|
''', map_data)
|
|
|
|
|
2018-07-06 16:27:17 +00:00
|
|
|
def testUnknownContents(self):
|
|
|
|
"""Test that obtaining the contents works as expected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('057_unknown_contents.dts', True)
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertIn("Image '/binman': Internal error: Could not complete "
|
2020-04-18 00:09:03 +00:00
|
|
|
"processing of contents: remaining ["
|
|
|
|
"<binman.etype._testing.Entry__testing ", str(e.exception))
|
2018-07-06 16:27:17 +00:00
|
|
|
|
2018-07-06 16:27:19 +00:00
|
|
|
def testBadChangeSize(self):
|
|
|
|
"""Test that trying to change the size of an entry fails"""
|
2019-07-08 20:25:37 +00:00
|
|
|
try:
|
|
|
|
state.SetAllowEntryExpansion(False)
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('059_change_size.dts', True)
|
2019-07-20 18:23:57 +00:00
|
|
|
self.assertIn("Node '/binman/_testing': Cannot update entry size from 2 to 3",
|
2019-07-08 20:25:37 +00:00
|
|
|
str(e.exception))
|
|
|
|
finally:
|
|
|
|
state.SetAllowEntryExpansion(True)
|
2018-07-06 16:27:19 +00:00
|
|
|
|
2018-07-06 16:27:42 +00:00
|
|
|
def testUpdateFdt(self):
|
2018-08-01 21:22:37 +00:00
|
|
|
"""Test that we can update the device tree with offset/size info"""
|
2018-10-01 18:22:30 +00:00
|
|
|
_, _, _, out_dtb_fname = self._DoReadFileDtb('060_fdt_update.dts',
|
2018-07-06 16:27:42 +00:00
|
|
|
update_dtb=True)
|
2018-07-17 19:25:52 +00:00
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
2019-07-20 18:23:51 +00:00
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS + REPACK_DTB_PROPS)
|
2018-07-06 16:27:42 +00:00
|
|
|
self.assertEqual({
|
2018-08-01 21:22:42 +00:00
|
|
|
'image-pos': 0,
|
2018-07-17 19:25:28 +00:00
|
|
|
'offset': 0,
|
2018-08-01 21:22:37 +00:00
|
|
|
'_testing:offset': 32,
|
2019-07-20 18:23:57 +00:00
|
|
|
'_testing:size': 2,
|
2018-08-01 21:22:42 +00:00
|
|
|
'_testing:image-pos': 32,
|
2018-08-01 21:22:37 +00:00
|
|
|
'section@0/u-boot:offset': 0,
|
2018-07-06 16:27:42 +00:00
|
|
|
'section@0/u-boot:size': len(U_BOOT_DATA),
|
2018-08-01 21:22:42 +00:00
|
|
|
'section@0/u-boot:image-pos': 0,
|
2018-08-01 21:22:37 +00:00
|
|
|
'section@0:offset': 0,
|
2018-07-06 16:27:42 +00:00
|
|
|
'section@0:size': 16,
|
2018-08-01 21:22:42 +00:00
|
|
|
'section@0:image-pos': 0,
|
2018-07-06 16:27:42 +00:00
|
|
|
|
2018-08-01 21:22:37 +00:00
|
|
|
'section@1/u-boot:offset': 0,
|
2018-07-06 16:27:42 +00:00
|
|
|
'section@1/u-boot:size': len(U_BOOT_DATA),
|
2018-08-01 21:22:42 +00:00
|
|
|
'section@1/u-boot:image-pos': 16,
|
2018-08-01 21:22:37 +00:00
|
|
|
'section@1:offset': 16,
|
2018-07-06 16:27:42 +00:00
|
|
|
'section@1:size': 16,
|
2018-08-01 21:22:42 +00:00
|
|
|
'section@1:image-pos': 16,
|
2018-07-06 16:27:42 +00:00
|
|
|
'size': 40
|
|
|
|
}, props)
|
|
|
|
|
|
|
|
def testUpdateFdtBad(self):
|
|
|
|
"""Test that we detect when ProcessFdt never completes"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('061_fdt_update_bad.dts', update_dtb=True)
|
2018-07-06 16:27:42 +00:00
|
|
|
self.assertIn('Could not complete processing of Fdt: remaining '
|
2020-04-18 00:09:03 +00:00
|
|
|
'[<binman.etype._testing.Entry__testing',
|
|
|
|
str(e.exception))
|
2018-07-06 16:27:19 +00:00
|
|
|
|
2018-07-17 19:25:32 +00:00
|
|
|
def testEntryArgs(self):
|
|
|
|
"""Test passing arguments to entries from the command line"""
|
|
|
|
entry_args = {
|
|
|
|
'test-str-arg': 'test1',
|
|
|
|
'test-int-arg': '456',
|
|
|
|
}
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('062_entry_args.dts', entry_args=entry_args)
|
2018-07-17 19:25:32 +00:00
|
|
|
self.assertIn('image', control.images)
|
|
|
|
entry = control.images['image'].GetEntries()['_testing']
|
|
|
|
self.assertEqual('test0', entry.test_str_fdt)
|
|
|
|
self.assertEqual('test1', entry.test_str_arg)
|
|
|
|
self.assertEqual(123, entry.test_int_fdt)
|
|
|
|
self.assertEqual(456, entry.test_int_arg)
|
|
|
|
|
|
|
|
def testEntryArgsMissing(self):
|
|
|
|
"""Test missing arguments and properties"""
|
|
|
|
entry_args = {
|
|
|
|
'test-int-arg': '456',
|
|
|
|
}
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('063_entry_args_missing.dts', entry_args=entry_args)
|
2018-07-17 19:25:32 +00:00
|
|
|
entry = control.images['image'].GetEntries()['_testing']
|
|
|
|
self.assertEqual('test0', entry.test_str_fdt)
|
|
|
|
self.assertEqual(None, entry.test_str_arg)
|
|
|
|
self.assertEqual(None, entry.test_int_fdt)
|
|
|
|
self.assertEqual(456, entry.test_int_arg)
|
|
|
|
|
|
|
|
def testEntryArgsRequired(self):
|
|
|
|
"""Test missing arguments and properties"""
|
|
|
|
entry_args = {
|
|
|
|
'test-int-arg': '456',
|
|
|
|
}
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('064_entry_args_required.dts')
|
2020-09-01 11:13:54 +00:00
|
|
|
self.assertIn("Node '/binman/_testing': "
|
|
|
|
'Missing required properties/entry args: test-str-arg, '
|
|
|
|
'test-int-fdt, test-int-arg',
|
2018-07-17 19:25:32 +00:00
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testEntryArgsInvalidFormat(self):
|
|
|
|
"""Test that an invalid entry-argument format is detected"""
|
2019-07-08 20:25:29 +00:00
|
|
|
args = ['build', '-d', self.TestFile('064_entry_args_required.dts'),
|
|
|
|
'-ano-value']
|
2018-07-17 19:25:32 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoBinman(*args)
|
|
|
|
self.assertIn("Invalid entry arguemnt 'no-value'", str(e.exception))
|
|
|
|
|
|
|
|
def testEntryArgsInvalidInteger(self):
|
|
|
|
"""Test that an invalid entry-argument integer is detected"""
|
|
|
|
entry_args = {
|
|
|
|
'test-int-arg': 'abc',
|
|
|
|
}
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('062_entry_args.dts', entry_args=entry_args)
|
2018-07-17 19:25:32 +00:00
|
|
|
self.assertIn("Node '/binman/_testing': Cannot convert entry arg "
|
|
|
|
"'test-int-arg' (value 'abc') to integer",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testEntryArgsInvalidDatatype(self):
|
|
|
|
"""Test that an invalid entry-argument datatype is detected
|
|
|
|
|
|
|
|
This test could be written in entry_test.py except that it needs
|
|
|
|
access to control.entry_args, which seems more than that module should
|
|
|
|
be able to see.
|
|
|
|
"""
|
|
|
|
entry_args = {
|
|
|
|
'test-bad-datatype-arg': '12',
|
|
|
|
}
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('065_entry_args_unknown_datatype.dts',
|
2018-07-17 19:25:32 +00:00
|
|
|
entry_args=entry_args)
|
|
|
|
self.assertIn('GetArg() internal error: Unknown data type ',
|
|
|
|
str(e.exception))
|
|
|
|
|
2018-07-17 19:25:33 +00:00
|
|
|
def testText(self):
|
|
|
|
"""Test for a text entry type"""
|
|
|
|
entry_args = {
|
|
|
|
'test-id': TEXT_DATA,
|
|
|
|
'test-id2': TEXT_DATA2,
|
|
|
|
'test-id3': TEXT_DATA3,
|
|
|
|
}
|
2018-10-01 18:22:30 +00:00
|
|
|
data, _, _, _ = self._DoReadFileDtb('066_text.dts',
|
2018-07-17 19:25:33 +00:00
|
|
|
entry_args=entry_args)
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (tools.to_bytes(TEXT_DATA) +
|
|
|
|
tools.get_bytes(0, 8 - len(TEXT_DATA)) +
|
|
|
|
tools.to_bytes(TEXT_DATA2) + tools.to_bytes(TEXT_DATA3) +
|
2019-07-08 19:18:40 +00:00
|
|
|
b'some text' + b'more text')
|
2018-07-17 19:25:33 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2018-07-17 19:25:36 +00:00
|
|
|
def testEntryDocs(self):
|
|
|
|
"""Test for creation of entry documentation"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
2020-08-05 19:27:46 +00:00
|
|
|
control.WriteEntryDocs(control.GetEntryModules())
|
2018-07-17 19:25:36 +00:00
|
|
|
self.assertTrue(len(stdout.getvalue()) > 0)
|
|
|
|
|
|
|
|
def testEntryDocsMissing(self):
|
|
|
|
"""Test handling of missing entry documentation"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
2020-08-05 19:27:46 +00:00
|
|
|
control.WriteEntryDocs(control.GetEntryModules(), 'u_boot')
|
2018-07-17 19:25:36 +00:00
|
|
|
self.assertIn('Documentation is missing for modules: u_boot',
|
|
|
|
str(e.exception))
|
|
|
|
|
2018-07-17 19:25:38 +00:00
|
|
|
def testFmap(self):
|
|
|
|
"""Basic test of generation of a flashrom fmap"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('067_fmap.dts')
|
2018-07-17 19:25:38 +00:00
|
|
|
fhdr, fentries = fmap_util.DecodeFmap(data[32:])
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (U_BOOT_DATA + tools.get_bytes(ord('!'), 12) +
|
|
|
|
U_BOOT_DATA + tools.get_bytes(ord('a'), 12))
|
2018-07-17 19:25:38 +00:00
|
|
|
self.assertEqual(expected, data[:32])
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'__FMAP__', fhdr.signature)
|
2018-07-17 19:25:38 +00:00
|
|
|
self.assertEqual(1, fhdr.ver_major)
|
|
|
|
self.assertEqual(0, fhdr.ver_minor)
|
|
|
|
self.assertEqual(0, fhdr.base)
|
2021-04-02 22:05:10 +00:00
|
|
|
expect_size = fmap_util.FMAP_HEADER_LEN + fmap_util.FMAP_AREA_LEN * 5
|
2021-04-02 22:05:09 +00:00
|
|
|
self.assertEqual(16 + 16 + expect_size, fhdr.image_size)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'FMAP', fhdr.name)
|
2021-04-02 22:05:10 +00:00
|
|
|
self.assertEqual(5, fhdr.nareas)
|
2021-04-02 22:05:09 +00:00
|
|
|
fiter = iter(fentries)
|
|
|
|
|
2021-04-02 22:05:10 +00:00
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'SECTION0', fentry.name)
|
|
|
|
self.assertEqual(0, fentry.offset)
|
|
|
|
self.assertEqual(16, fentry.size)
|
|
|
|
self.assertEqual(0, fentry.flags)
|
|
|
|
|
2021-04-02 22:05:09 +00:00
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'RO_U_BOOT', fentry.name)
|
|
|
|
self.assertEqual(0, fentry.offset)
|
|
|
|
self.assertEqual(4, fentry.size)
|
|
|
|
self.assertEqual(0, fentry.flags)
|
|
|
|
|
2021-04-02 22:05:10 +00:00
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'SECTION1', fentry.name)
|
|
|
|
self.assertEqual(16, fentry.offset)
|
|
|
|
self.assertEqual(16, fentry.size)
|
|
|
|
self.assertEqual(0, fentry.flags)
|
|
|
|
|
2021-04-02 22:05:09 +00:00
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'RW_U_BOOT', fentry.name)
|
|
|
|
self.assertEqual(16, fentry.offset)
|
|
|
|
self.assertEqual(4, fentry.size)
|
|
|
|
self.assertEqual(0, fentry.flags)
|
|
|
|
|
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'FMAP', fentry.name)
|
|
|
|
self.assertEqual(32, fentry.offset)
|
|
|
|
self.assertEqual(expect_size, fentry.size)
|
|
|
|
self.assertEqual(0, fentry.flags)
|
2018-07-17 19:25:38 +00:00
|
|
|
|
2018-07-17 19:25:39 +00:00
|
|
|
def testBlobNamedByArg(self):
|
|
|
|
"""Test we can add a blob with the filename coming from an entry arg"""
|
|
|
|
entry_args = {
|
|
|
|
'cros-ec-rw-path': 'ecrw.bin',
|
|
|
|
}
|
2020-09-01 11:13:54 +00:00
|
|
|
self._DoReadFileDtb('068_blob_named_by_arg.dts', entry_args=entry_args)
|
2018-07-17 19:25:39 +00:00
|
|
|
|
2018-07-17 19:25:40 +00:00
|
|
|
def testFill(self):
|
|
|
|
"""Test for an fill entry type"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('069_fill.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = tools.get_bytes(0xff, 8) + tools.get_bytes(0, 8)
|
2018-07-17 19:25:40 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testFillNoSize(self):
|
|
|
|
"""Test for an fill entry type with no size"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('070_fill_no_size.dts')
|
2022-08-13 17:40:44 +00:00
|
|
|
self.assertIn("'fill' entry is missing properties: size",
|
2018-07-17 19:25:40 +00:00
|
|
|
str(e.exception))
|
|
|
|
|
2018-07-17 19:25:44 +00:00
|
|
|
def _HandleGbbCommand(self, pipe_list):
|
|
|
|
"""Fake calls to the futility utility"""
|
|
|
|
if pipe_list[0][0] == 'futility':
|
|
|
|
fname = pipe_list[0][-1]
|
|
|
|
# Append our GBB data to the file, which will happen every time the
|
|
|
|
# futility command is called.
|
2019-05-14 21:53:42 +00:00
|
|
|
with open(fname, 'ab') as fd:
|
2018-07-17 19:25:44 +00:00
|
|
|
fd.write(GBB_DATA)
|
|
|
|
return command.CommandResult()
|
|
|
|
|
|
|
|
def testGbb(self):
|
|
|
|
"""Test for the Chromium OS Google Binary Block"""
|
|
|
|
command.test_result = self._HandleGbbCommand
|
|
|
|
entry_args = {
|
|
|
|
'keydir': 'devkeys',
|
|
|
|
'bmpblk': 'bmpblk.bin',
|
|
|
|
}
|
2018-10-01 18:22:30 +00:00
|
|
|
data, _, _, _ = self._DoReadFileDtb('071_gbb.dts', entry_args=entry_args)
|
2018-07-17 19:25:44 +00:00
|
|
|
|
|
|
|
# Since futility
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (GBB_DATA + GBB_DATA + tools.get_bytes(0, 8) +
|
|
|
|
tools.get_bytes(0, 0x2180 - 16))
|
2018-07-17 19:25:44 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testGbbTooSmall(self):
|
|
|
|
"""Test for the Chromium OS Google Binary Block being large enough"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('072_gbb_too_small.dts')
|
2018-07-17 19:25:44 +00:00
|
|
|
self.assertIn("Node '/binman/gbb': GBB is too small",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testGbbNoSize(self):
|
|
|
|
"""Test for the Chromium OS Google Binary Block having a size"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('073_gbb_no_size.dts')
|
2018-07-17 19:25:44 +00:00
|
|
|
self.assertIn("Node '/binman/gbb': GBB must have a fixed size",
|
|
|
|
str(e.exception))
|
|
|
|
|
2022-01-10 03:14:09 +00:00
|
|
|
def testGbbMissing(self):
|
|
|
|
"""Test that binman still produces an image if futility is missing"""
|
|
|
|
entry_args = {
|
|
|
|
'keydir': 'devkeys',
|
|
|
|
}
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoTestFile('071_gbb.dts', force_missing_bintools='futility',
|
|
|
|
entry_args=entry_args)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing bintools.*: futility")
|
|
|
|
|
2018-07-17 19:25:47 +00:00
|
|
|
def _HandleVblockCommand(self, pipe_list):
|
2021-01-07 04:35:17 +00:00
|
|
|
"""Fake calls to the futility utility
|
|
|
|
|
|
|
|
The expected pipe is:
|
|
|
|
|
|
|
|
[('futility', 'vbutil_firmware', '--vblock',
|
|
|
|
'vblock.vblock', '--keyblock', 'devkeys/firmware.keyblock',
|
|
|
|
'--signprivate', 'devkeys/firmware_data_key.vbprivk',
|
|
|
|
'--version', '1', '--fv', 'input.vblock', '--kernelkey',
|
|
|
|
'devkeys/kernel_subkey.vbpubk', '--flags', '1')]
|
|
|
|
|
|
|
|
This writes to the output file (here, 'vblock.vblock'). If
|
|
|
|
self._hash_data is False, it writes VBLOCK_DATA, else it writes a hash
|
|
|
|
of the input data (here, 'input.vblock').
|
|
|
|
"""
|
2018-07-17 19:25:47 +00:00
|
|
|
if pipe_list[0][0] == 'futility':
|
|
|
|
fname = pipe_list[0][3]
|
2018-09-14 10:57:11 +00:00
|
|
|
with open(fname, 'wb') as fd:
|
2021-01-07 04:35:17 +00:00
|
|
|
if self._hash_data:
|
|
|
|
infile = pipe_list[0][11]
|
|
|
|
m = hashlib.sha256()
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(infile)
|
2021-01-07 04:35:17 +00:00
|
|
|
m.update(data)
|
|
|
|
fd.write(m.digest())
|
|
|
|
else:
|
|
|
|
fd.write(VBLOCK_DATA)
|
|
|
|
|
2018-07-17 19:25:47 +00:00
|
|
|
return command.CommandResult()
|
|
|
|
|
|
|
|
def testVblock(self):
|
|
|
|
"""Test for the Chromium OS Verified Boot Block"""
|
2021-01-07 04:35:17 +00:00
|
|
|
self._hash_data = False
|
2018-07-17 19:25:47 +00:00
|
|
|
command.test_result = self._HandleVblockCommand
|
|
|
|
entry_args = {
|
|
|
|
'keydir': 'devkeys',
|
|
|
|
}
|
2018-10-01 18:22:30 +00:00
|
|
|
data, _, _, _ = self._DoReadFileDtb('074_vblock.dts',
|
2018-07-17 19:25:47 +00:00
|
|
|
entry_args=entry_args)
|
|
|
|
expected = U_BOOT_DATA + VBLOCK_DATA + U_BOOT_DTB_DATA
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testVblockNoContent(self):
|
|
|
|
"""Test we detect a vblock which has no content to sign"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('075_vblock_no_content.dts')
|
2021-03-21 05:24:31 +00:00
|
|
|
self.assertIn("Node '/binman/vblock': Collection must have a 'content' "
|
2018-07-17 19:25:47 +00:00
|
|
|
'property', str(e.exception))
|
|
|
|
|
|
|
|
def testVblockBadPhandle(self):
|
|
|
|
"""Test that we detect a vblock with an invalid phandle in contents"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('076_vblock_bad_phandle.dts')
|
2018-07-17 19:25:47 +00:00
|
|
|
self.assertIn("Node '/binman/vblock': Cannot find node for phandle "
|
|
|
|
'1000', str(e.exception))
|
|
|
|
|
|
|
|
def testVblockBadEntry(self):
|
|
|
|
"""Test that we detect an entry that points to a non-entry"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFile('077_vblock_bad_entry.dts')
|
2018-07-17 19:25:47 +00:00
|
|
|
self.assertIn("Node '/binman/vblock': Cannot find entry for node "
|
|
|
|
"'other'", str(e.exception))
|
|
|
|
|
2021-01-07 04:35:17 +00:00
|
|
|
def testVblockContent(self):
|
|
|
|
"""Test that the vblock signs the right data"""
|
|
|
|
self._hash_data = True
|
|
|
|
command.test_result = self._HandleVblockCommand
|
|
|
|
entry_args = {
|
|
|
|
'keydir': 'devkeys',
|
|
|
|
}
|
|
|
|
data = self._DoReadFileDtb(
|
|
|
|
'189_vblock_content.dts', use_real_dtb=True, update_dtb=True,
|
|
|
|
entry_args=entry_args)[0]
|
|
|
|
hashlen = 32 # SHA256 hash is 32 bytes
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[:len(U_BOOT_DATA)])
|
|
|
|
hashval = data[-hashlen:]
|
|
|
|
dtb = data[len(U_BOOT_DATA):-hashlen]
|
|
|
|
|
|
|
|
expected_data = U_BOOT_DATA + dtb
|
|
|
|
|
|
|
|
# The hashval should be a hash of the dtb
|
|
|
|
m = hashlib.sha256()
|
|
|
|
m.update(expected_data)
|
|
|
|
expected_hashval = m.digest()
|
|
|
|
self.assertEqual(expected_hashval, hashval)
|
|
|
|
|
2022-01-10 03:14:09 +00:00
|
|
|
def testVblockMissing(self):
|
|
|
|
"""Test that binman still produces an image if futility is missing"""
|
|
|
|
entry_args = {
|
|
|
|
'keydir': 'devkeys',
|
|
|
|
}
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoTestFile('074_vblock.dts',
|
|
|
|
force_missing_bintools='futility',
|
|
|
|
entry_args=entry_args)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing bintools.*: futility")
|
|
|
|
|
2018-07-17 19:25:48 +00:00
|
|
|
def testTpl(self):
|
2019-08-24 13:23:00 +00:00
|
|
|
"""Test that an image with TPL and its device tree can be created"""
|
2018-07-17 19:25:48 +00:00
|
|
|
# ELF file with a '__bss_size' symbol
|
2019-08-24 13:23:00 +00:00
|
|
|
self._SetupTplElf()
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('078_u_boot_tpl.dts')
|
2018-07-17 19:25:48 +00:00
|
|
|
self.assertEqual(U_BOOT_TPL_DATA + U_BOOT_TPL_DTB_DATA, data)
|
|
|
|
|
2018-07-17 19:25:51 +00:00
|
|
|
def testUsesPos(self):
|
|
|
|
"""Test that the 'pos' property cannot be used anymore"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('079_uses_pos.dts')
|
2018-07-17 19:25:51 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Please use 'offset' instead of "
|
|
|
|
"'pos'", str(e.exception))
|
|
|
|
|
2018-09-14 10:57:08 +00:00
|
|
|
def testFillZero(self):
|
|
|
|
"""Test for an fill entry type with a size of 0"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('080_fill_empty.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0, 16), data)
|
2018-09-14 10:57:08 +00:00
|
|
|
|
2018-09-14 10:57:09 +00:00
|
|
|
def testTextMissing(self):
|
|
|
|
"""Test for a text entry type where there is no text"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('066_text.dts',)
|
2018-09-14 10:57:09 +00:00
|
|
|
self.assertIn("Node '/binman/text': No value provided for text label "
|
|
|
|
"'test-id'", str(e.exception))
|
|
|
|
|
2018-09-14 10:57:10 +00:00
|
|
|
def testPackStart16Tpl(self):
|
|
|
|
"""Test that an image with an x86 start16 TPL region can be created"""
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('081_x86_start16_tpl.dts')
|
2018-09-14 10:57:10 +00:00
|
|
|
self.assertEqual(X86_START16_TPL_DATA, data[:len(X86_START16_TPL_DATA)])
|
|
|
|
|
2018-09-14 10:57:12 +00:00
|
|
|
def testSelectImage(self):
|
|
|
|
"""Test that we can select which images to build"""
|
2019-04-26 03:58:34 +00:00
|
|
|
expected = 'Skipping images: image1'
|
|
|
|
|
|
|
|
# We should only get the expected message in verbose mode
|
2019-07-08 19:18:48 +00:00
|
|
|
for verbosity in (0, 2):
|
2019-04-26 03:58:34 +00:00
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
retcode = self._DoTestFile('006_dual_image.dts',
|
|
|
|
verbosity=verbosity,
|
|
|
|
images=['image2'])
|
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
if verbosity:
|
|
|
|
self.assertIn(expected, stdout.getvalue())
|
|
|
|
else:
|
|
|
|
self.assertNotIn(expected, stdout.getvalue())
|
2018-09-14 10:57:12 +00:00
|
|
|
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertFalse(os.path.exists(tools.get_output_filename('image1.bin')))
|
|
|
|
self.assertTrue(os.path.exists(tools.get_output_filename('image2.bin')))
|
2019-07-20 18:24:10 +00:00
|
|
|
self._CleanupOutputDir()
|
2018-09-14 10:57:12 +00:00
|
|
|
|
2018-09-14 10:57:24 +00:00
|
|
|
def testUpdateFdtAll(self):
|
|
|
|
"""Test that all device trees are updated with offset/size info"""
|
2019-07-08 20:25:26 +00:00
|
|
|
data = self._DoReadFileRealDtb('082_fdt_update_all.dts')
|
2018-09-14 10:57:24 +00:00
|
|
|
|
|
|
|
base_expected = {
|
|
|
|
'offset': 0,
|
2022-02-28 14:16:54 +00:00
|
|
|
'image-pos': 0,
|
|
|
|
'size': 2320,
|
2018-09-14 10:57:24 +00:00
|
|
|
'section:offset': 0,
|
2022-02-28 14:16:54 +00:00
|
|
|
'section:image-pos': 0,
|
|
|
|
'section:size': 565,
|
|
|
|
'section/u-boot-dtb:offset': 0,
|
|
|
|
'section/u-boot-dtb:image-pos': 0,
|
|
|
|
'section/u-boot-dtb:size': 565,
|
|
|
|
'u-boot-spl-dtb:offset': 565,
|
|
|
|
'u-boot-spl-dtb:image-pos': 565,
|
|
|
|
'u-boot-spl-dtb:size': 585,
|
|
|
|
'u-boot-tpl-dtb:offset': 1150,
|
|
|
|
'u-boot-tpl-dtb:image-pos': 1150,
|
|
|
|
'u-boot-tpl-dtb:size': 585,
|
|
|
|
'u-boot-vpl-dtb:image-pos': 1735,
|
|
|
|
'u-boot-vpl-dtb:offset': 1735,
|
|
|
|
'u-boot-vpl-dtb:size': 585,
|
2018-09-14 10:57:24 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
# We expect three device-tree files in the output, one after the other.
|
|
|
|
# Read them in sequence. We look for an 'spl' property in the SPL tree,
|
|
|
|
# and 'tpl' in the TPL tree, to make sure they are distinct from the
|
|
|
|
# main U-Boot tree. All three should have the same postions and offset.
|
|
|
|
start = 0
|
2022-02-28 14:16:54 +00:00
|
|
|
self.maxDiff = None
|
|
|
|
for item in ['', 'spl', 'tpl', 'vpl']:
|
2018-09-14 10:57:24 +00:00
|
|
|
dtb = fdt.Fdt.FromData(data[start:])
|
|
|
|
dtb.Scan()
|
2019-07-20 18:23:51 +00:00
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS + REPACK_DTB_PROPS +
|
2022-02-28 14:16:54 +00:00
|
|
|
['spl', 'tpl', 'vpl'])
|
2018-09-14 10:57:24 +00:00
|
|
|
expected = dict(base_expected)
|
|
|
|
if item:
|
|
|
|
expected[item] = 0
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
start += dtb._fdt_obj.totalsize()
|
|
|
|
|
|
|
|
def testUpdateFdtOutput(self):
|
|
|
|
"""Test that output DTB files are updated"""
|
|
|
|
try:
|
2018-10-01 18:22:30 +00:00
|
|
|
data, dtb_data, _, _ = self._DoReadFileDtb('082_fdt_update_all.dts',
|
2018-09-14 10:57:24 +00:00
|
|
|
use_real_dtb=True, update_dtb=True, reset_dtbs=False)
|
|
|
|
|
|
|
|
# Unfortunately, compiling a source file always results in a file
|
|
|
|
# called source.dtb (see fdt_util.EnsureCompiled()). The test
|
2018-10-01 18:22:30 +00:00
|
|
|
# source file (e.g. test/075_fdt_update_all.dts) thus does not enter
|
2018-09-14 10:57:24 +00:00
|
|
|
# binman as a file called u-boot.dtb. To fix this, copy the file
|
|
|
|
# over to the expected place.
|
|
|
|
start = 0
|
|
|
|
for fname in ['u-boot.dtb.out', 'spl/u-boot-spl.dtb.out',
|
2022-02-28 14:16:54 +00:00
|
|
|
'tpl/u-boot-tpl.dtb.out', 'vpl/u-boot-vpl.dtb.out']:
|
2018-09-14 10:57:24 +00:00
|
|
|
dtb = fdt.Fdt.FromData(data[start:])
|
|
|
|
size = dtb._fdt_obj.totalsize()
|
2022-01-29 21:14:04 +00:00
|
|
|
pathname = tools.get_output_filename(os.path.split(fname)[1])
|
|
|
|
outdata = tools.read_file(pathname)
|
2018-09-14 10:57:24 +00:00
|
|
|
name = os.path.split(fname)[0]
|
|
|
|
|
|
|
|
if name:
|
2022-02-28 14:16:54 +00:00
|
|
|
orig_indata = self._GetDtbContentsForSpls(dtb_data, name)
|
2018-09-14 10:57:24 +00:00
|
|
|
else:
|
|
|
|
orig_indata = dtb_data
|
|
|
|
self.assertNotEqual(outdata, orig_indata,
|
|
|
|
"Expected output file '%s' be updated" % pathname)
|
|
|
|
self.assertEqual(outdata, data[start:start + size],
|
|
|
|
"Expected output file '%s' to match output image" %
|
|
|
|
pathname)
|
|
|
|
start += size
|
|
|
|
finally:
|
|
|
|
self._ResetDtbs()
|
|
|
|
|
2018-09-14 10:57:26 +00:00
|
|
|
def _decompress(self, data):
|
2022-08-19 14:25:28 +00:00
|
|
|
bintool = self.comp_bintools['lz4']
|
|
|
|
return bintool.decompress(data)
|
2018-09-14 10:57:26 +00:00
|
|
|
|
|
|
|
def testCompress(self):
|
|
|
|
"""Test compression of blobs"""
|
2019-07-08 19:18:53 +00:00
|
|
|
self._CheckLz4()
|
2018-10-01 18:22:30 +00:00
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb('083_compress.dts',
|
2018-09-14 10:57:26 +00:00
|
|
|
use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['size', 'uncomp-size'])
|
|
|
|
orig = self._decompress(data)
|
|
|
|
self.assertEquals(COMPRESS_DATA, orig)
|
2020-10-26 23:40:15 +00:00
|
|
|
|
|
|
|
# Do a sanity check on various fields
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(1, len(entries))
|
|
|
|
|
|
|
|
entry = entries['blob']
|
|
|
|
self.assertEqual(COMPRESS_DATA, entry.uncomp_data)
|
|
|
|
self.assertEqual(len(COMPRESS_DATA), entry.uncomp_size)
|
|
|
|
orig = self._decompress(entry.data)
|
|
|
|
self.assertEqual(orig, entry.uncomp_data)
|
|
|
|
|
2020-10-26 23:40:16 +00:00
|
|
|
self.assertEqual(image.data, entry.data)
|
|
|
|
|
2018-09-14 10:57:26 +00:00
|
|
|
expected = {
|
|
|
|
'blob:uncomp-size': len(COMPRESS_DATA),
|
|
|
|
'blob:size': len(data),
|
|
|
|
'size': len(data),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
2018-09-14 10:57:28 +00:00
|
|
|
def testFiles(self):
|
|
|
|
"""Test bringing in multiple files"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('084_files.dts')
|
2018-09-14 10:57:28 +00:00
|
|
|
self.assertEqual(FILES_DATA, data)
|
|
|
|
|
|
|
|
def testFilesCompress(self):
|
|
|
|
"""Test bringing in multiple files and compressing them"""
|
2019-07-08 19:18:53 +00:00
|
|
|
self._CheckLz4()
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('085_files_compress.dts')
|
2018-09-14 10:57:28 +00:00
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
files = entries['files']
|
2019-07-08 20:25:47 +00:00
|
|
|
entries = files._entries
|
2018-09-14 10:57:28 +00:00
|
|
|
|
2019-05-18 04:00:46 +00:00
|
|
|
orig = b''
|
2018-09-14 10:57:28 +00:00
|
|
|
for i in range(1, 3):
|
|
|
|
key = '%d.dat' % i
|
|
|
|
start = entries[key].image_pos
|
|
|
|
len = entries[key].size
|
|
|
|
chunk = data[start:start + len]
|
|
|
|
orig += self._decompress(chunk)
|
|
|
|
|
|
|
|
self.assertEqual(FILES_DATA, orig)
|
|
|
|
|
|
|
|
def testFilesMissing(self):
|
|
|
|
"""Test missing files"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('086_files_none.dts')
|
2018-09-14 10:57:28 +00:00
|
|
|
self.assertIn("Node '/binman/files': Pattern \'files/*.none\' matched "
|
|
|
|
'no files', str(e.exception))
|
|
|
|
|
|
|
|
def testFilesNoPattern(self):
|
|
|
|
"""Test missing files"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('087_files_no_pattern.dts')
|
2018-09-14 10:57:28 +00:00
|
|
|
self.assertIn("Node '/binman/files': Missing 'pattern' property",
|
|
|
|
str(e.exception))
|
|
|
|
|
2022-03-06 03:18:59 +00:00
|
|
|
def testExtendSize(self):
|
|
|
|
"""Test an extending entry"""
|
|
|
|
data, _, map_data, _ = self._DoReadFileDtb('088_extend_size.dts',
|
2018-09-14 10:57:29 +00:00
|
|
|
map=True)
|
2022-01-29 21:14:04 +00:00
|
|
|
expect = (tools.get_bytes(ord('a'), 8) + U_BOOT_DATA +
|
|
|
|
MRC_DATA + tools.get_bytes(ord('b'), 1) + U_BOOT_DATA +
|
|
|
|
tools.get_bytes(ord('c'), 8) + U_BOOT_DATA +
|
|
|
|
tools.get_bytes(ord('d'), 8))
|
2018-09-14 10:57:29 +00:00
|
|
|
self.assertEqual(expect, data)
|
|
|
|
self.assertEqual('''ImagePos Offset Size Name
|
|
|
|
00000000 00000000 00000028 main-section
|
|
|
|
00000000 00000000 00000008 fill
|
|
|
|
00000008 00000008 00000004 u-boot
|
|
|
|
0000000c 0000000c 00000004 section
|
|
|
|
0000000c 00000000 00000003 intel-mrc
|
|
|
|
00000010 00000010 00000004 u-boot2
|
|
|
|
00000014 00000014 0000000c section2
|
|
|
|
00000014 00000000 00000008 fill
|
|
|
|
0000001c 00000008 00000004 u-boot
|
|
|
|
00000020 00000020 00000008 fill2
|
|
|
|
''', map_data)
|
|
|
|
|
2022-03-06 03:18:59 +00:00
|
|
|
def testExtendSizeBad(self):
|
|
|
|
"""Test an extending entry which fails to provide contents"""
|
2018-09-14 10:57:36 +00:00
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-03-06 03:18:59 +00:00
|
|
|
self._DoReadFileDtb('089_extend_size_bad.dts', map=True)
|
2018-09-14 10:57:29 +00:00
|
|
|
self.assertIn("Node '/binman/_testing': Cannot obtain contents when "
|
|
|
|
'expanding entry', str(e.exception))
|
|
|
|
|
2018-09-14 10:57:31 +00:00
|
|
|
def testHash(self):
|
|
|
|
"""Test hashing of the contents of an entry"""
|
2018-10-01 18:22:30 +00:00
|
|
|
_, _, _, out_dtb_fname = self._DoReadFileDtb('090_hash.dts',
|
2018-09-14 10:57:31 +00:00
|
|
|
use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
hash_node = dtb.GetNode('/binman/u-boot/hash').props['value']
|
|
|
|
m = hashlib.sha256()
|
|
|
|
m.update(U_BOOT_DATA)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(m.digest(), b''.join(hash_node.value))
|
2018-09-14 10:57:31 +00:00
|
|
|
|
|
|
|
def testHashNoAlgo(self):
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('091_hash_no_algo.dts', update_dtb=True)
|
2018-09-14 10:57:31 +00:00
|
|
|
self.assertIn("Node \'/binman/u-boot\': Missing \'algo\' property for "
|
|
|
|
'hash node', str(e.exception))
|
|
|
|
|
|
|
|
def testHashBadAlgo(self):
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoReadFileDtb('092_hash_bad_algo.dts', update_dtb=True)
|
2022-02-08 17:59:44 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot': Unknown hash algorithm 'invalid'",
|
2018-09-14 10:57:31 +00:00
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testHashSection(self):
|
|
|
|
"""Test hashing of the contents of an entry"""
|
2018-10-01 18:22:30 +00:00
|
|
|
_, _, _, out_dtb_fname = self._DoReadFileDtb('099_hash_section.dts',
|
2018-09-14 10:57:31 +00:00
|
|
|
use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
hash_node = dtb.GetNode('/binman/section/hash').props['value']
|
|
|
|
m = hashlib.sha256()
|
|
|
|
m.update(U_BOOT_DATA)
|
2022-01-29 21:14:04 +00:00
|
|
|
m.update(tools.get_bytes(ord('a'), 16))
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(m.digest(), b''.join(hash_node.value))
|
2018-09-14 10:57:31 +00:00
|
|
|
|
2018-09-14 10:57:32 +00:00
|
|
|
def testPackUBootTplMicrocode(self):
|
|
|
|
"""Test that x86 microcode can be handled correctly in TPL
|
|
|
|
|
|
|
|
We expect to see the following in the image, in order:
|
|
|
|
u-boot-tpl-nodtb.bin with a microcode pointer inserted at the correct
|
|
|
|
place
|
|
|
|
u-boot-tpl.dtb with the microcode removed
|
|
|
|
the microcode
|
|
|
|
"""
|
2019-08-24 13:23:00 +00:00
|
|
|
self._SetupTplElf('u_boot_ucode_ptr')
|
2018-10-01 18:22:30 +00:00
|
|
|
first, pos_and_size = self._RunMicrocodeTest('093_x86_tpl_ucode.dts',
|
2018-09-14 10:57:32 +00:00
|
|
|
U_BOOT_TPL_NODTB_DATA)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'tplnodtb with microc' + pos_and_size +
|
|
|
|
b'ter somewhere in here', first)
|
2018-09-14 10:57:32 +00:00
|
|
|
|
2018-09-14 10:57:34 +00:00
|
|
|
def testFmapX86(self):
|
|
|
|
"""Basic test of generation of a flashrom fmap"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('094_fmap_x86.dts')
|
2018-09-14 10:57:34 +00:00
|
|
|
fhdr, fentries = fmap_util.DecodeFmap(data[32:])
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = U_BOOT_DATA + MRC_DATA + tools.get_bytes(ord('a'), 32 - 7)
|
2018-09-14 10:57:34 +00:00
|
|
|
self.assertEqual(expected, data[:32])
|
|
|
|
fhdr, fentries = fmap_util.DecodeFmap(data[32:])
|
|
|
|
|
|
|
|
self.assertEqual(0x100, fhdr.image_size)
|
|
|
|
|
|
|
|
self.assertEqual(0, fentries[0].offset)
|
|
|
|
self.assertEqual(4, fentries[0].size)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'U_BOOT', fentries[0].name)
|
2018-09-14 10:57:34 +00:00
|
|
|
|
|
|
|
self.assertEqual(4, fentries[1].offset)
|
|
|
|
self.assertEqual(3, fentries[1].size)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'INTEL_MRC', fentries[1].name)
|
2018-09-14 10:57:34 +00:00
|
|
|
|
|
|
|
self.assertEqual(32, fentries[2].offset)
|
|
|
|
self.assertEqual(fmap_util.FMAP_HEADER_LEN +
|
|
|
|
fmap_util.FMAP_AREA_LEN * 3, fentries[2].size)
|
2019-05-18 04:00:46 +00:00
|
|
|
self.assertEqual(b'FMAP', fentries[2].name)
|
2018-09-14 10:57:34 +00:00
|
|
|
|
|
|
|
def testFmapX86Section(self):
|
|
|
|
"""Basic test of generation of a flashrom fmap"""
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('095_fmap_x86_section.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = U_BOOT_DATA + MRC_DATA + tools.get_bytes(ord('b'), 32 - 7)
|
2018-09-14 10:57:34 +00:00
|
|
|
self.assertEqual(expected, data[:32])
|
|
|
|
fhdr, fentries = fmap_util.DecodeFmap(data[36:])
|
|
|
|
|
2021-04-02 22:05:10 +00:00
|
|
|
self.assertEqual(0x180, fhdr.image_size)
|
|
|
|
expect_size = fmap_util.FMAP_HEADER_LEN + fmap_util.FMAP_AREA_LEN * 4
|
2021-04-02 22:05:09 +00:00
|
|
|
fiter = iter(fentries)
|
|
|
|
|
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'U_BOOT', fentry.name)
|
|
|
|
self.assertEqual(0, fentry.offset)
|
|
|
|
self.assertEqual(4, fentry.size)
|
|
|
|
|
2021-04-02 22:05:10 +00:00
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'SECTION', fentry.name)
|
|
|
|
self.assertEqual(4, fentry.offset)
|
|
|
|
self.assertEqual(0x20 + expect_size, fentry.size)
|
|
|
|
|
2021-04-02 22:05:09 +00:00
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'INTEL_MRC', fentry.name)
|
|
|
|
self.assertEqual(4, fentry.offset)
|
|
|
|
self.assertEqual(3, fentry.size)
|
|
|
|
|
|
|
|
fentry = next(fiter)
|
|
|
|
self.assertEqual(b'FMAP', fentry.name)
|
|
|
|
self.assertEqual(36, fentry.offset)
|
|
|
|
self.assertEqual(expect_size, fentry.size)
|
2018-09-14 10:57:34 +00:00
|
|
|
|
2018-09-14 10:57:35 +00:00
|
|
|
def testElf(self):
|
|
|
|
"""Basic test of ELF entries"""
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf()
|
2019-08-24 13:23:00 +00:00
|
|
|
self._SetupTplElf()
|
2019-08-24 13:22:53 +00:00
|
|
|
with open(self.ElfTestFile('bss_data'), 'rb') as fd:
|
2018-09-14 10:57:35 +00:00
|
|
|
TestFunctional._MakeInputFile('-boot', fd.read())
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('096_elf.dts')
|
2018-09-14 10:57:35 +00:00
|
|
|
|
2019-07-08 19:18:25 +00:00
|
|
|
def testElfStrip(self):
|
2018-09-14 10:57:35 +00:00
|
|
|
"""Basic test of ELF entries"""
|
2018-10-02 03:12:47 +00:00
|
|
|
self._SetupSplElf()
|
2019-08-24 13:22:53 +00:00
|
|
|
with open(self.ElfTestFile('bss_data'), 'rb') as fd:
|
2018-09-14 10:57:35 +00:00
|
|
|
TestFunctional._MakeInputFile('-boot', fd.read())
|
2018-10-01 18:22:30 +00:00
|
|
|
data = self._DoReadFile('097_elf_strip.dts')
|
2018-09-14 10:57:35 +00:00
|
|
|
|
2018-09-14 10:57:36 +00:00
|
|
|
def testPackOverlapMap(self):
|
|
|
|
"""Test that overlapping regions are detected"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2018-10-01 18:22:30 +00:00
|
|
|
self._DoTestFile('014_pack_overlap.dts', map=True)
|
2022-01-29 21:14:04 +00:00
|
|
|
map_fname = tools.get_output_filename('image.map')
|
2018-09-14 10:57:36 +00:00
|
|
|
self.assertEqual("Wrote map file '%s' to show errors\n" % map_fname,
|
|
|
|
stdout.getvalue())
|
|
|
|
|
|
|
|
# We should not get an inmage, but there should be a map file
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertFalse(os.path.exists(tools.get_output_filename('image.bin')))
|
2018-09-14 10:57:36 +00:00
|
|
|
self.assertTrue(os.path.exists(map_fname))
|
2022-01-29 21:14:04 +00:00
|
|
|
map_data = tools.read_file(map_fname, binary=False)
|
2018-09-14 10:57:36 +00:00
|
|
|
self.assertEqual('''ImagePos Offset Size Name
|
2020-10-26 23:40:24 +00:00
|
|
|
<none> 00000000 00000008 main-section
|
2018-09-14 10:57:36 +00:00
|
|
|
<none> 00000000 00000004 u-boot
|
|
|
|
<none> 00000003 00000004 u-boot-align
|
|
|
|
''', map_data)
|
|
|
|
|
2019-07-08 19:18:25 +00:00
|
|
|
def testPackRefCode(self):
|
2018-10-01 18:22:31 +00:00
|
|
|
"""Test that an image with an Intel Reference code binary works"""
|
|
|
|
data = self._DoReadFile('100_intel_refcode.dts')
|
|
|
|
self.assertEqual(REFCODE_DATA, data[:len(REFCODE_DATA)])
|
|
|
|
|
2019-04-26 03:58:39 +00:00
|
|
|
def testSectionOffset(self):
|
|
|
|
"""Tests use of a section with an offset"""
|
|
|
|
data, _, map_data, _ = self._DoReadFileDtb('101_sections_offset.dts',
|
|
|
|
map=True)
|
|
|
|
self.assertEqual('''ImagePos Offset Size Name
|
|
|
|
00000000 00000000 00000038 main-section
|
|
|
|
00000004 00000004 00000010 section@0
|
|
|
|
00000004 00000000 00000004 u-boot
|
|
|
|
00000018 00000018 00000010 section@1
|
|
|
|
00000018 00000000 00000004 u-boot
|
|
|
|
0000002c 0000002c 00000004 section@2
|
|
|
|
0000002c 00000000 00000004 u-boot
|
|
|
|
''', map_data)
|
|
|
|
self.assertEqual(data,
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.get_bytes(0x26, 4) + U_BOOT_DATA +
|
|
|
|
tools.get_bytes(0x21, 12) +
|
|
|
|
tools.get_bytes(0x26, 4) + U_BOOT_DATA +
|
|
|
|
tools.get_bytes(0x61, 12) +
|
|
|
|
tools.get_bytes(0x26, 4) + U_BOOT_DATA +
|
|
|
|
tools.get_bytes(0x26, 8))
|
2019-04-26 03:58:39 +00:00
|
|
|
|
2019-07-08 19:18:53 +00:00
|
|
|
def testCbfsRaw(self):
|
|
|
|
"""Test base handling of a Coreboot Filesystem (CBFS)
|
|
|
|
|
|
|
|
The exact contents of the CBFS is verified by similar tests in
|
|
|
|
cbfs_util_test.py. The tests here merely check that the files added to
|
|
|
|
the CBFS can be found in the final image.
|
|
|
|
"""
|
|
|
|
data = self._DoReadFile('102_cbfs_raw.dts')
|
|
|
|
size = 0xb0
|
|
|
|
|
|
|
|
cbfs = cbfs_util.CbfsReader(data)
|
|
|
|
self.assertEqual(size, cbfs.rom_size)
|
|
|
|
|
|
|
|
self.assertIn('u-boot-dtb', cbfs.files)
|
|
|
|
cfile = cbfs.files['u-boot-dtb']
|
|
|
|
self.assertEqual(U_BOOT_DTB_DATA, cfile.data)
|
|
|
|
|
|
|
|
def testCbfsArch(self):
|
|
|
|
"""Test on non-x86 architecture"""
|
|
|
|
data = self._DoReadFile('103_cbfs_raw_ppc.dts')
|
|
|
|
size = 0x100
|
|
|
|
|
|
|
|
cbfs = cbfs_util.CbfsReader(data)
|
|
|
|
self.assertEqual(size, cbfs.rom_size)
|
|
|
|
|
|
|
|
self.assertIn('u-boot-dtb', cbfs.files)
|
|
|
|
cfile = cbfs.files['u-boot-dtb']
|
|
|
|
self.assertEqual(U_BOOT_DTB_DATA, cfile.data)
|
|
|
|
|
|
|
|
def testCbfsStage(self):
|
|
|
|
"""Tests handling of a Coreboot Filesystem (CBFS)"""
|
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
|
|
|
elf_fname = os.path.join(self._indir, 'cbfs-stage.elf')
|
|
|
|
elf.MakeElf(elf_fname, U_BOOT_DATA, U_BOOT_DTB_DATA)
|
|
|
|
size = 0xb0
|
|
|
|
|
|
|
|
data = self._DoReadFile('104_cbfs_stage.dts')
|
|
|
|
cbfs = cbfs_util.CbfsReader(data)
|
|
|
|
self.assertEqual(size, cbfs.rom_size)
|
|
|
|
|
|
|
|
self.assertIn('u-boot', cbfs.files)
|
|
|
|
cfile = cbfs.files['u-boot']
|
|
|
|
self.assertEqual(U_BOOT_DATA + U_BOOT_DTB_DATA, cfile.data)
|
|
|
|
|
|
|
|
def testCbfsRawCompress(self):
|
|
|
|
"""Test handling of compressing raw files"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data = self._DoReadFile('105_cbfs_raw_compress.dts')
|
|
|
|
size = 0x140
|
|
|
|
|
|
|
|
cbfs = cbfs_util.CbfsReader(data)
|
|
|
|
self.assertIn('u-boot', cbfs.files)
|
|
|
|
cfile = cbfs.files['u-boot']
|
|
|
|
self.assertEqual(COMPRESS_DATA, cfile.data)
|
|
|
|
|
|
|
|
def testCbfsBadArch(self):
|
|
|
|
"""Test handling of a bad architecture"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('106_cbfs_bad_arch.dts')
|
|
|
|
self.assertIn("Invalid architecture 'bad-arch'", str(e.exception))
|
|
|
|
|
|
|
|
def testCbfsNoSize(self):
|
|
|
|
"""Test handling of a missing size property"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('107_cbfs_no_size.dts')
|
|
|
|
self.assertIn('entry must have a size property', str(e.exception))
|
|
|
|
|
2021-11-23 18:03:54 +00:00
|
|
|
def testCbfsNoContents(self):
|
2019-07-08 19:18:53 +00:00
|
|
|
"""Test handling of a CBFS entry which does not provide contentsy"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('108_cbfs_no_contents.dts')
|
|
|
|
self.assertIn('Could not complete processing of contents',
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testCbfsBadCompress(self):
|
|
|
|
"""Test handling of a bad architecture"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('109_cbfs_bad_compress.dts')
|
|
|
|
self.assertIn("Invalid compression in 'u-boot': 'invalid-algo'",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testCbfsNamedEntries(self):
|
|
|
|
"""Test handling of named entries"""
|
|
|
|
data = self._DoReadFile('110_cbfs_name.dts')
|
|
|
|
|
|
|
|
cbfs = cbfs_util.CbfsReader(data)
|
|
|
|
self.assertIn('FRED', cbfs.files)
|
|
|
|
cfile1 = cbfs.files['FRED']
|
|
|
|
self.assertEqual(U_BOOT_DATA, cfile1.data)
|
|
|
|
|
|
|
|
self.assertIn('hello', cbfs.files)
|
|
|
|
cfile2 = cbfs.files['hello']
|
|
|
|
self.assertEqual(U_BOOT_DTB_DATA, cfile2.data)
|
|
|
|
|
2019-07-08 19:18:54 +00:00
|
|
|
def _SetupIfwi(self, fname):
|
|
|
|
"""Set up to run an IFWI test
|
|
|
|
|
|
|
|
Args:
|
|
|
|
fname: Filename of input file to provide (fitimage.bin or ifwi.bin)
|
|
|
|
"""
|
|
|
|
self._SetupSplElf()
|
2019-08-24 13:23:00 +00:00
|
|
|
self._SetupTplElf()
|
2019-07-08 19:18:54 +00:00
|
|
|
|
|
|
|
# Intel Integrated Firmware Image (IFWI) file
|
|
|
|
with gzip.open(self.TestFile('%s.gz' % fname), 'rb') as fd:
|
|
|
|
data = fd.read()
|
|
|
|
TestFunctional._MakeInputFile(fname,data)
|
|
|
|
|
|
|
|
def _CheckIfwi(self, data):
|
|
|
|
"""Check that an image with an IFWI contains the correct output
|
|
|
|
|
|
|
|
Args:
|
|
|
|
data: Conents of output file
|
|
|
|
"""
|
2022-01-29 21:14:04 +00:00
|
|
|
expected_desc = tools.read_file(self.TestFile('descriptor.bin'))
|
2019-07-08 19:18:54 +00:00
|
|
|
if data[:0x1000] != expected_desc:
|
|
|
|
self.fail('Expected descriptor binary at start of image')
|
|
|
|
|
|
|
|
# We expect to find the TPL wil in subpart IBBP entry IBBL
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
tpl_fname = tools.get_output_filename('tpl.out')
|
2022-01-10 03:14:01 +00:00
|
|
|
ifwitool = bintool.Bintool.create('ifwitool')
|
|
|
|
ifwitool.extract(image_fname, 'IBBP', 'IBBL', tpl_fname)
|
2019-07-08 19:18:54 +00:00
|
|
|
|
2022-01-29 21:14:04 +00:00
|
|
|
tpl_data = tools.read_file(tpl_fname)
|
2019-08-24 13:22:51 +00:00
|
|
|
self.assertEqual(U_BOOT_TPL_DATA, tpl_data[:len(U_BOOT_TPL_DATA)])
|
2019-07-08 19:18:54 +00:00
|
|
|
|
|
|
|
def testPackX86RomIfwi(self):
|
|
|
|
"""Test that an x86 ROM with Integrated Firmware Image can be created"""
|
|
|
|
self._SetupIfwi('fitimage.bin')
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('111_x86_rom_ifwi.dts')
|
2019-07-08 19:18:54 +00:00
|
|
|
self._CheckIfwi(data)
|
|
|
|
|
|
|
|
def testPackX86RomIfwiNoDesc(self):
|
|
|
|
"""Test that an x86 ROM with IFWI can be created from an ifwi.bin file"""
|
|
|
|
self._SetupIfwi('ifwi.bin')
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('112_x86_rom_ifwi_nodesc.dts')
|
2019-07-08 19:18:54 +00:00
|
|
|
self._CheckIfwi(data)
|
|
|
|
|
|
|
|
def testPackX86RomIfwiNoData(self):
|
|
|
|
"""Test that an x86 ROM with IFWI handles missing data"""
|
|
|
|
self._SetupIfwi('ifwi.bin')
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2019-08-24 13:23:01 +00:00
|
|
|
data = self._DoReadFile('113_x86_rom_ifwi_nodata.dts')
|
2019-07-08 19:18:54 +00:00
|
|
|
self.assertIn('Could not complete processing of contents',
|
|
|
|
str(e.exception))
|
2018-07-17 19:25:32 +00:00
|
|
|
|
2022-01-10 03:14:09 +00:00
|
|
|
def testIfwiMissing(self):
|
|
|
|
"""Test that binman still produces an image if ifwitool is missing"""
|
|
|
|
self._SetupIfwi('fitimage.bin')
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoTestFile('111_x86_rom_ifwi.dts',
|
|
|
|
force_missing_bintools='ifwitool')
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing bintools.*: ifwitool")
|
|
|
|
|
2019-07-08 19:18:56 +00:00
|
|
|
def testCbfsOffset(self):
|
|
|
|
"""Test a CBFS with files at particular offsets
|
|
|
|
|
|
|
|
Like all CFBS tests, this is just checking the logic that calls
|
|
|
|
cbfs_util. See cbfs_util_test for fully tests (e.g. test_cbfs_offset()).
|
|
|
|
"""
|
|
|
|
data = self._DoReadFile('114_cbfs_offset.dts')
|
|
|
|
size = 0x200
|
|
|
|
|
|
|
|
cbfs = cbfs_util.CbfsReader(data)
|
|
|
|
self.assertEqual(size, cbfs.rom_size)
|
|
|
|
|
|
|
|
self.assertIn('u-boot', cbfs.files)
|
|
|
|
cfile = cbfs.files['u-boot']
|
|
|
|
self.assertEqual(U_BOOT_DATA, cfile.data)
|
|
|
|
self.assertEqual(0x40, cfile.cbfs_offset)
|
|
|
|
|
|
|
|
self.assertIn('u-boot-dtb', cbfs.files)
|
|
|
|
cfile2 = cbfs.files['u-boot-dtb']
|
|
|
|
self.assertEqual(U_BOOT_DTB_DATA, cfile2.data)
|
|
|
|
self.assertEqual(0x140, cfile2.cbfs_offset)
|
|
|
|
|
2019-07-08 20:25:27 +00:00
|
|
|
def testFdtmap(self):
|
|
|
|
"""Test an FDT map can be inserted in the image"""
|
|
|
|
data = self.data = self._DoReadFileRealDtb('115_fdtmap.dts')
|
|
|
|
fdtmap_data = data[len(U_BOOT_DATA):]
|
|
|
|
magic = fdtmap_data[:8]
|
2019-10-31 13:43:03 +00:00
|
|
|
self.assertEqual(b'_FDTMAP_', magic)
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0, 8), fdtmap_data[8:16])
|
2019-07-08 20:25:27 +00:00
|
|
|
|
|
|
|
fdt_data = fdtmap_data[16:]
|
|
|
|
dtb = fdt.Fdt.FromData(fdt_data)
|
|
|
|
dtb.Scan()
|
2019-07-20 18:23:47 +00:00
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS, prefix='/')
|
2019-07-08 20:25:27 +00:00
|
|
|
self.assertEqual({
|
|
|
|
'image-pos': 0,
|
|
|
|
'offset': 0,
|
|
|
|
'u-boot:offset': 0,
|
|
|
|
'u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'u-boot:image-pos': 0,
|
|
|
|
'fdtmap:image-pos': 4,
|
|
|
|
'fdtmap:offset': 4,
|
|
|
|
'fdtmap:size': len(fdtmap_data),
|
|
|
|
'size': len(data),
|
|
|
|
}, props)
|
|
|
|
|
|
|
|
def testFdtmapNoMatch(self):
|
|
|
|
"""Check handling of an FDT map when the section cannot be found"""
|
|
|
|
self.data = self._DoReadFileRealDtb('115_fdtmap.dts')
|
|
|
|
|
|
|
|
# Mangle the section name, which should cause a mismatch between the
|
|
|
|
# correct FDT path and the one expected by the section
|
|
|
|
image = control.images['image']
|
2019-07-08 20:25:28 +00:00
|
|
|
image._node.path += '-suffix'
|
2019-07-08 20:25:27 +00:00
|
|
|
entries = image.GetEntries()
|
|
|
|
fdtmap = entries['fdtmap']
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
fdtmap._GetFdtmap()
|
|
|
|
self.assertIn("Cannot locate node for path '/binman-suffix'",
|
|
|
|
str(e.exception))
|
|
|
|
|
2019-07-08 20:25:28 +00:00
|
|
|
def testFdtmapHeader(self):
|
|
|
|
"""Test an FDT map and image header can be inserted in the image"""
|
|
|
|
data = self.data = self._DoReadFileRealDtb('116_fdtmap_hdr.dts')
|
|
|
|
fdtmap_pos = len(U_BOOT_DATA)
|
|
|
|
fdtmap_data = data[fdtmap_pos:]
|
|
|
|
fdt_data = fdtmap_data[16:]
|
|
|
|
dtb = fdt.Fdt.FromData(fdt_data)
|
|
|
|
fdt_size = dtb.GetFdtObj().totalsize()
|
|
|
|
hdr_data = data[-8:]
|
2019-10-31 13:43:03 +00:00
|
|
|
self.assertEqual(b'BinM', hdr_data[:4])
|
2019-07-08 20:25:28 +00:00
|
|
|
offset = struct.unpack('<I', hdr_data[4:])[0] & 0xffffffff
|
|
|
|
self.assertEqual(fdtmap_pos - 0x400, offset - (1 << 32))
|
|
|
|
|
|
|
|
def testFdtmapHeaderStart(self):
|
|
|
|
"""Test an image header can be inserted at the image start"""
|
|
|
|
data = self.data = self._DoReadFileRealDtb('117_fdtmap_hdr_start.dts')
|
|
|
|
fdtmap_pos = 0x100 + len(U_BOOT_DATA)
|
|
|
|
hdr_data = data[:8]
|
2019-10-31 13:43:03 +00:00
|
|
|
self.assertEqual(b'BinM', hdr_data[:4])
|
2019-07-08 20:25:28 +00:00
|
|
|
offset = struct.unpack('<I', hdr_data[4:])[0]
|
|
|
|
self.assertEqual(fdtmap_pos, offset)
|
|
|
|
|
|
|
|
def testFdtmapHeaderPos(self):
|
|
|
|
"""Test an image header can be inserted at a chosen position"""
|
|
|
|
data = self.data = self._DoReadFileRealDtb('118_fdtmap_hdr_pos.dts')
|
|
|
|
fdtmap_pos = 0x100 + len(U_BOOT_DATA)
|
|
|
|
hdr_data = data[0x80:0x88]
|
2019-10-31 13:43:03 +00:00
|
|
|
self.assertEqual(b'BinM', hdr_data[:4])
|
2019-07-08 20:25:28 +00:00
|
|
|
offset = struct.unpack('<I', hdr_data[4:])[0]
|
|
|
|
self.assertEqual(fdtmap_pos, offset)
|
|
|
|
|
|
|
|
def testHeaderMissingFdtmap(self):
|
|
|
|
"""Test an image header requires an fdtmap"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self.data = self._DoReadFileRealDtb('119_fdtmap_hdr_missing.dts')
|
|
|
|
self.assertIn("'image_header' section must have an 'fdtmap' sibling",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testHeaderNoLocation(self):
|
|
|
|
"""Test an image header with a no specified location is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self.data = self._DoReadFileRealDtb('120_hdr_no_location.dts')
|
|
|
|
self.assertIn("Invalid location 'None', expected 'start' or 'end'",
|
|
|
|
str(e.exception))
|
|
|
|
|
2019-07-08 20:25:37 +00:00
|
|
|
def testEntryExpand(self):
|
2022-03-06 03:18:59 +00:00
|
|
|
"""Test extending an entry after it is packed"""
|
|
|
|
data = self._DoReadFile('121_entry_extend.dts')
|
2019-07-20 18:23:57 +00:00
|
|
|
self.assertEqual(b'aaa', data[:3])
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[3:3 + len(U_BOOT_DATA)])
|
|
|
|
self.assertEqual(b'aaa', data[-3:])
|
2019-07-08 20:25:37 +00:00
|
|
|
|
2022-03-06 03:18:59 +00:00
|
|
|
def testEntryExtendBad(self):
|
|
|
|
"""Test extending an entry after it is packed, twice"""
|
2019-07-08 20:25:37 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-03-06 03:18:59 +00:00
|
|
|
self._DoReadFile('122_entry_extend_twice.dts')
|
2019-07-20 18:23:58 +00:00
|
|
|
self.assertIn("Image '/binman': Entries changed size after packing",
|
2019-07-08 20:25:37 +00:00
|
|
|
str(e.exception))
|
|
|
|
|
2022-03-06 03:18:59 +00:00
|
|
|
def testEntryExtendSection(self):
|
|
|
|
"""Test extending an entry within a section after it is packed"""
|
|
|
|
data = self._DoReadFile('123_entry_extend_section.dts')
|
2019-07-20 18:23:57 +00:00
|
|
|
self.assertEqual(b'aaa', data[:3])
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[3:3 + len(U_BOOT_DATA)])
|
|
|
|
self.assertEqual(b'aaa', data[-3:])
|
2019-07-08 20:25:37 +00:00
|
|
|
|
2019-07-08 20:25:38 +00:00
|
|
|
def testCompressDtb(self):
|
|
|
|
"""Test that compress of device-tree files is supported"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data = self.data = self._DoReadFileRealDtb('124_compress_dtb.dts')
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[:len(U_BOOT_DATA)])
|
|
|
|
comp_data = data[len(U_BOOT_DATA):]
|
|
|
|
orig = self._decompress(comp_data)
|
|
|
|
dtb = fdt.Fdt.FromData(orig)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['size', 'uncomp-size'])
|
|
|
|
expected = {
|
|
|
|
'u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'u-boot-dtb:uncomp-size': len(orig),
|
|
|
|
'u-boot-dtb:size': len(comp_data),
|
|
|
|
'size': len(data),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
2019-07-08 20:25:41 +00:00
|
|
|
def testCbfsUpdateFdt(self):
|
|
|
|
"""Test that we can update the device tree with CBFS offset/size info"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb('125_cbfs_update.dts',
|
|
|
|
update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
2019-07-20 18:23:47 +00:00
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS + ['uncomp-size'])
|
2019-07-08 20:25:41 +00:00
|
|
|
del props['cbfs/u-boot:size']
|
|
|
|
self.assertEqual({
|
|
|
|
'offset': 0,
|
|
|
|
'size': len(data),
|
|
|
|
'image-pos': 0,
|
|
|
|
'cbfs:offset': 0,
|
|
|
|
'cbfs:size': len(data),
|
|
|
|
'cbfs:image-pos': 0,
|
|
|
|
'cbfs/u-boot:offset': 0x38,
|
|
|
|
'cbfs/u-boot:uncomp-size': len(U_BOOT_DATA),
|
|
|
|
'cbfs/u-boot:image-pos': 0x38,
|
|
|
|
'cbfs/u-boot-dtb:offset': 0xb8,
|
|
|
|
'cbfs/u-boot-dtb:size': len(U_BOOT_DATA),
|
|
|
|
'cbfs/u-boot-dtb:image-pos': 0xb8,
|
|
|
|
}, props)
|
|
|
|
|
2019-07-08 20:25:42 +00:00
|
|
|
def testCbfsBadType(self):
|
|
|
|
"""Test an image header with a no specified location is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('126_cbfs_bad_type.dts')
|
|
|
|
self.assertIn("Unknown cbfs-type 'badtype'", str(e.exception))
|
|
|
|
|
2019-07-08 20:25:43 +00:00
|
|
|
def testList(self):
|
|
|
|
"""Test listing the files in an image"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data = self._DoReadFile('127_list.dts')
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.BuildEntryList()
|
|
|
|
self.assertEqual(7, len(entries))
|
|
|
|
|
|
|
|
ent = entries[0]
|
|
|
|
self.assertEqual(0, ent.indent)
|
|
|
|
self.assertEqual('main-section', ent.name)
|
|
|
|
self.assertEqual('section', ent.etype)
|
|
|
|
self.assertEqual(len(data), ent.size)
|
|
|
|
self.assertEqual(0, ent.image_pos)
|
|
|
|
self.assertEqual(None, ent.uncomp_size)
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(0, ent.offset)
|
2019-07-08 20:25:43 +00:00
|
|
|
|
|
|
|
ent = entries[1]
|
|
|
|
self.assertEqual(1, ent.indent)
|
|
|
|
self.assertEqual('u-boot', ent.name)
|
|
|
|
self.assertEqual('u-boot', ent.etype)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), ent.size)
|
|
|
|
self.assertEqual(0, ent.image_pos)
|
|
|
|
self.assertEqual(None, ent.uncomp_size)
|
|
|
|
self.assertEqual(0, ent.offset)
|
|
|
|
|
|
|
|
ent = entries[2]
|
|
|
|
self.assertEqual(1, ent.indent)
|
|
|
|
self.assertEqual('section', ent.name)
|
|
|
|
self.assertEqual('section', ent.etype)
|
|
|
|
section_size = ent.size
|
|
|
|
self.assertEqual(0x100, ent.image_pos)
|
|
|
|
self.assertEqual(None, ent.uncomp_size)
|
2019-07-08 20:25:47 +00:00
|
|
|
self.assertEqual(0x100, ent.offset)
|
2019-07-08 20:25:43 +00:00
|
|
|
|
|
|
|
ent = entries[3]
|
|
|
|
self.assertEqual(2, ent.indent)
|
|
|
|
self.assertEqual('cbfs', ent.name)
|
|
|
|
self.assertEqual('cbfs', ent.etype)
|
|
|
|
self.assertEqual(0x400, ent.size)
|
|
|
|
self.assertEqual(0x100, ent.image_pos)
|
|
|
|
self.assertEqual(None, ent.uncomp_size)
|
|
|
|
self.assertEqual(0, ent.offset)
|
|
|
|
|
|
|
|
ent = entries[4]
|
|
|
|
self.assertEqual(3, ent.indent)
|
|
|
|
self.assertEqual('u-boot', ent.name)
|
|
|
|
self.assertEqual('u-boot', ent.etype)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), ent.size)
|
|
|
|
self.assertEqual(0x138, ent.image_pos)
|
|
|
|
self.assertEqual(None, ent.uncomp_size)
|
|
|
|
self.assertEqual(0x38, ent.offset)
|
|
|
|
|
|
|
|
ent = entries[5]
|
|
|
|
self.assertEqual(3, ent.indent)
|
|
|
|
self.assertEqual('u-boot-dtb', ent.name)
|
|
|
|
self.assertEqual('text', ent.etype)
|
|
|
|
self.assertGreater(len(COMPRESS_DATA), ent.size)
|
|
|
|
self.assertEqual(0x178, ent.image_pos)
|
|
|
|
self.assertEqual(len(COMPRESS_DATA), ent.uncomp_size)
|
|
|
|
self.assertEqual(0x78, ent.offset)
|
|
|
|
|
|
|
|
ent = entries[6]
|
|
|
|
self.assertEqual(2, ent.indent)
|
|
|
|
self.assertEqual('u-boot-dtb', ent.name)
|
|
|
|
self.assertEqual('u-boot-dtb', ent.etype)
|
|
|
|
self.assertEqual(0x500, ent.image_pos)
|
|
|
|
self.assertEqual(len(U_BOOT_DTB_DATA), ent.uncomp_size)
|
|
|
|
dtb_size = ent.size
|
|
|
|
# Compressing this data expands it since headers are added
|
|
|
|
self.assertGreater(dtb_size, len(U_BOOT_DTB_DATA))
|
|
|
|
self.assertEqual(0x400, ent.offset)
|
|
|
|
|
|
|
|
self.assertEqual(len(data), 0x100 + section_size)
|
|
|
|
self.assertEqual(section_size, 0x400 + dtb_size)
|
|
|
|
|
2019-07-08 20:25:44 +00:00
|
|
|
def testFindFdtmap(self):
|
|
|
|
"""Test locating an FDT map in an image"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data = self.data = self._DoReadFileRealDtb('128_decode_image.dts')
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
entry = entries['fdtmap']
|
|
|
|
self.assertEqual(entry.image_pos, fdtmap.LocateFdtmap(data))
|
|
|
|
|
|
|
|
def testFindFdtmapMissing(self):
|
|
|
|
"""Test failing to locate an FDP map"""
|
|
|
|
data = self._DoReadFile('005_simple.dts')
|
|
|
|
self.assertEqual(None, fdtmap.LocateFdtmap(data))
|
|
|
|
|
2019-07-08 20:25:45 +00:00
|
|
|
def testFindImageHeader(self):
|
|
|
|
"""Test locating a image header"""
|
|
|
|
self._CheckLz4()
|
2019-07-08 20:25:46 +00:00
|
|
|
data = self.data = self._DoReadFileRealDtb('128_decode_image.dts')
|
2019-07-08 20:25:45 +00:00
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
entry = entries['fdtmap']
|
|
|
|
# The header should point to the FDT map
|
|
|
|
self.assertEqual(entry.image_pos, image_header.LocateHeaderOffset(data))
|
|
|
|
|
|
|
|
def testFindImageHeaderStart(self):
|
|
|
|
"""Test locating a image header located at the start of an image"""
|
2019-07-08 20:25:46 +00:00
|
|
|
data = self.data = self._DoReadFileRealDtb('117_fdtmap_hdr_start.dts')
|
2019-07-08 20:25:45 +00:00
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
entry = entries['fdtmap']
|
|
|
|
# The header should point to the FDT map
|
|
|
|
self.assertEqual(entry.image_pos, image_header.LocateHeaderOffset(data))
|
|
|
|
|
|
|
|
def testFindImageHeaderMissing(self):
|
|
|
|
"""Test failing to locate an image header"""
|
|
|
|
data = self._DoReadFile('005_simple.dts')
|
|
|
|
self.assertEqual(None, image_header.LocateHeaderOffset(data))
|
|
|
|
|
2019-07-08 20:25:46 +00:00
|
|
|
def testReadImage(self):
|
|
|
|
"""Test reading an image and accessing its FDT map"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data = self.data = self._DoReadFileRealDtb('128_decode_image.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:46 +00:00
|
|
|
orig_image = control.images['image']
|
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
self.assertEqual(orig_image.GetEntries().keys(),
|
|
|
|
image.GetEntries().keys())
|
|
|
|
|
|
|
|
orig_entry = orig_image.GetEntries()['fdtmap']
|
|
|
|
entry = image.GetEntries()['fdtmap']
|
|
|
|
self.assertEquals(orig_entry.offset, entry.offset)
|
|
|
|
self.assertEquals(orig_entry.size, entry.size)
|
|
|
|
self.assertEquals(orig_entry.image_pos, entry.image_pos)
|
|
|
|
|
|
|
|
def testReadImageNoHeader(self):
|
|
|
|
"""Test accessing an image's FDT map without an image header"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data = self._DoReadFileRealDtb('129_decode_image_nohdr.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:46 +00:00
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
self.assertTrue(isinstance(image, Image))
|
2019-07-20 18:23:50 +00:00
|
|
|
self.assertEqual('image', image.image_name[-5:])
|
2019-07-08 20:25:46 +00:00
|
|
|
|
|
|
|
def testReadImageFail(self):
|
|
|
|
"""Test failing to read an image image's FDT map"""
|
|
|
|
self._DoReadFile('005_simple.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:46 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
self.assertIn("Cannot find FDT map in image", str(e.exception))
|
2019-07-08 19:18:56 +00:00
|
|
|
|
2019-07-08 20:25:48 +00:00
|
|
|
def testListCmd(self):
|
|
|
|
"""Test listing the files in an image using an Fdtmap"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data = self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
|
|
|
|
|
|
|
# lz4 compression size differs depending on the version
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
section_size = entries['section'].size
|
|
|
|
fdt_size = entries['section'].GetEntries()['u-boot-dtb'].size
|
|
|
|
fdtmap_offset = entries['fdtmap'].offset
|
|
|
|
|
2019-07-20 18:24:10 +00:00
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoBinman('ls', '-i', updated_fname)
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
2019-07-08 20:25:48 +00:00
|
|
|
lines = stdout.getvalue().splitlines()
|
|
|
|
expected = [
|
|
|
|
'Name Image-pos Size Entry-type Offset Uncomp-size',
|
|
|
|
'----------------------------------------------------------------------',
|
|
|
|
'main-section 0 c00 section 0',
|
|
|
|
' u-boot 0 4 u-boot 0',
|
|
|
|
' section 100 %x section 100' % section_size,
|
|
|
|
' cbfs 100 400 cbfs 0',
|
|
|
|
' u-boot 138 4 u-boot 38',
|
2019-10-31 13:43:03 +00:00
|
|
|
' u-boot-dtb 180 105 u-boot-dtb 80 3c9',
|
2019-07-08 20:25:48 +00:00
|
|
|
' u-boot-dtb 500 %x u-boot-dtb 400 3c9' % fdt_size,
|
2019-10-31 13:43:03 +00:00
|
|
|
' fdtmap %x 3bd fdtmap %x' %
|
2019-07-08 20:25:48 +00:00
|
|
|
(fdtmap_offset, fdtmap_offset),
|
|
|
|
' image-header bf8 8 image-header bf8',
|
|
|
|
]
|
|
|
|
self.assertEqual(expected, lines)
|
|
|
|
|
|
|
|
def testListCmdFail(self):
|
|
|
|
"""Test failing to list an image"""
|
|
|
|
self._DoReadFile('005_simple.dts')
|
2019-07-20 18:24:10 +00:00
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoBinman('ls', '-i', updated_fname)
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
2019-07-08 20:25:48 +00:00
|
|
|
self.assertIn("Cannot find FDT map in image", str(e.exception))
|
|
|
|
|
|
|
|
def _RunListCmd(self, paths, expected):
|
|
|
|
"""List out entries and check the result
|
|
|
|
|
|
|
|
Args:
|
|
|
|
paths: List of paths to pass to the list command
|
|
|
|
expected: Expected list of filenames to be returned, in order
|
|
|
|
"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:48 +00:00
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
lines = image.GetListEntries(paths)[1]
|
|
|
|
files = [line[0].strip() for line in lines[1:]]
|
|
|
|
self.assertEqual(expected, files)
|
|
|
|
|
|
|
|
def testListCmdSection(self):
|
|
|
|
"""Test listing the files in a section"""
|
|
|
|
self._RunListCmd(['section'],
|
|
|
|
['section', 'cbfs', 'u-boot', 'u-boot-dtb', 'u-boot-dtb'])
|
|
|
|
|
|
|
|
def testListCmdFile(self):
|
|
|
|
"""Test listing a particular file"""
|
|
|
|
self._RunListCmd(['*u-boot-dtb'], ['u-boot-dtb', 'u-boot-dtb'])
|
|
|
|
|
|
|
|
def testListCmdWildcard(self):
|
|
|
|
"""Test listing a wildcarded file"""
|
|
|
|
self._RunListCmd(['*boot*'],
|
|
|
|
['u-boot', 'u-boot', 'u-boot-dtb', 'u-boot-dtb'])
|
|
|
|
|
|
|
|
def testListCmdWildcardMulti(self):
|
|
|
|
"""Test listing a wildcarded file"""
|
|
|
|
self._RunListCmd(['*cb*', '*head*'],
|
|
|
|
['cbfs', 'u-boot', 'u-boot-dtb', 'image-header'])
|
|
|
|
|
|
|
|
def testListCmdEmpty(self):
|
|
|
|
"""Test listing a wildcarded file"""
|
|
|
|
self._RunListCmd(['nothing'], [])
|
|
|
|
|
|
|
|
def testListCmdPath(self):
|
|
|
|
"""Test listing the files in a sub-entry of a section"""
|
|
|
|
self._RunListCmd(['section/cbfs'], ['cbfs', 'u-boot', 'u-boot-dtb'])
|
|
|
|
|
2019-07-08 20:25:50 +00:00
|
|
|
def _RunExtractCmd(self, entry_name, decomp=True):
|
|
|
|
"""Extract an entry from an image
|
|
|
|
|
|
|
|
Args:
|
|
|
|
entry_name: Entry name to extract
|
|
|
|
decomp: True to decompress the data if compressed, False to leave
|
|
|
|
it in its raw uncompressed format
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
data from entry
|
|
|
|
"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:50 +00:00
|
|
|
return control.ReadEntry(image_fname, entry_name, decomp)
|
|
|
|
|
|
|
|
def testExtractSimple(self):
|
|
|
|
"""Test extracting a single file"""
|
|
|
|
data = self._RunExtractCmd('u-boot')
|
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
2019-07-08 20:25:52 +00:00
|
|
|
def testExtractSection(self):
|
|
|
|
"""Test extracting the files in a section"""
|
|
|
|
data = self._RunExtractCmd('section')
|
|
|
|
cbfs_data = data[:0x400]
|
|
|
|
cbfs = cbfs_util.CbfsReader(cbfs_data)
|
2019-10-31 13:43:03 +00:00
|
|
|
self.assertEqual(['u-boot', 'u-boot-dtb', ''], list(cbfs.files.keys()))
|
2019-07-08 20:25:52 +00:00
|
|
|
dtb_data = data[0x400:]
|
|
|
|
dtb = self._decompress(dtb_data)
|
|
|
|
self.assertEqual(EXTRACT_DTB_SIZE, len(dtb))
|
|
|
|
|
|
|
|
def testExtractCompressed(self):
|
|
|
|
"""Test extracting compressed data"""
|
|
|
|
data = self._RunExtractCmd('section/u-boot-dtb')
|
|
|
|
self.assertEqual(EXTRACT_DTB_SIZE, len(data))
|
|
|
|
|
|
|
|
def testExtractRaw(self):
|
|
|
|
"""Test extracting compressed data without decompressing it"""
|
|
|
|
data = self._RunExtractCmd('section/u-boot-dtb', decomp=False)
|
|
|
|
dtb = self._decompress(data)
|
|
|
|
self.assertEqual(EXTRACT_DTB_SIZE, len(dtb))
|
|
|
|
|
|
|
|
def testExtractCbfs(self):
|
|
|
|
"""Test extracting CBFS data"""
|
|
|
|
data = self._RunExtractCmd('section/cbfs/u-boot')
|
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
def testExtractCbfsCompressed(self):
|
|
|
|
"""Test extracting CBFS compressed data"""
|
|
|
|
data = self._RunExtractCmd('section/cbfs/u-boot-dtb')
|
|
|
|
self.assertEqual(EXTRACT_DTB_SIZE, len(data))
|
|
|
|
|
|
|
|
def testExtractCbfsRaw(self):
|
|
|
|
"""Test extracting CBFS compressed data without decompressing it"""
|
2022-08-19 14:25:28 +00:00
|
|
|
bintool = self.comp_bintools['lzma_alone']
|
|
|
|
self._CheckBintool(bintool)
|
2019-07-08 20:25:52 +00:00
|
|
|
data = self._RunExtractCmd('section/cbfs/u-boot-dtb', decomp=False)
|
2022-08-19 14:25:28 +00:00
|
|
|
dtb = bintool.decompress(data)
|
2019-07-08 20:25:52 +00:00
|
|
|
self.assertEqual(EXTRACT_DTB_SIZE, len(dtb))
|
|
|
|
|
2019-07-08 20:25:50 +00:00
|
|
|
def testExtractBadEntry(self):
|
|
|
|
"""Test extracting a bad section path"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._RunExtractCmd('section/does-not-exist')
|
|
|
|
self.assertIn("Entry 'does-not-exist' not found in '/section'",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testExtractMissingFile(self):
|
|
|
|
"""Test extracting file that does not exist"""
|
|
|
|
with self.assertRaises(IOError) as e:
|
|
|
|
control.ReadEntry('missing-file', 'name')
|
|
|
|
|
|
|
|
def testExtractBadFile(self):
|
|
|
|
"""Test extracting an invalid file"""
|
|
|
|
fname = os.path.join(self._indir, 'badfile')
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(fname, b'')
|
2019-07-08 20:25:50 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
control.ReadEntry(fname, 'name')
|
|
|
|
|
2019-07-08 20:25:52 +00:00
|
|
|
def testExtractCmd(self):
|
|
|
|
"""Test extracting a file fron an image on the command line"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
|
|
|
fname = os.path.join(self._indir, 'output.extact')
|
2019-07-20 18:24:10 +00:00
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoBinman('extract', '-i', updated_fname, 'u-boot',
|
|
|
|
'-f', fname)
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(fname)
|
2019-07-08 20:25:52 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
def testExtractOneEntry(self):
|
|
|
|
"""Test extracting a single entry fron an image """
|
|
|
|
self._CheckLz4()
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:52 +00:00
|
|
|
fname = os.path.join(self._indir, 'output.extact')
|
|
|
|
control.ExtractEntries(image_fname, fname, None, ['u-boot'])
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(fname)
|
2019-07-08 20:25:52 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
def _CheckExtractOutput(self, decomp):
|
|
|
|
"""Helper to test file output with and without decompression
|
|
|
|
|
|
|
|
Args:
|
|
|
|
decomp: True to decompress entry data, False to output it raw
|
|
|
|
"""
|
|
|
|
def _CheckPresent(entry_path, expect_data, expect_size=None):
|
|
|
|
"""Check and remove expected file
|
|
|
|
|
|
|
|
This checks the data/size of a file and removes the file both from
|
|
|
|
the outfiles set and from the output directory. Once all files are
|
|
|
|
processed, both the set and directory should be empty.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
entry_path: Entry path
|
|
|
|
expect_data: Data to expect in file, or None to skip check
|
|
|
|
expect_size: Size of data to expect in file, or None to skip
|
|
|
|
"""
|
|
|
|
path = os.path.join(outdir, entry_path)
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(path)
|
2019-07-08 20:25:52 +00:00
|
|
|
os.remove(path)
|
|
|
|
if expect_data:
|
|
|
|
self.assertEqual(expect_data, data)
|
|
|
|
elif expect_size:
|
|
|
|
self.assertEqual(expect_size, len(data))
|
|
|
|
outfiles.remove(path)
|
|
|
|
|
|
|
|
def _CheckDirPresent(name):
|
|
|
|
"""Remove expected directory
|
|
|
|
|
|
|
|
This gives an error if the directory does not exist as expected
|
|
|
|
|
|
|
|
Args:
|
|
|
|
name: Name of directory to remove
|
|
|
|
"""
|
|
|
|
path = os.path.join(outdir, name)
|
|
|
|
os.rmdir(path)
|
|
|
|
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:52 +00:00
|
|
|
outdir = os.path.join(self._indir, 'extract')
|
|
|
|
einfos = control.ExtractEntries(image_fname, None, outdir, [], decomp)
|
|
|
|
|
|
|
|
# Create a set of all file that were output (should be 9)
|
|
|
|
outfiles = set()
|
|
|
|
for root, dirs, files in os.walk(outdir):
|
|
|
|
outfiles |= set([os.path.join(root, fname) for fname in files])
|
|
|
|
self.assertEqual(9, len(outfiles))
|
|
|
|
self.assertEqual(9, len(einfos))
|
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
|
|
|
|
# Check the 9 files in various ways
|
|
|
|
section = entries['section']
|
|
|
|
section_entries = section.GetEntries()
|
|
|
|
cbfs_entries = section_entries['cbfs'].GetEntries()
|
|
|
|
_CheckPresent('u-boot', U_BOOT_DATA)
|
|
|
|
_CheckPresent('section/cbfs/u-boot', U_BOOT_DATA)
|
|
|
|
dtb_len = EXTRACT_DTB_SIZE
|
|
|
|
if not decomp:
|
|
|
|
dtb_len = cbfs_entries['u-boot-dtb'].size
|
|
|
|
_CheckPresent('section/cbfs/u-boot-dtb', None, dtb_len)
|
|
|
|
if not decomp:
|
|
|
|
dtb_len = section_entries['u-boot-dtb'].size
|
|
|
|
_CheckPresent('section/u-boot-dtb', None, dtb_len)
|
|
|
|
|
|
|
|
fdtmap = entries['fdtmap']
|
|
|
|
_CheckPresent('fdtmap', fdtmap.data)
|
|
|
|
hdr = entries['image-header']
|
|
|
|
_CheckPresent('image-header', hdr.data)
|
|
|
|
|
|
|
|
_CheckPresent('section/root', section.data)
|
|
|
|
cbfs = section_entries['cbfs']
|
|
|
|
_CheckPresent('section/cbfs/root', cbfs.data)
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(image_fname)
|
2019-07-08 20:25:52 +00:00
|
|
|
_CheckPresent('root', data)
|
|
|
|
|
|
|
|
# There should be no files left. Remove all the directories to check.
|
|
|
|
# If there are any files/dirs remaining, one of these checks will fail.
|
|
|
|
self.assertEqual(0, len(outfiles))
|
|
|
|
_CheckDirPresent('section/cbfs')
|
|
|
|
_CheckDirPresent('section')
|
|
|
|
_CheckDirPresent('')
|
|
|
|
self.assertFalse(os.path.exists(outdir))
|
|
|
|
|
|
|
|
def testExtractAllEntries(self):
|
|
|
|
"""Test extracting all entries"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._CheckExtractOutput(decomp=True)
|
|
|
|
|
|
|
|
def testExtractAllEntriesRaw(self):
|
|
|
|
"""Test extracting all entries without decompressing them"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._CheckExtractOutput(decomp=False)
|
|
|
|
|
|
|
|
def testExtractSelectedEntries(self):
|
|
|
|
"""Test extracting some entries"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:52 +00:00
|
|
|
outdir = os.path.join(self._indir, 'extract')
|
|
|
|
einfos = control.ExtractEntries(image_fname, None, outdir,
|
|
|
|
['*cb*', '*head*'])
|
|
|
|
|
|
|
|
# File output is tested by testExtractAllEntries(), so just check that
|
|
|
|
# the expected entries are selected
|
|
|
|
names = [einfo.name for einfo in einfos]
|
|
|
|
self.assertEqual(names,
|
|
|
|
['cbfs', 'u-boot', 'u-boot-dtb', 'image-header'])
|
|
|
|
|
|
|
|
def testExtractNoEntryPaths(self):
|
|
|
|
"""Test extracting some entries"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:52 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
control.ExtractEntries(image_fname, 'fname', None, [])
|
2019-07-20 18:24:14 +00:00
|
|
|
self.assertIn('Must specify an entry path to write with -f',
|
2019-07-08 20:25:52 +00:00
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testExtractTooManyEntryPaths(self):
|
|
|
|
"""Test extracting some entries"""
|
|
|
|
self._CheckLz4()
|
|
|
|
self._DoReadFileRealDtb('130_list_fdtmap.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-08 20:25:52 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
control.ExtractEntries(image_fname, 'fname', None, ['a', 'b'])
|
2019-07-20 18:24:14 +00:00
|
|
|
self.assertIn('Must specify exactly one entry path to write with -f',
|
2019-07-08 20:25:52 +00:00
|
|
|
str(e.exception))
|
|
|
|
|
2019-07-08 20:25:53 +00:00
|
|
|
def testPackAlignSection(self):
|
|
|
|
"""Test that sections can have alignment"""
|
|
|
|
self._DoReadFile('131_pack_align_section.dts')
|
|
|
|
|
|
|
|
self.assertIn('image', control.images)
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(3, len(entries))
|
|
|
|
|
|
|
|
# First u-boot
|
|
|
|
self.assertIn('u-boot', entries)
|
|
|
|
entry = entries['u-boot']
|
|
|
|
self.assertEqual(0, entry.offset)
|
|
|
|
self.assertEqual(0, entry.image_pos)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.contents_size)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
|
|
|
# Section0
|
|
|
|
self.assertIn('section0', entries)
|
|
|
|
section0 = entries['section0']
|
|
|
|
self.assertEqual(0x10, section0.offset)
|
|
|
|
self.assertEqual(0x10, section0.image_pos)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), section0.size)
|
|
|
|
|
|
|
|
# Second u-boot
|
|
|
|
section_entries = section0.GetEntries()
|
|
|
|
self.assertIn('u-boot', section_entries)
|
|
|
|
entry = section_entries['u-boot']
|
|
|
|
self.assertEqual(0, entry.offset)
|
|
|
|
self.assertEqual(0x10, entry.image_pos)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.contents_size)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
|
|
|
# Section1
|
|
|
|
self.assertIn('section1', entries)
|
|
|
|
section1 = entries['section1']
|
|
|
|
self.assertEqual(0x14, section1.offset)
|
|
|
|
self.assertEqual(0x14, section1.image_pos)
|
|
|
|
self.assertEqual(0x20, section1.size)
|
|
|
|
|
|
|
|
# Second u-boot
|
|
|
|
section_entries = section1.GetEntries()
|
|
|
|
self.assertIn('u-boot', section_entries)
|
|
|
|
entry = section_entries['u-boot']
|
|
|
|
self.assertEqual(0, entry.offset)
|
|
|
|
self.assertEqual(0x14, entry.image_pos)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.contents_size)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
|
|
|
# Section2
|
|
|
|
self.assertIn('section2', section_entries)
|
|
|
|
section2 = section_entries['section2']
|
|
|
|
self.assertEqual(0x4, section2.offset)
|
|
|
|
self.assertEqual(0x18, section2.image_pos)
|
|
|
|
self.assertEqual(4, section2.size)
|
|
|
|
|
|
|
|
# Third u-boot
|
|
|
|
section_entries = section2.GetEntries()
|
|
|
|
self.assertIn('u-boot', section_entries)
|
|
|
|
entry = section_entries['u-boot']
|
|
|
|
self.assertEqual(0, entry.offset)
|
|
|
|
self.assertEqual(0x18, entry.image_pos)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.contents_size)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
2019-07-20 18:23:56 +00:00
|
|
|
def _RunReplaceCmd(self, entry_name, data, decomp=True, allow_resize=True,
|
|
|
|
dts='132_replace.dts'):
|
2019-07-20 18:23:50 +00:00
|
|
|
"""Replace an entry in an image
|
|
|
|
|
|
|
|
This writes the entry data to update it, then opens the updated file and
|
|
|
|
returns the value that it now finds there.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
entry_name: Entry name to replace
|
|
|
|
data: Data to replace it with
|
|
|
|
decomp: True to compress the data if needed, False if data is
|
|
|
|
already compressed so should be used as is
|
2019-07-20 18:23:56 +00:00
|
|
|
allow_resize: True to allow entries to change size, False to raise
|
|
|
|
an exception
|
2019-07-20 18:23:50 +00:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
Tuple:
|
|
|
|
data from entry
|
|
|
|
data from fdtmap (excluding header)
|
2019-07-20 18:23:56 +00:00
|
|
|
Image object that was modified
|
2019-07-20 18:23:50 +00:00
|
|
|
"""
|
2019-07-20 18:23:56 +00:00
|
|
|
dtb_data = self._DoReadFileDtb(dts, use_real_dtb=True,
|
2019-07-20 18:23:50 +00:00
|
|
|
update_dtb=True)[1]
|
|
|
|
|
|
|
|
self.assertIn('image', control.images)
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
orig_dtb_data = entries['u-boot-dtb'].data
|
|
|
|
orig_fdtmap_data = entries['fdtmap'].data
|
|
|
|
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
updated_fname = tools.get_output_filename('image-updated.bin')
|
|
|
|
tools.write_file(updated_fname, tools.read_file(image_fname))
|
2019-07-20 18:23:56 +00:00
|
|
|
image = control.WriteEntry(updated_fname, entry_name, data, decomp,
|
|
|
|
allow_resize)
|
2019-07-20 18:23:50 +00:00
|
|
|
data = control.ReadEntry(updated_fname, entry_name, decomp)
|
|
|
|
|
2019-07-20 18:23:56 +00:00
|
|
|
# The DT data should not change unless resized:
|
|
|
|
if not allow_resize:
|
|
|
|
new_dtb_data = entries['u-boot-dtb'].data
|
|
|
|
self.assertEqual(new_dtb_data, orig_dtb_data)
|
|
|
|
new_fdtmap_data = entries['fdtmap'].data
|
|
|
|
self.assertEqual(new_fdtmap_data, orig_fdtmap_data)
|
2019-07-20 18:23:50 +00:00
|
|
|
|
2019-07-20 18:23:56 +00:00
|
|
|
return data, orig_fdtmap_data[fdtmap.FDTMAP_HDR_LEN:], image
|
2019-07-20 18:23:50 +00:00
|
|
|
|
|
|
|
def testReplaceSimple(self):
|
|
|
|
"""Test replacing a single file"""
|
|
|
|
expected = b'x' * len(U_BOOT_DATA)
|
2019-07-20 18:23:56 +00:00
|
|
|
data, expected_fdtmap, _ = self._RunReplaceCmd('u-boot', expected,
|
|
|
|
allow_resize=False)
|
2019-07-20 18:23:50 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
# Test that the state looks right. There should be an FDT for the fdtmap
|
|
|
|
# that we jsut read back in, and it should match what we find in the
|
|
|
|
# 'control' tables. Checking for an FDT that does not exist should
|
|
|
|
# return None.
|
|
|
|
path, fdtmap = state.GetFdtContents('fdtmap')
|
2019-07-20 18:23:56 +00:00
|
|
|
self.assertIsNotNone(path)
|
2019-07-20 18:23:50 +00:00
|
|
|
self.assertEqual(expected_fdtmap, fdtmap)
|
|
|
|
|
|
|
|
dtb = state.GetFdtForEtype('fdtmap')
|
|
|
|
self.assertEqual(dtb.GetContents(), fdtmap)
|
|
|
|
|
|
|
|
missing_path, missing_fdtmap = state.GetFdtContents('missing')
|
|
|
|
self.assertIsNone(missing_path)
|
|
|
|
self.assertIsNone(missing_fdtmap)
|
|
|
|
|
|
|
|
missing_dtb = state.GetFdtForEtype('missing')
|
|
|
|
self.assertIsNone(missing_dtb)
|
|
|
|
|
|
|
|
self.assertEqual('/binman', state.fdt_path_prefix)
|
|
|
|
|
|
|
|
def testReplaceResizeFail(self):
|
|
|
|
"""Test replacing a file by something larger"""
|
|
|
|
expected = U_BOOT_DATA + b'x'
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2019-07-20 18:23:56 +00:00
|
|
|
self._RunReplaceCmd('u-boot', expected, allow_resize=False,
|
|
|
|
dts='139_replace_repack.dts')
|
2019-07-20 18:23:50 +00:00
|
|
|
self.assertIn("Node '/u-boot': Entry data size does not match, but resize is disabled",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testReplaceMulti(self):
|
|
|
|
"""Test replacing entry data where multiple images are generated"""
|
|
|
|
data = self._DoReadFileDtb('133_replace_multi.dts', use_real_dtb=True,
|
|
|
|
update_dtb=True)[0]
|
|
|
|
expected = b'x' * len(U_BOOT_DATA)
|
2022-01-29 21:14:04 +00:00
|
|
|
updated_fname = tools.get_output_filename('image-updated.bin')
|
|
|
|
tools.write_file(updated_fname, data)
|
2019-07-20 18:23:50 +00:00
|
|
|
entry_name = 'u-boot'
|
2019-07-20 18:23:56 +00:00
|
|
|
control.WriteEntry(updated_fname, entry_name, expected,
|
|
|
|
allow_resize=False)
|
2019-07-20 18:23:50 +00:00
|
|
|
data = control.ReadEntry(updated_fname, entry_name)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
# Check the state looks right.
|
|
|
|
self.assertEqual('/binman/image', state.fdt_path_prefix)
|
|
|
|
|
|
|
|
# Now check we can write the first image
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('first-image.bin')
|
|
|
|
updated_fname = tools.get_output_filename('first-updated.bin')
|
|
|
|
tools.write_file(updated_fname, tools.read_file(image_fname))
|
2019-07-20 18:23:50 +00:00
|
|
|
entry_name = 'u-boot'
|
2019-07-20 18:23:56 +00:00
|
|
|
control.WriteEntry(updated_fname, entry_name, expected,
|
|
|
|
allow_resize=False)
|
2019-07-20 18:23:50 +00:00
|
|
|
data = control.ReadEntry(updated_fname, entry_name)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
# Check the state looks right.
|
|
|
|
self.assertEqual('/binman/first-image', state.fdt_path_prefix)
|
2019-07-08 20:25:47 +00:00
|
|
|
|
2019-07-20 18:23:51 +00:00
|
|
|
def testUpdateFdtAllRepack(self):
|
|
|
|
"""Test that all device trees are updated with offset/size info"""
|
|
|
|
data = self._DoReadFileRealDtb('134_fdt_update_all_repack.dts')
|
|
|
|
SECTION_SIZE = 0x300
|
|
|
|
DTB_SIZE = 602
|
|
|
|
FDTMAP_SIZE = 608
|
|
|
|
base_expected = {
|
|
|
|
'offset': 0,
|
|
|
|
'size': SECTION_SIZE + DTB_SIZE * 2 + FDTMAP_SIZE,
|
|
|
|
'image-pos': 0,
|
|
|
|
'section:offset': 0,
|
|
|
|
'section:size': SECTION_SIZE,
|
|
|
|
'section:image-pos': 0,
|
|
|
|
'section/u-boot-dtb:offset': 4,
|
|
|
|
'section/u-boot-dtb:size': 636,
|
|
|
|
'section/u-boot-dtb:image-pos': 4,
|
|
|
|
'u-boot-spl-dtb:offset': SECTION_SIZE,
|
|
|
|
'u-boot-spl-dtb:size': DTB_SIZE,
|
|
|
|
'u-boot-spl-dtb:image-pos': SECTION_SIZE,
|
|
|
|
'u-boot-tpl-dtb:offset': SECTION_SIZE + DTB_SIZE,
|
|
|
|
'u-boot-tpl-dtb:image-pos': SECTION_SIZE + DTB_SIZE,
|
|
|
|
'u-boot-tpl-dtb:size': DTB_SIZE,
|
|
|
|
'fdtmap:offset': SECTION_SIZE + DTB_SIZE * 2,
|
|
|
|
'fdtmap:size': FDTMAP_SIZE,
|
|
|
|
'fdtmap:image-pos': SECTION_SIZE + DTB_SIZE * 2,
|
|
|
|
}
|
|
|
|
main_expected = {
|
|
|
|
'section:orig-size': SECTION_SIZE,
|
|
|
|
'section/u-boot-dtb:orig-offset': 4,
|
|
|
|
}
|
|
|
|
|
|
|
|
# We expect three device-tree files in the output, with the first one
|
|
|
|
# within a fixed-size section.
|
|
|
|
# Read them in sequence. We look for an 'spl' property in the SPL tree,
|
|
|
|
# and 'tpl' in the TPL tree, to make sure they are distinct from the
|
|
|
|
# main U-Boot tree. All three should have the same positions and offset
|
|
|
|
# except that the main tree should include the main_expected properties
|
|
|
|
start = 4
|
|
|
|
for item in ['', 'spl', 'tpl', None]:
|
|
|
|
if item is None:
|
|
|
|
start += 16 # Move past fdtmap header
|
|
|
|
dtb = fdt.Fdt.FromData(data[start:])
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb,
|
|
|
|
BASE_DTB_PROPS + REPACK_DTB_PROPS + ['spl', 'tpl'],
|
|
|
|
prefix='/' if item is None else '/binman/')
|
|
|
|
expected = dict(base_expected)
|
|
|
|
if item:
|
|
|
|
expected[item] = 0
|
|
|
|
else:
|
|
|
|
# Main DTB and fdtdec should include the 'orig-' properties
|
|
|
|
expected.update(main_expected)
|
|
|
|
# Helpful for debugging:
|
|
|
|
#for prop in sorted(props):
|
|
|
|
#print('prop %s %s %s' % (prop, props[prop], expected[prop]))
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
if item == '':
|
|
|
|
start = SECTION_SIZE
|
|
|
|
else:
|
|
|
|
start += dtb._fdt_obj.totalsize()
|
|
|
|
|
2019-07-20 18:23:55 +00:00
|
|
|
def testFdtmapHeaderMiddle(self):
|
|
|
|
"""Test an FDT map in the middle of an image when it should be at end"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFileRealDtb('135_fdtmap_hdr_middle.dts')
|
|
|
|
self.assertIn("Invalid sibling order 'middle' for image-header: Must be at 'end' to match location",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testFdtmapHeaderStartBad(self):
|
|
|
|
"""Test an FDT map in middle of an image when it should be at start"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFileRealDtb('136_fdtmap_hdr_startbad.dts')
|
|
|
|
self.assertIn("Invalid sibling order 'end' for image-header: Must be at 'start' to match location",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testFdtmapHeaderEndBad(self):
|
|
|
|
"""Test an FDT map at the start of an image when it should be at end"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFileRealDtb('137_fdtmap_hdr_endbad.dts')
|
|
|
|
self.assertIn("Invalid sibling order 'start' for image-header: Must be at 'end' to match location",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testFdtmapHeaderNoSize(self):
|
|
|
|
"""Test an image header at the end of an image with undefined size"""
|
|
|
|
self._DoReadFileRealDtb('138_fdtmap_hdr_nosize.dts')
|
|
|
|
|
2019-07-20 18:23:56 +00:00
|
|
|
def testReplaceResize(self):
|
|
|
|
"""Test replacing a single file in an entry with a larger file"""
|
|
|
|
expected = U_BOOT_DATA + b'x'
|
|
|
|
data, _, image = self._RunReplaceCmd('u-boot', expected,
|
|
|
|
dts='139_replace_repack.dts')
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
entries = image.GetEntries()
|
|
|
|
dtb_data = entries['u-boot-dtb'].data
|
|
|
|
dtb = fdt.Fdt.FromData(dtb_data)
|
|
|
|
dtb.Scan()
|
|
|
|
|
|
|
|
# The u-boot section should now be larger in the dtb
|
|
|
|
node = dtb.GetNode('/binman/u-boot')
|
|
|
|
self.assertEqual(len(expected), fdt_util.GetInt(node, 'size'))
|
|
|
|
|
|
|
|
# Same for the fdtmap
|
|
|
|
fdata = entries['fdtmap'].data
|
|
|
|
fdtb = fdt.Fdt.FromData(fdata[fdtmap.FDTMAP_HDR_LEN:])
|
|
|
|
fdtb.Scan()
|
|
|
|
fnode = fdtb.GetNode('/u-boot')
|
|
|
|
self.assertEqual(len(expected), fdt_util.GetInt(fnode, 'size'))
|
|
|
|
|
|
|
|
def testReplaceResizeNoRepack(self):
|
|
|
|
"""Test replacing an entry with a larger file when not allowed"""
|
|
|
|
expected = U_BOOT_DATA + b'x'
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._RunReplaceCmd('u-boot', expected)
|
|
|
|
self.assertIn('Entry data size does not match, but allow-repack is not present for this image',
|
|
|
|
str(e.exception))
|
|
|
|
|
2019-07-20 18:23:58 +00:00
|
|
|
def testEntryShrink(self):
|
|
|
|
"""Test contracting an entry after it is packed"""
|
|
|
|
try:
|
|
|
|
state.SetAllowEntryContraction(True)
|
|
|
|
data = self._DoReadFileDtb('140_entry_shrink.dts',
|
|
|
|
update_dtb=True)[0]
|
|
|
|
finally:
|
|
|
|
state.SetAllowEntryContraction(False)
|
|
|
|
self.assertEqual(b'a', data[:1])
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[1:1 + len(U_BOOT_DATA)])
|
|
|
|
self.assertEqual(b'a', data[-1:])
|
|
|
|
|
|
|
|
def testEntryShrinkFail(self):
|
|
|
|
"""Test not being allowed to contract an entry after it is packed"""
|
|
|
|
data = self._DoReadFileDtb('140_entry_shrink.dts', update_dtb=True)[0]
|
|
|
|
|
|
|
|
# In this case there is a spare byte at the end of the data. The size of
|
|
|
|
# the contents is only 1 byte but we still have the size before it
|
|
|
|
# shrunk.
|
|
|
|
self.assertEqual(b'a\0', data[:2])
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[2:2 + len(U_BOOT_DATA)])
|
|
|
|
self.assertEqual(b'a\0', data[-2:])
|
|
|
|
|
2019-07-20 18:24:01 +00:00
|
|
|
def testDescriptorOffset(self):
|
|
|
|
"""Test that the Intel descriptor is always placed at at the start"""
|
|
|
|
data = self._DoReadFileDtb('141_descriptor_offset.dts')
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
desc = entries['intel-descriptor']
|
|
|
|
self.assertEqual(0xff800000, desc.offset);
|
|
|
|
self.assertEqual(0xff800000, desc.image_pos);
|
|
|
|
|
2019-07-20 18:24:06 +00:00
|
|
|
def testReplaceCbfs(self):
|
|
|
|
"""Test replacing a single file in CBFS without changing the size"""
|
|
|
|
self._CheckLz4()
|
|
|
|
expected = b'x' * len(U_BOOT_DATA)
|
|
|
|
data = self._DoReadFileRealDtb('142_replace_cbfs.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
updated_fname = tools.get_output_filename('image-updated.bin')
|
|
|
|
tools.write_file(updated_fname, data)
|
2019-07-20 18:24:06 +00:00
|
|
|
entry_name = 'section/cbfs/u-boot'
|
|
|
|
control.WriteEntry(updated_fname, entry_name, expected,
|
|
|
|
allow_resize=True)
|
|
|
|
data = control.ReadEntry(updated_fname, entry_name)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testReplaceResizeCbfs(self):
|
|
|
|
"""Test replacing a single file in CBFS with one of a different size"""
|
|
|
|
self._CheckLz4()
|
|
|
|
expected = U_BOOT_DATA + b'x'
|
|
|
|
data = self._DoReadFileRealDtb('142_replace_cbfs.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
updated_fname = tools.get_output_filename('image-updated.bin')
|
|
|
|
tools.write_file(updated_fname, data)
|
2019-07-20 18:24:06 +00:00
|
|
|
entry_name = 'section/cbfs/u-boot'
|
|
|
|
control.WriteEntry(updated_fname, entry_name, expected,
|
|
|
|
allow_resize=True)
|
|
|
|
data = control.ReadEntry(updated_fname, entry_name)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2019-07-20 18:24:15 +00:00
|
|
|
def _SetupForReplace(self):
|
|
|
|
"""Set up some files to use to replace entries
|
|
|
|
|
|
|
|
This generates an image, copies it to a new file, extracts all the files
|
|
|
|
in it and updates some of them
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
List
|
|
|
|
Image filename
|
|
|
|
Output directory
|
|
|
|
Expected values for updated entries, each a string
|
|
|
|
"""
|
|
|
|
data = self._DoReadFileRealDtb('143_replace_all.dts')
|
|
|
|
|
2022-01-29 21:14:04 +00:00
|
|
|
updated_fname = tools.get_output_filename('image-updated.bin')
|
|
|
|
tools.write_file(updated_fname, data)
|
2019-07-20 18:24:15 +00:00
|
|
|
|
|
|
|
outdir = os.path.join(self._indir, 'extract')
|
|
|
|
einfos = control.ExtractEntries(updated_fname, None, outdir, [])
|
|
|
|
|
|
|
|
expected1 = b'x' + U_BOOT_DATA + b'y'
|
|
|
|
u_boot_fname1 = os.path.join(outdir, 'u-boot')
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(u_boot_fname1, expected1)
|
2019-07-20 18:24:15 +00:00
|
|
|
|
|
|
|
expected2 = b'a' + U_BOOT_DATA + b'b'
|
|
|
|
u_boot_fname2 = os.path.join(outdir, 'u-boot2')
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(u_boot_fname2, expected2)
|
2019-07-20 18:24:15 +00:00
|
|
|
|
|
|
|
expected_text = b'not the same text'
|
|
|
|
text_fname = os.path.join(outdir, 'text')
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(text_fname, expected_text)
|
2019-07-20 18:24:15 +00:00
|
|
|
|
|
|
|
dtb_fname = os.path.join(outdir, 'u-boot-dtb')
|
|
|
|
dtb = fdt.FdtScan(dtb_fname)
|
|
|
|
node = dtb.GetNode('/binman/text')
|
|
|
|
node.AddString('my-property', 'the value')
|
|
|
|
dtb.Sync(auto_resize=True)
|
|
|
|
dtb.Flush()
|
|
|
|
|
|
|
|
return updated_fname, outdir, expected1, expected2, expected_text
|
|
|
|
|
|
|
|
def _CheckReplaceMultiple(self, entry_paths):
|
|
|
|
"""Handle replacing the contents of multiple entries
|
|
|
|
|
|
|
|
Args:
|
|
|
|
entry_paths: List of entry paths to replace
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
List
|
|
|
|
Dict of entries in the image:
|
|
|
|
key: Entry name
|
|
|
|
Value: Entry object
|
|
|
|
Expected values for updated entries, each a string
|
|
|
|
"""
|
|
|
|
updated_fname, outdir, expected1, expected2, expected_text = (
|
|
|
|
self._SetupForReplace())
|
|
|
|
control.ReplaceEntries(updated_fname, None, outdir, entry_paths)
|
|
|
|
|
|
|
|
image = Image.FromFile(updated_fname)
|
|
|
|
image.LoadData()
|
|
|
|
return image.GetEntries(), expected1, expected2, expected_text
|
|
|
|
|
|
|
|
def testReplaceAll(self):
|
|
|
|
"""Test replacing the contents of all entries"""
|
|
|
|
entries, expected1, expected2, expected_text = (
|
|
|
|
self._CheckReplaceMultiple([]))
|
|
|
|
data = entries['u-boot'].data
|
|
|
|
self.assertEqual(expected1, data)
|
|
|
|
|
|
|
|
data = entries['u-boot2'].data
|
|
|
|
self.assertEqual(expected2, data)
|
|
|
|
|
|
|
|
data = entries['text'].data
|
|
|
|
self.assertEqual(expected_text, data)
|
|
|
|
|
|
|
|
# Check that the device tree is updated
|
|
|
|
data = entries['u-boot-dtb'].data
|
|
|
|
dtb = fdt.Fdt.FromData(data)
|
|
|
|
dtb.Scan()
|
|
|
|
node = dtb.GetNode('/binman/text')
|
|
|
|
self.assertEqual('the value', node.props['my-property'].value)
|
|
|
|
|
|
|
|
def testReplaceSome(self):
|
|
|
|
"""Test replacing the contents of a few entries"""
|
|
|
|
entries, expected1, expected2, expected_text = (
|
|
|
|
self._CheckReplaceMultiple(['u-boot2', 'text']))
|
|
|
|
|
|
|
|
# This one should not change
|
|
|
|
data = entries['u-boot'].data
|
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
data = entries['u-boot2'].data
|
|
|
|
self.assertEqual(expected2, data)
|
|
|
|
|
|
|
|
data = entries['text'].data
|
|
|
|
self.assertEqual(expected_text, data)
|
|
|
|
|
|
|
|
def testReplaceCmd(self):
|
|
|
|
"""Test replacing a file fron an image on the command line"""
|
|
|
|
self._DoReadFileRealDtb('143_replace_all.dts')
|
|
|
|
|
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
|
|
|
|
fname = os.path.join(tmpdir, 'update-u-boot.bin')
|
|
|
|
expected = b'x' * len(U_BOOT_DATA)
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(fname, expected)
|
2019-07-20 18:24:15 +00:00
|
|
|
|
|
|
|
self._DoBinman('replace', '-i', updated_fname, 'u-boot', '-f', fname)
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(updated_fname)
|
2019-07-20 18:24:15 +00:00
|
|
|
self.assertEqual(expected, data[:len(expected)])
|
|
|
|
map_fname = os.path.join(tmpdir, 'image-updated.map')
|
|
|
|
self.assertFalse(os.path.exists(map_fname))
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
|
|
|
|
|
|
|
def testReplaceCmdSome(self):
|
|
|
|
"""Test replacing some files fron an image on the command line"""
|
|
|
|
updated_fname, outdir, expected1, expected2, expected_text = (
|
|
|
|
self._SetupForReplace())
|
|
|
|
|
|
|
|
self._DoBinman('replace', '-i', updated_fname, '-I', outdir,
|
|
|
|
'u-boot2', 'text')
|
|
|
|
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.prepare_output_dir(None)
|
2019-07-20 18:24:15 +00:00
|
|
|
image = Image.FromFile(updated_fname)
|
|
|
|
image.LoadData()
|
|
|
|
entries = image.GetEntries()
|
|
|
|
|
|
|
|
# This one should not change
|
|
|
|
data = entries['u-boot'].data
|
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
data = entries['u-boot2'].data
|
|
|
|
self.assertEqual(expected2, data)
|
|
|
|
|
|
|
|
data = entries['text'].data
|
|
|
|
self.assertEqual(expected_text, data)
|
|
|
|
|
|
|
|
def testReplaceMissing(self):
|
|
|
|
"""Test replacing entries where the file is missing"""
|
|
|
|
updated_fname, outdir, expected1, expected2, expected_text = (
|
|
|
|
self._SetupForReplace())
|
|
|
|
|
|
|
|
# Remove one of the files, to generate a warning
|
|
|
|
u_boot_fname1 = os.path.join(outdir, 'u-boot')
|
|
|
|
os.remove(u_boot_fname1)
|
|
|
|
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
control.ReplaceEntries(updated_fname, None, outdir, [])
|
|
|
|
self.assertIn("Skipping entry '/u-boot' from missing file",
|
2020-07-10 00:39:39 +00:00
|
|
|
stderr.getvalue())
|
2019-07-20 18:24:15 +00:00
|
|
|
|
|
|
|
def testReplaceCmdMap(self):
|
|
|
|
"""Test replacing a file fron an image on the command line"""
|
|
|
|
self._DoReadFileRealDtb('143_replace_all.dts')
|
|
|
|
|
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
|
|
|
|
fname = os.path.join(self._indir, 'update-u-boot.bin')
|
|
|
|
expected = b'x' * len(U_BOOT_DATA)
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(fname, expected)
|
2019-07-20 18:24:15 +00:00
|
|
|
|
|
|
|
self._DoBinman('replace', '-i', updated_fname, 'u-boot',
|
|
|
|
'-f', fname, '-m')
|
|
|
|
map_fname = os.path.join(tmpdir, 'image-updated.map')
|
|
|
|
self.assertTrue(os.path.exists(map_fname))
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
|
|
|
|
|
|
|
def testReplaceNoEntryPaths(self):
|
|
|
|
"""Test replacing an entry without an entry path"""
|
|
|
|
self._DoReadFileRealDtb('143_replace_all.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-20 18:24:15 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
control.ReplaceEntries(image_fname, 'fname', None, [])
|
|
|
|
self.assertIn('Must specify an entry path to read with -f',
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testReplaceTooManyEntryPaths(self):
|
|
|
|
"""Test extracting some entries"""
|
|
|
|
self._DoReadFileRealDtb('143_replace_all.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2019-07-20 18:24:15 +00:00
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
control.ReplaceEntries(image_fname, 'fname', None, ['a', 'b'])
|
|
|
|
self.assertIn('Must specify exactly one entry path to write with -f',
|
|
|
|
str(e.exception))
|
|
|
|
|
2019-08-24 13:22:48 +00:00
|
|
|
def testPackReset16(self):
|
|
|
|
"""Test that an image with an x86 reset16 region can be created"""
|
|
|
|
data = self._DoReadFile('144_x86_reset16.dts')
|
|
|
|
self.assertEqual(X86_RESET16_DATA, data[:len(X86_RESET16_DATA)])
|
|
|
|
|
|
|
|
def testPackReset16Spl(self):
|
|
|
|
"""Test that an image with an x86 reset16-spl region can be created"""
|
|
|
|
data = self._DoReadFile('145_x86_reset16_spl.dts')
|
|
|
|
self.assertEqual(X86_RESET16_SPL_DATA, data[:len(X86_RESET16_SPL_DATA)])
|
|
|
|
|
|
|
|
def testPackReset16Tpl(self):
|
|
|
|
"""Test that an image with an x86 reset16-tpl region can be created"""
|
|
|
|
data = self._DoReadFile('146_x86_reset16_tpl.dts')
|
|
|
|
self.assertEqual(X86_RESET16_TPL_DATA, data[:len(X86_RESET16_TPL_DATA)])
|
|
|
|
|
2019-08-24 13:22:50 +00:00
|
|
|
def testPackIntelFit(self):
|
|
|
|
"""Test that an image with an Intel FIT and pointer can be created"""
|
|
|
|
data = self._DoReadFile('147_intel_fit.dts')
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[:len(U_BOOT_DATA)])
|
|
|
|
fit = data[16:32];
|
|
|
|
self.assertEqual(b'_FIT_ \x01\x00\x00\x00\x00\x01\x80}' , fit)
|
|
|
|
ptr = struct.unpack('<i', data[0x40:0x44])[0]
|
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
expected_ptr = entries['intel-fit'].image_pos - (1 << 32)
|
|
|
|
self.assertEqual(expected_ptr, ptr)
|
|
|
|
|
|
|
|
def testPackIntelFitMissing(self):
|
|
|
|
"""Test detection of a FIT pointer with not FIT region"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('148_intel_fit_missing.dts')
|
|
|
|
self.assertIn("'intel-fit-ptr' section must have an 'intel-fit' sibling",
|
|
|
|
str(e.exception))
|
|
|
|
|
2019-11-07 00:22:44 +00:00
|
|
|
def _CheckSymbolsTplSection(self, dts, expected_vals):
|
|
|
|
data = self._DoReadFile(dts)
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
sym_values = struct.pack('<LLQLL', elf.BINMAN_SYM_MAGIC_VALUE, *expected_vals)
|
2019-08-24 13:23:00 +00:00
|
|
|
upto1 = 4 + len(U_BOOT_SPL_DATA)
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
expected1 = tools.get_bytes(0xff, 4) + sym_values + U_BOOT_SPL_DATA[24:]
|
2019-08-24 13:23:00 +00:00
|
|
|
self.assertEqual(expected1, data[:upto1])
|
|
|
|
|
|
|
|
upto2 = upto1 + 1 + len(U_BOOT_SPL_DATA)
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
expected2 = tools.get_bytes(0xff, 1) + sym_values + U_BOOT_SPL_DATA[24:]
|
2019-08-24 13:23:00 +00:00
|
|
|
self.assertEqual(expected2, data[upto1:upto2])
|
|
|
|
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
upto3 = 0x3c + len(U_BOOT_DATA)
|
2022-01-29 21:14:04 +00:00
|
|
|
expected3 = tools.get_bytes(0xff, 1) + U_BOOT_DATA
|
2019-08-24 13:23:00 +00:00
|
|
|
self.assertEqual(expected3, data[upto2:upto3])
|
|
|
|
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
expected4 = sym_values + U_BOOT_TPL_DATA[24:]
|
2019-11-07 00:22:44 +00:00
|
|
|
self.assertEqual(expected4, data[upto3:upto3 + len(U_BOOT_TPL_DATA)])
|
|
|
|
|
|
|
|
def testSymbolsTplSection(self):
|
|
|
|
"""Test binman can assign symbols embedded in U-Boot TPL in a section"""
|
|
|
|
self._SetupSplElf('u_boot_binman_syms')
|
|
|
|
self._SetupTplElf('u_boot_binman_syms')
|
|
|
|
self._CheckSymbolsTplSection('149_symbols_tpl.dts',
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
[0x04, 0x20, 0x10 + 0x3c, 0x04])
|
2019-11-07 00:22:44 +00:00
|
|
|
|
|
|
|
def testSymbolsTplSectionX86(self):
|
|
|
|
"""Test binman can assign symbols in a section with end-at-4gb"""
|
|
|
|
self._SetupSplElf('u_boot_binman_syms_x86')
|
|
|
|
self._SetupTplElf('u_boot_binman_syms_x86')
|
|
|
|
self._CheckSymbolsTplSection('155_symbols_tpl_x86.dts',
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
[0xffffff04, 0xffffff20, 0xffffff3c,
|
2019-11-07 00:22:44 +00:00
|
|
|
0x04])
|
2019-08-24 13:23:00 +00:00
|
|
|
|
2019-08-24 13:23:03 +00:00
|
|
|
def testPackX86RomIfwiSectiom(self):
|
|
|
|
"""Test that a section can be placed in an IFWI region"""
|
|
|
|
self._SetupIfwi('fitimage.bin')
|
|
|
|
data = self._DoReadFile('151_x86_rom_ifwi_section.dts')
|
|
|
|
self._CheckIfwi(data)
|
|
|
|
|
2019-08-24 13:23:07 +00:00
|
|
|
def testPackFspM(self):
|
|
|
|
"""Test that an image with a FSP memory-init binary can be created"""
|
|
|
|
data = self._DoReadFile('152_intel_fsp_m.dts')
|
|
|
|
self.assertEqual(FSP_M_DATA, data[:len(FSP_M_DATA)])
|
|
|
|
|
2019-10-21 03:31:35 +00:00
|
|
|
def testPackFspS(self):
|
|
|
|
"""Test that an image with a FSP silicon-init binary can be created"""
|
|
|
|
data = self._DoReadFile('153_intel_fsp_s.dts')
|
|
|
|
self.assertEqual(FSP_S_DATA, data[:len(FSP_S_DATA)])
|
2019-08-24 13:23:07 +00:00
|
|
|
|
2019-10-21 03:31:36 +00:00
|
|
|
def testPackFspT(self):
|
|
|
|
"""Test that an image with a FSP temp-ram-init binary can be created"""
|
|
|
|
data = self._DoReadFile('154_intel_fsp_t.dts')
|
|
|
|
self.assertEqual(FSP_T_DATA, data[:len(FSP_T_DATA)])
|
|
|
|
|
2020-07-10 00:39:31 +00:00
|
|
|
def testMkimage(self):
|
|
|
|
"""Test using mkimage to build an image"""
|
|
|
|
data = self._DoReadFile('156_mkimage.dts')
|
|
|
|
|
|
|
|
# Just check that the data appears in the file somewhere
|
|
|
|
self.assertIn(U_BOOT_SPL_DATA, data)
|
|
|
|
|
2022-01-10 03:14:09 +00:00
|
|
|
def testMkimageMissing(self):
|
|
|
|
"""Test that binman still produces an image if mkimage is missing"""
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoTestFile('156_mkimage.dts',
|
|
|
|
force_missing_bintools='mkimage')
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing bintools.*: mkimage")
|
|
|
|
|
2020-07-10 00:39:36 +00:00
|
|
|
def testExtblob(self):
|
|
|
|
"""Test an image with an external blob"""
|
|
|
|
data = self._DoReadFile('157_blob_ext.dts')
|
|
|
|
self.assertEqual(REFCODE_DATA, data)
|
|
|
|
|
|
|
|
def testExtblobMissing(self):
|
|
|
|
"""Test an image with a missing external blob"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('158_blob_ext_missing.dts')
|
|
|
|
self.assertIn("Filename 'missing-file' not found in input path",
|
|
|
|
str(e.exception))
|
|
|
|
|
2020-07-10 00:39:38 +00:00
|
|
|
def testExtblobMissingOk(self):
|
|
|
|
"""Test an image with an missing external blob that is allowed"""
|
2020-07-10 00:39:40 +00:00
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
2022-11-10 02:14:42 +00:00
|
|
|
ret = self._DoTestFile('158_blob_ext_missing.dts',
|
|
|
|
allow_missing=True)
|
|
|
|
self.assertEqual(103, ret)
|
2020-07-10 00:39:40 +00:00
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err, "Image 'main-section'.*missing.*: blob-ext")
|
2022-11-10 02:14:42 +00:00
|
|
|
self.assertIn('Some images are invalid', err)
|
|
|
|
|
|
|
|
def testExtblobMissingOkFlag(self):
|
|
|
|
"""Test an image with an missing external blob allowed with -W"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
ret = self._DoTestFile('158_blob_ext_missing.dts',
|
|
|
|
allow_missing=True, ignore_missing=True)
|
|
|
|
self.assertEqual(0, ret)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err, "Image 'main-section'.*missing.*: blob-ext")
|
|
|
|
self.assertIn('Some images are invalid', err)
|
2020-07-10 00:39:40 +00:00
|
|
|
|
|
|
|
def testExtblobMissingOkSect(self):
|
|
|
|
"""Test an image with an missing external blob that is allowed"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('159_blob_ext_missing_sect.dts',
|
|
|
|
allow_missing=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err, "Image 'main-section'.*missing.*: "
|
|
|
|
"blob-ext blob-ext2")
|
2020-07-10 00:39:38 +00:00
|
|
|
|
2020-07-10 00:39:41 +00:00
|
|
|
def testPackX86RomMeMissingDesc(self):
|
|
|
|
"""Test that an missing Intel descriptor entry is allowed"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
2020-07-25 21:11:19 +00:00
|
|
|
self._DoTestFile('164_x86_rom_me_missing.dts', allow_missing=True)
|
2020-07-10 00:39:41 +00:00
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing.*: intel-descriptor")
|
|
|
|
|
|
|
|
def testPackX86RomMissingIfwi(self):
|
|
|
|
"""Test that an x86 ROM with Integrated Firmware Image can be created"""
|
|
|
|
self._SetupIfwi('fitimage.bin')
|
|
|
|
pathname = os.path.join(self._indir, 'fitimage.bin')
|
|
|
|
os.remove(pathname)
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('111_x86_rom_ifwi.dts', allow_missing=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err, "Image 'main-section'.*missing.*: intel-ifwi")
|
|
|
|
|
2022-02-11 20:23:21 +00:00
|
|
|
def testPackOverlapZero(self):
|
2020-07-10 00:39:42 +00:00
|
|
|
"""Test that zero-size overlapping regions are ignored"""
|
|
|
|
self._DoTestFile('160_pack_overlap_zero.dts')
|
|
|
|
|
2022-02-07 22:08:04 +00:00
|
|
|
def _CheckSimpleFitData(self, fit_data, kernel_data, fdt1_data):
|
2020-07-10 00:39:45 +00:00
|
|
|
# The data should be inside the FIT
|
|
|
|
dtb = fdt.Fdt.FromData(fit_data)
|
|
|
|
dtb.Scan()
|
|
|
|
fnode = dtb.GetNode('/images/kernel')
|
|
|
|
self.assertIn('data', fnode.props)
|
|
|
|
|
|
|
|
fname = os.path.join(self._indir, 'fit_data.fit')
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(fname, fit_data)
|
|
|
|
out = tools.run('dumpimage', '-l', fname)
|
2020-07-10 00:39:45 +00:00
|
|
|
|
|
|
|
# Check a few features to make sure the plumbing works. We don't need
|
|
|
|
# to test the operation of mkimage or dumpimage here. First convert the
|
|
|
|
# output into a dict where the keys are the fields printed by dumpimage
|
|
|
|
# and the values are a list of values for each field
|
|
|
|
lines = out.splitlines()
|
|
|
|
|
|
|
|
# Converts "Compression: gzip compressed" into two groups:
|
|
|
|
# 'Compression' and 'gzip compressed'
|
|
|
|
re_line = re.compile(r'^ *([^:]*)(?:: *(.*))?$')
|
|
|
|
vals = collections.defaultdict(list)
|
|
|
|
for line in lines:
|
|
|
|
mat = re_line.match(line)
|
|
|
|
vals[mat.group(1)].append(mat.group(2))
|
|
|
|
|
|
|
|
self.assertEquals('FIT description: test-desc', lines[0])
|
|
|
|
self.assertIn('Created:', lines[1])
|
|
|
|
self.assertIn('Image 0 (kernel)', vals)
|
|
|
|
self.assertIn('Hash value', vals)
|
|
|
|
data_sizes = vals.get('Data Size')
|
|
|
|
self.assertIsNotNone(data_sizes)
|
|
|
|
self.assertEqual(2, len(data_sizes))
|
|
|
|
# Format is "4 Bytes = 0.00 KiB = 0.00 MiB" so take the first word
|
2022-02-07 22:08:04 +00:00
|
|
|
self.assertEqual(len(kernel_data), int(data_sizes[0].split()[0]))
|
|
|
|
self.assertEqual(len(fdt1_data), int(data_sizes[1].split()[0]))
|
|
|
|
|
2022-03-27 15:31:47 +00:00
|
|
|
# Check if entry listing correctly omits /images/
|
|
|
|
image = control.images['image']
|
|
|
|
fit_entry = image.GetEntries()['fit']
|
|
|
|
subentries = list(fit_entry.GetEntries().keys())
|
|
|
|
expected = ['kernel', 'fdt-1']
|
|
|
|
self.assertEqual(expected, subentries)
|
|
|
|
|
2022-02-07 22:08:04 +00:00
|
|
|
def testSimpleFit(self):
|
|
|
|
"""Test an image with a FIT inside"""
|
|
|
|
data = self._DoReadFile('161_fit.dts')
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[:len(U_BOOT_DATA)])
|
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA, data[-len(U_BOOT_NODTB_DATA):])
|
|
|
|
fit_data = data[len(U_BOOT_DATA):-len(U_BOOT_NODTB_DATA)]
|
|
|
|
|
|
|
|
self._CheckSimpleFitData(fit_data, U_BOOT_DATA, U_BOOT_SPL_DTB_DATA)
|
|
|
|
|
|
|
|
def testSimpleFitExpandsSubentries(self):
|
|
|
|
"""Test that FIT images expand their subentries"""
|
|
|
|
data = self._DoReadFileDtb('161_fit.dts', use_expanded=True)[0]
|
|
|
|
self.assertEqual(U_BOOT_EXP_DATA, data[:len(U_BOOT_EXP_DATA)])
|
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA, data[-len(U_BOOT_NODTB_DATA):])
|
|
|
|
fit_data = data[len(U_BOOT_EXP_DATA):-len(U_BOOT_NODTB_DATA)]
|
|
|
|
|
|
|
|
self._CheckSimpleFitData(fit_data, U_BOOT_EXP_DATA, U_BOOT_SPL_DTB_DATA)
|
2020-07-10 00:39:45 +00:00
|
|
|
|
2022-02-07 22:08:08 +00:00
|
|
|
def testSimpleFitImagePos(self):
|
|
|
|
"""Test that we have correct image-pos for FIT subentries"""
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb('161_fit.dts',
|
|
|
|
update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS + REPACK_DTB_PROPS)
|
|
|
|
|
2022-03-06 03:19:01 +00:00
|
|
|
self.maxDiff = None
|
2022-02-07 22:08:08 +00:00
|
|
|
self.assertEqual({
|
|
|
|
'image-pos': 0,
|
|
|
|
'offset': 0,
|
|
|
|
'size': 1890,
|
|
|
|
|
|
|
|
'u-boot:image-pos': 0,
|
|
|
|
'u-boot:offset': 0,
|
|
|
|
'u-boot:size': 4,
|
|
|
|
|
|
|
|
'fit:image-pos': 4,
|
|
|
|
'fit:offset': 4,
|
|
|
|
'fit:size': 1840,
|
|
|
|
|
2022-03-06 03:19:01 +00:00
|
|
|
'fit/images/kernel:image-pos': 304,
|
|
|
|
'fit/images/kernel:offset': 300,
|
2022-02-07 22:08:08 +00:00
|
|
|
'fit/images/kernel:size': 4,
|
|
|
|
|
2022-03-06 03:19:01 +00:00
|
|
|
'fit/images/kernel/u-boot:image-pos': 304,
|
2022-02-07 22:08:08 +00:00
|
|
|
'fit/images/kernel/u-boot:offset': 0,
|
|
|
|
'fit/images/kernel/u-boot:size': 4,
|
|
|
|
|
2022-03-06 03:19:01 +00:00
|
|
|
'fit/images/fdt-1:image-pos': 552,
|
|
|
|
'fit/images/fdt-1:offset': 548,
|
2022-02-07 22:08:08 +00:00
|
|
|
'fit/images/fdt-1:size': 6,
|
|
|
|
|
2022-03-06 03:19:01 +00:00
|
|
|
'fit/images/fdt-1/u-boot-spl-dtb:image-pos': 552,
|
2022-02-07 22:08:08 +00:00
|
|
|
'fit/images/fdt-1/u-boot-spl-dtb:offset': 0,
|
|
|
|
'fit/images/fdt-1/u-boot-spl-dtb:size': 6,
|
|
|
|
|
|
|
|
'u-boot-nodtb:image-pos': 1844,
|
|
|
|
'u-boot-nodtb:offset': 1844,
|
|
|
|
'u-boot-nodtb:size': 46,
|
|
|
|
}, props)
|
|
|
|
|
|
|
|
# Actually check the data is where we think it is
|
|
|
|
for node, expected in [
|
|
|
|
("u-boot", U_BOOT_DATA),
|
|
|
|
("fit/images/kernel", U_BOOT_DATA),
|
|
|
|
("fit/images/kernel/u-boot", U_BOOT_DATA),
|
|
|
|
("fit/images/fdt-1", U_BOOT_SPL_DTB_DATA),
|
|
|
|
("fit/images/fdt-1/u-boot-spl-dtb", U_BOOT_SPL_DTB_DATA),
|
|
|
|
("u-boot-nodtb", U_BOOT_NODTB_DATA),
|
|
|
|
]:
|
|
|
|
image_pos = props[f"{node}:image-pos"]
|
|
|
|
size = props[f"{node}:size"]
|
|
|
|
self.assertEqual(len(expected), size)
|
|
|
|
self.assertEqual(expected, data[image_pos:image_pos+size])
|
|
|
|
|
2020-07-10 00:39:45 +00:00
|
|
|
def testFitExternal(self):
|
2020-09-01 11:13:55 +00:00
|
|
|
"""Test an image with an FIT with external images"""
|
2020-07-10 00:39:45 +00:00
|
|
|
data = self._DoReadFile('162_fit_external.dts')
|
|
|
|
fit_data = data[len(U_BOOT_DATA):-2] # _testing is 2 bytes
|
|
|
|
|
2022-01-10 03:13:39 +00:00
|
|
|
# Size of the external-data region as set up by mkimage
|
|
|
|
external_data_size = len(U_BOOT_DATA) + 2
|
|
|
|
expected_size = (len(U_BOOT_DATA) + 0x400 +
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.align(external_data_size, 4) +
|
2022-01-10 03:13:39 +00:00
|
|
|
len(U_BOOT_NODTB_DATA))
|
|
|
|
|
2020-07-10 00:39:45 +00:00
|
|
|
# The data should be outside the FIT
|
|
|
|
dtb = fdt.Fdt.FromData(fit_data)
|
|
|
|
dtb.Scan()
|
|
|
|
fnode = dtb.GetNode('/images/kernel')
|
|
|
|
self.assertNotIn('data', fnode.props)
|
2022-01-10 03:13:39 +00:00
|
|
|
self.assertEqual(len(U_BOOT_DATA),
|
|
|
|
fdt_util.fdt32_to_cpu(fnode.props['data-size'].value))
|
|
|
|
fit_pos = 0x400;
|
|
|
|
self.assertEqual(
|
|
|
|
fit_pos,
|
|
|
|
fdt_util.fdt32_to_cpu(fnode.props['data-position'].value))
|
|
|
|
|
|
|
|
self.assertEquals(expected_size, len(data))
|
|
|
|
actual_pos = len(U_BOOT_DATA) + fit_pos
|
|
|
|
self.assertEqual(U_BOOT_DATA + b'aa',
|
|
|
|
data[actual_pos:actual_pos + external_data_size])
|
2019-07-20 18:23:51 +00:00
|
|
|
|
2022-02-07 22:08:08 +00:00
|
|
|
def testFitExternalImagePos(self):
|
|
|
|
"""Test that we have correct image-pos for external FIT subentries"""
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb('162_fit_external.dts',
|
|
|
|
update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS + REPACK_DTB_PROPS)
|
|
|
|
|
|
|
|
self.assertEqual({
|
|
|
|
'image-pos': 0,
|
|
|
|
'offset': 0,
|
|
|
|
'size': 1082,
|
|
|
|
|
|
|
|
'u-boot:image-pos': 0,
|
|
|
|
'u-boot:offset': 0,
|
|
|
|
'u-boot:size': 4,
|
|
|
|
|
|
|
|
'fit:size': 1032,
|
|
|
|
'fit:offset': 4,
|
|
|
|
'fit:image-pos': 4,
|
|
|
|
|
|
|
|
'fit/images/kernel:size': 4,
|
|
|
|
'fit/images/kernel:offset': 1024,
|
|
|
|
'fit/images/kernel:image-pos': 1028,
|
|
|
|
|
|
|
|
'fit/images/kernel/u-boot:size': 4,
|
|
|
|
'fit/images/kernel/u-boot:offset': 0,
|
|
|
|
'fit/images/kernel/u-boot:image-pos': 1028,
|
|
|
|
|
|
|
|
'fit/images/fdt-1:size': 2,
|
|
|
|
'fit/images/fdt-1:offset': 1028,
|
|
|
|
'fit/images/fdt-1:image-pos': 1032,
|
|
|
|
|
|
|
|
'fit/images/fdt-1/_testing:size': 2,
|
|
|
|
'fit/images/fdt-1/_testing:offset': 0,
|
|
|
|
'fit/images/fdt-1/_testing:image-pos': 1032,
|
|
|
|
|
|
|
|
'u-boot-nodtb:image-pos': 1036,
|
|
|
|
'u-boot-nodtb:offset': 1036,
|
|
|
|
'u-boot-nodtb:size': 46,
|
|
|
|
}, props)
|
|
|
|
|
|
|
|
# Actually check the data is where we think it is
|
|
|
|
for node, expected in [
|
|
|
|
("u-boot", U_BOOT_DATA),
|
|
|
|
("fit/images/kernel", U_BOOT_DATA),
|
|
|
|
("fit/images/kernel/u-boot", U_BOOT_DATA),
|
|
|
|
("fit/images/fdt-1", b'aa'),
|
|
|
|
("fit/images/fdt-1/_testing", b'aa'),
|
|
|
|
("u-boot-nodtb", U_BOOT_NODTB_DATA),
|
|
|
|
]:
|
|
|
|
image_pos = props[f"{node}:image-pos"]
|
|
|
|
size = props[f"{node}:size"]
|
|
|
|
self.assertEqual(len(expected), size)
|
|
|
|
self.assertEqual(expected, data[image_pos:image_pos+size])
|
|
|
|
|
2022-01-10 03:14:09 +00:00
|
|
|
def testFitMissing(self):
|
|
|
|
"""Test that binman still produces a FIT image if mkimage is missing"""
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoTestFile('162_fit_external.dts',
|
|
|
|
force_missing_bintools='mkimage')
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing bintools.*: mkimage")
|
|
|
|
|
2020-08-31 09:58:18 +00:00
|
|
|
def testSectionIgnoreHashSignature(self):
|
|
|
|
"""Test that sections ignore hash, signature nodes for its data"""
|
|
|
|
data = self._DoReadFile('165_section_ignore_hash_signature.dts')
|
|
|
|
expected = (U_BOOT_DATA + U_BOOT_DATA)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2020-08-31 09:58:19 +00:00
|
|
|
def testPadInSections(self):
|
|
|
|
"""Test pad-before, pad-after for entries in sections"""
|
2020-10-26 23:40:09 +00:00
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb(
|
|
|
|
'166_pad_in_sections.dts', update_dtb=True)
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (U_BOOT_DATA + tools.get_bytes(ord('!'), 12) +
|
|
|
|
U_BOOT_DATA + tools.get_bytes(ord('!'), 6) +
|
2020-08-31 09:58:19 +00:00
|
|
|
U_BOOT_DATA)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2020-10-26 23:40:09 +00:00
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['size', 'image-pos', 'offset'])
|
|
|
|
expected = {
|
|
|
|
'image-pos': 0,
|
|
|
|
'offset': 0,
|
|
|
|
'size': 12 + 6 + 3 * len(U_BOOT_DATA),
|
|
|
|
|
|
|
|
'section:image-pos': 0,
|
|
|
|
'section:offset': 0,
|
|
|
|
'section:size': 12 + 6 + 3 * len(U_BOOT_DATA),
|
|
|
|
|
|
|
|
'section/before:image-pos': 0,
|
|
|
|
'section/before:offset': 0,
|
|
|
|
'section/before:size': len(U_BOOT_DATA),
|
|
|
|
|
|
|
|
'section/u-boot:image-pos': 4,
|
|
|
|
'section/u-boot:offset': 4,
|
|
|
|
'section/u-boot:size': 12 + len(U_BOOT_DATA) + 6,
|
|
|
|
|
|
|
|
'section/after:image-pos': 26,
|
|
|
|
'section/after:offset': 26,
|
|
|
|
'section/after:size': len(U_BOOT_DATA),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
binman: Build FIT image subentries with the section etype
When reading subentries of each image, the FIT entry type directly
concatenates their contents without padding them according to their
offset, size, align, align-size, align-end, pad-before, pad-after
properties.
This patch makes sure these properties are respected by offloading this
image-data building to the section etype, where each subnode of the
"images" node is processed as a section. Alignments and offsets are
respective to the beginning of each image. For example, the following
fragment can end up having "u-boot-spl" start at 0x88 within the final
FIT binary, while "u-boot" would then end up starting at e.g. 0x20088.
fit {
description = "example";
images {
kernel-1 {
description = "U-Boot with SPL";
type = "kernel";
arch = "arm64";
os = "linux";
compression = "none";
u-boot-spl {
};
u-boot {
align = <0x10000>;
};
};
};
}
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
Reviewed-by: Simon Glass <sjg@chromium.org>
Reinstate check in testPadInSections(), squash in
"binman: Allow FIT binaries to have missing external blobs"
Signed-off-by: Simon Glass <sjg@chromium.org>
Reviewed-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2020-08-31 09:58:20 +00:00
|
|
|
def testFitImageSubentryAlignment(self):
|
|
|
|
"""Test relative alignability of FIT image subentries"""
|
2022-02-07 22:08:07 +00:00
|
|
|
self._SetupSplElf()
|
binman: Build FIT image subentries with the section etype
When reading subentries of each image, the FIT entry type directly
concatenates their contents without padding them according to their
offset, size, align, align-size, align-end, pad-before, pad-after
properties.
This patch makes sure these properties are respected by offloading this
image-data building to the section etype, where each subnode of the
"images" node is processed as a section. Alignments and offsets are
respective to the beginning of each image. For example, the following
fragment can end up having "u-boot-spl" start at 0x88 within the final
FIT binary, while "u-boot" would then end up starting at e.g. 0x20088.
fit {
description = "example";
images {
kernel-1 {
description = "U-Boot with SPL";
type = "kernel";
arch = "arm64";
os = "linux";
compression = "none";
u-boot-spl {
};
u-boot {
align = <0x10000>;
};
};
};
}
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
Reviewed-by: Simon Glass <sjg@chromium.org>
Reinstate check in testPadInSections(), squash in
"binman: Allow FIT binaries to have missing external blobs"
Signed-off-by: Simon Glass <sjg@chromium.org>
Reviewed-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2020-08-31 09:58:20 +00:00
|
|
|
entry_args = {
|
|
|
|
'test-id': TEXT_DATA,
|
|
|
|
}
|
|
|
|
data, _, _, _ = self._DoReadFileDtb('167_fit_image_subentry_alignment.dts',
|
|
|
|
entry_args=entry_args)
|
|
|
|
dtb = fdt.Fdt.FromData(data)
|
|
|
|
dtb.Scan()
|
|
|
|
|
|
|
|
node = dtb.GetNode('/images/kernel')
|
|
|
|
data = dtb.GetProps(node)["data"].bytes
|
|
|
|
align_pad = 0x10 - (len(U_BOOT_SPL_DATA) % 0x10)
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (tools.get_bytes(0, 0x20) + U_BOOT_SPL_DATA +
|
|
|
|
tools.get_bytes(0, align_pad) + U_BOOT_DATA)
|
binman: Build FIT image subentries with the section etype
When reading subentries of each image, the FIT entry type directly
concatenates their contents without padding them according to their
offset, size, align, align-size, align-end, pad-before, pad-after
properties.
This patch makes sure these properties are respected by offloading this
image-data building to the section etype, where each subnode of the
"images" node is processed as a section. Alignments and offsets are
respective to the beginning of each image. For example, the following
fragment can end up having "u-boot-spl" start at 0x88 within the final
FIT binary, while "u-boot" would then end up starting at e.g. 0x20088.
fit {
description = "example";
images {
kernel-1 {
description = "U-Boot with SPL";
type = "kernel";
arch = "arm64";
os = "linux";
compression = "none";
u-boot-spl {
};
u-boot {
align = <0x10000>;
};
};
};
}
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
Reviewed-by: Simon Glass <sjg@chromium.org>
Reinstate check in testPadInSections(), squash in
"binman: Allow FIT binaries to have missing external blobs"
Signed-off-by: Simon Glass <sjg@chromium.org>
Reviewed-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2020-08-31 09:58:20 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
node = dtb.GetNode('/images/fdt-1')
|
|
|
|
data = dtb.GetProps(node)["data"].bytes
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (U_BOOT_SPL_DTB_DATA + tools.get_bytes(0, 20) +
|
|
|
|
tools.to_bytes(TEXT_DATA) + tools.get_bytes(0, 30) +
|
binman: Build FIT image subentries with the section etype
When reading subentries of each image, the FIT entry type directly
concatenates their contents without padding them according to their
offset, size, align, align-size, align-end, pad-before, pad-after
properties.
This patch makes sure these properties are respected by offloading this
image-data building to the section etype, where each subnode of the
"images" node is processed as a section. Alignments and offsets are
respective to the beginning of each image. For example, the following
fragment can end up having "u-boot-spl" start at 0x88 within the final
FIT binary, while "u-boot" would then end up starting at e.g. 0x20088.
fit {
description = "example";
images {
kernel-1 {
description = "U-Boot with SPL";
type = "kernel";
arch = "arm64";
os = "linux";
compression = "none";
u-boot-spl {
};
u-boot {
align = <0x10000>;
};
};
};
}
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
Reviewed-by: Simon Glass <sjg@chromium.org>
Reinstate check in testPadInSections(), squash in
"binman: Allow FIT binaries to have missing external blobs"
Signed-off-by: Simon Glass <sjg@chromium.org>
Reviewed-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2020-08-31 09:58:20 +00:00
|
|
|
U_BOOT_DTB_DATA)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testFitExtblobMissingOk(self):
|
|
|
|
"""Test a FIT with a missing external blob that is allowed"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('168_fit_missing_blob.dts',
|
|
|
|
allow_missing=True)
|
|
|
|
err = stderr.getvalue()
|
2020-09-06 16:39:09 +00:00
|
|
|
self.assertRegex(err, "Image 'main-section'.*missing.*: atf-bl31")
|
binman: Build FIT image subentries with the section etype
When reading subentries of each image, the FIT entry type directly
concatenates their contents without padding them according to their
offset, size, align, align-size, align-end, pad-before, pad-after
properties.
This patch makes sure these properties are respected by offloading this
image-data building to the section etype, where each subnode of the
"images" node is processed as a section. Alignments and offsets are
respective to the beginning of each image. For example, the following
fragment can end up having "u-boot-spl" start at 0x88 within the final
FIT binary, while "u-boot" would then end up starting at e.g. 0x20088.
fit {
description = "example";
images {
kernel-1 {
description = "U-Boot with SPL";
type = "kernel";
arch = "arm64";
os = "linux";
compression = "none";
u-boot-spl {
};
u-boot {
align = <0x10000>;
};
};
};
}
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
Reviewed-by: Simon Glass <sjg@chromium.org>
Reinstate check in testPadInSections(), squash in
"binman: Allow FIT binaries to have missing external blobs"
Signed-off-by: Simon Glass <sjg@chromium.org>
Reviewed-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2020-08-31 09:58:20 +00:00
|
|
|
|
2020-09-01 11:13:54 +00:00
|
|
|
def testBlobNamedByArgMissing(self):
|
|
|
|
"""Test handling of a missing entry arg"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('068_blob_named_by_arg.dts')
|
|
|
|
self.assertIn("Missing required properties/entry args: cros-ec-rw-path",
|
|
|
|
str(e.exception))
|
|
|
|
|
2020-09-01 11:13:58 +00:00
|
|
|
def testPackBl31(self):
|
|
|
|
"""Test that an image with an ATF BL31 binary can be created"""
|
|
|
|
data = self._DoReadFile('169_atf_bl31.dts')
|
|
|
|
self.assertEqual(ATF_BL31_DATA, data[:len(ATF_BL31_DATA)])
|
|
|
|
|
2020-10-22 02:12:15 +00:00
|
|
|
def testPackScp(self):
|
|
|
|
"""Test that an image with an SCP binary can be created"""
|
|
|
|
data = self._DoReadFile('172_scp.dts')
|
|
|
|
self.assertEqual(SCP_DATA, data[:len(SCP_DATA)])
|
|
|
|
|
2020-09-01 11:13:59 +00:00
|
|
|
def testFitFdt(self):
|
|
|
|
"""Test an image with an FIT with multiple FDT images"""
|
|
|
|
def _CheckFdt(seq, expected_data):
|
|
|
|
"""Check the FDT nodes
|
|
|
|
|
|
|
|
Args:
|
|
|
|
seq: Sequence number to check (0 or 1)
|
|
|
|
expected_data: Expected contents of 'data' property
|
|
|
|
"""
|
|
|
|
name = 'fdt-%d' % seq
|
|
|
|
fnode = dtb.GetNode('/images/%s' % name)
|
|
|
|
self.assertIsNotNone(fnode)
|
|
|
|
self.assertEqual({'description','type', 'compression', 'data'},
|
|
|
|
set(fnode.props.keys()))
|
|
|
|
self.assertEqual(expected_data, fnode.props['data'].bytes)
|
|
|
|
self.assertEqual('fdt-test-fdt%d.dtb' % seq,
|
|
|
|
fnode.props['description'].value)
|
2022-02-28 16:06:20 +00:00
|
|
|
self.assertEqual(fnode.subnodes[0].name, 'hash')
|
2020-09-01 11:13:59 +00:00
|
|
|
|
|
|
|
def _CheckConfig(seq, expected_data):
|
|
|
|
"""Check the configuration nodes
|
|
|
|
|
|
|
|
Args:
|
|
|
|
seq: Sequence number to check (0 or 1)
|
|
|
|
expected_data: Expected contents of 'data' property
|
|
|
|
"""
|
|
|
|
cnode = dtb.GetNode('/configurations')
|
|
|
|
self.assertIn('default', cnode.props)
|
2020-09-06 16:39:08 +00:00
|
|
|
self.assertEqual('config-2', cnode.props['default'].value)
|
2020-09-01 11:13:59 +00:00
|
|
|
|
|
|
|
name = 'config-%d' % seq
|
|
|
|
fnode = dtb.GetNode('/configurations/%s' % name)
|
|
|
|
self.assertIsNotNone(fnode)
|
|
|
|
self.assertEqual({'description','firmware', 'loadables', 'fdt'},
|
|
|
|
set(fnode.props.keys()))
|
|
|
|
self.assertEqual('conf-test-fdt%d.dtb' % seq,
|
|
|
|
fnode.props['description'].value)
|
|
|
|
self.assertEqual('fdt-%d' % seq, fnode.props['fdt'].value)
|
|
|
|
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
2020-09-06 16:39:08 +00:00
|
|
|
'default-dt': 'test-fdt2',
|
2020-09-01 11:13:59 +00:00
|
|
|
}
|
|
|
|
data = self._DoReadFileDtb(
|
2021-05-10 12:23:32 +00:00
|
|
|
'170_fit_fdt.dts',
|
2020-09-01 11:13:59 +00:00
|
|
|
entry_args=entry_args,
|
|
|
|
extra_indirs=[os.path.join(self._indir, TEST_FDT_SUBDIR)])[0]
|
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA, data[-len(U_BOOT_NODTB_DATA):])
|
|
|
|
fit_data = data[len(U_BOOT_DATA):-len(U_BOOT_NODTB_DATA)]
|
|
|
|
|
|
|
|
dtb = fdt.Fdt.FromData(fit_data)
|
|
|
|
dtb.Scan()
|
|
|
|
fnode = dtb.GetNode('/images/kernel')
|
|
|
|
self.assertIn('data', fnode.props)
|
|
|
|
|
|
|
|
# Check all the properties in fdt-1 and fdt-2
|
|
|
|
_CheckFdt(1, TEST_FDT1_DATA)
|
|
|
|
_CheckFdt(2, TEST_FDT2_DATA)
|
|
|
|
|
|
|
|
# Check configurations
|
|
|
|
_CheckConfig(1, TEST_FDT1_DATA)
|
|
|
|
_CheckConfig(2, TEST_FDT2_DATA)
|
|
|
|
|
|
|
|
def testFitFdtMissingList(self):
|
|
|
|
"""Test handling of a missing 'of-list' entry arg"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2021-05-10 12:23:32 +00:00
|
|
|
self._DoReadFile('170_fit_fdt.dts')
|
2020-09-01 11:13:59 +00:00
|
|
|
self.assertIn("Generator node requires 'of-list' entry argument",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testFitFdtEmptyList(self):
|
|
|
|
"""Test handling of an empty 'of-list' entry arg"""
|
|
|
|
entry_args = {
|
|
|
|
'of-list': '',
|
|
|
|
}
|
|
|
|
data = self._DoReadFileDtb('170_fit_fdt.dts', entry_args=entry_args)[0]
|
|
|
|
|
|
|
|
def testFitFdtMissingProp(self):
|
|
|
|
"""Test handling of a missing 'fit,fdt-list' property"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('171_fit_fdt_missing_prop.dts')
|
|
|
|
self.assertIn("Generator node requires 'fit,fdt-list' property",
|
|
|
|
str(e.exception))
|
2020-09-01 11:13:58 +00:00
|
|
|
|
2020-09-06 16:39:08 +00:00
|
|
|
def testFitFdtMissing(self):
|
|
|
|
"""Test handling of a missing 'default-dt' entry arg"""
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
}
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFileDtb(
|
2021-05-10 12:23:32 +00:00
|
|
|
'170_fit_fdt.dts',
|
2020-09-06 16:39:08 +00:00
|
|
|
entry_args=entry_args,
|
|
|
|
extra_indirs=[os.path.join(self._indir, TEST_FDT_SUBDIR)])[0]
|
|
|
|
self.assertIn("Generated 'default' node requires default-dt entry argument",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testFitFdtNotInList(self):
|
|
|
|
"""Test handling of a default-dt that is not in the of-list"""
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
'default-dt': 'test-fdt3',
|
|
|
|
}
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFileDtb(
|
2021-05-10 12:23:32 +00:00
|
|
|
'170_fit_fdt.dts',
|
2020-09-06 16:39:08 +00:00
|
|
|
entry_args=entry_args,
|
|
|
|
extra_indirs=[os.path.join(self._indir, TEST_FDT_SUBDIR)])[0]
|
|
|
|
self.assertIn("default-dt entry argument 'test-fdt3' not found in fdt list: test-fdt1, test-fdt2",
|
|
|
|
str(e.exception))
|
|
|
|
|
2020-09-06 16:39:09 +00:00
|
|
|
def testFitExtblobMissingHelp(self):
|
|
|
|
"""Test display of help messages when an external blob is missing"""
|
|
|
|
control.missing_blob_help = control._ReadMissingBlobHelp()
|
|
|
|
control.missing_blob_help['wibble'] = 'Wibble test'
|
|
|
|
control.missing_blob_help['another'] = 'Another test'
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('168_fit_missing_blob.dts',
|
|
|
|
allow_missing=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
|
|
|
|
# We can get the tag from the name, the type or the missing-msg
|
|
|
|
# property. Check all three.
|
|
|
|
self.assertIn('You may need to build ARM Trusted', err)
|
|
|
|
self.assertIn('Wibble test', err)
|
|
|
|
self.assertIn('Another test', err)
|
|
|
|
|
2020-09-06 16:35:32 +00:00
|
|
|
def testMissingBlob(self):
|
|
|
|
"""Test handling of a blob containing a missing file"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoTestFile('173_missing_blob.dts', allow_missing=True)
|
|
|
|
self.assertIn("Filename 'missing' not found in input path",
|
|
|
|
str(e.exception))
|
|
|
|
|
2020-09-06 16:35:33 +00:00
|
|
|
def testEnvironment(self):
|
|
|
|
"""Test adding a U-Boot environment"""
|
|
|
|
data = self._DoReadFile('174_env.dts')
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[:len(U_BOOT_DATA)])
|
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA, data[-len(U_BOOT_NODTB_DATA):])
|
|
|
|
env = data[len(U_BOOT_DATA):-len(U_BOOT_NODTB_DATA)]
|
|
|
|
self.assertEqual(b'\x1b\x97\x22\x7c\x01var1=1\0var2="2"\0\0\xff\xff',
|
|
|
|
env)
|
|
|
|
|
|
|
|
def testEnvironmentNoSize(self):
|
|
|
|
"""Test that a missing 'size' property is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2020-10-26 23:40:00 +00:00
|
|
|
self._DoTestFile('175_env_no_size.dts')
|
2020-09-06 16:35:33 +00:00
|
|
|
self.assertIn("'u-boot-env' entry must have a size property",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testEnvironmentTooSmall(self):
|
|
|
|
"""Test handling of an environment that does not fit"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2020-10-26 23:40:00 +00:00
|
|
|
self._DoTestFile('176_env_too_small.dts')
|
2020-09-06 16:35:33 +00:00
|
|
|
|
|
|
|
# checksum, start byte, environment with \0 terminator, final \0
|
|
|
|
need = 4 + 1 + len(ENV_DATA) + 1 + 1
|
|
|
|
short = need - 0x8
|
|
|
|
self.assertIn("too small to hold data (need %#x more bytes)" % short,
|
|
|
|
str(e.exception))
|
|
|
|
|
2020-10-26 23:40:01 +00:00
|
|
|
def testSkipAtStart(self):
|
|
|
|
"""Test handling of skip-at-start section"""
|
|
|
|
data = self._DoReadFile('177_skip_at_start.dts')
|
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
section = entries['section']
|
|
|
|
self.assertEqual(0, section.offset)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), section.size)
|
|
|
|
self.assertEqual(U_BOOT_DATA, section.GetData())
|
|
|
|
|
|
|
|
entry = section.GetEntries()['u-boot']
|
|
|
|
self.assertEqual(16, entry.offset)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
|
|
|
|
|
|
|
def testSkipAtStartPad(self):
|
|
|
|
"""Test handling of skip-at-start section with padded entry"""
|
|
|
|
data = self._DoReadFile('178_skip_at_start_pad.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
before = tools.get_bytes(0, 8)
|
|
|
|
after = tools.get_bytes(0, 4)
|
2020-10-26 23:40:01 +00:00
|
|
|
all = before + U_BOOT_DATA + after
|
|
|
|
self.assertEqual(all, data)
|
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
section = entries['section']
|
|
|
|
self.assertEqual(0, section.offset)
|
|
|
|
self.assertEqual(len(all), section.size)
|
|
|
|
self.assertEqual(all, section.GetData())
|
|
|
|
|
|
|
|
entry = section.GetEntries()['u-boot']
|
|
|
|
self.assertEqual(16, entry.offset)
|
|
|
|
self.assertEqual(len(all), entry.size)
|
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
|
|
|
|
|
|
|
def testSkipAtStartSectionPad(self):
|
|
|
|
"""Test handling of skip-at-start section with padding"""
|
|
|
|
data = self._DoReadFile('179_skip_at_start_section_pad.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
before = tools.get_bytes(0, 8)
|
|
|
|
after = tools.get_bytes(0, 4)
|
2020-10-26 23:40:01 +00:00
|
|
|
all = before + U_BOOT_DATA + after
|
2020-10-26 23:40:13 +00:00
|
|
|
self.assertEqual(all, data)
|
2020-10-26 23:40:01 +00:00
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
section = entries['section']
|
|
|
|
self.assertEqual(0, section.offset)
|
|
|
|
self.assertEqual(len(all), section.size)
|
2020-10-26 23:40:16 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, section.data)
|
2020-10-26 23:40:13 +00:00
|
|
|
self.assertEqual(all, section.GetPaddedData())
|
2020-10-26 23:40:01 +00:00
|
|
|
|
|
|
|
entry = section.GetEntries()['u-boot']
|
|
|
|
self.assertEqual(16, entry.offset)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
2020-09-06 16:35:33 +00:00
|
|
|
|
2020-10-26 23:40:14 +00:00
|
|
|
def testSectionPad(self):
|
|
|
|
"""Testing padding with sections"""
|
|
|
|
data = self._DoReadFile('180_section_pad.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (tools.get_bytes(ord('&'), 3) +
|
|
|
|
tools.get_bytes(ord('!'), 5) +
|
2020-10-26 23:40:14 +00:00
|
|
|
U_BOOT_DATA +
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.get_bytes(ord('!'), 1) +
|
|
|
|
tools.get_bytes(ord('&'), 2))
|
2020-10-26 23:40:14 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testSectionAlign(self):
|
|
|
|
"""Testing alignment with sections"""
|
|
|
|
data = self._DoReadFileDtb('181_section_align.dts', map=True)[0]
|
|
|
|
expected = (b'\0' + # fill section
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.get_bytes(ord('&'), 1) + # padding to section align
|
2020-10-26 23:40:14 +00:00
|
|
|
b'\0' + # fill section
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.get_bytes(ord('!'), 3) + # padding to u-boot align
|
2020-10-26 23:40:14 +00:00
|
|
|
U_BOOT_DATA +
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.get_bytes(ord('!'), 4) + # padding to u-boot size
|
|
|
|
tools.get_bytes(ord('!'), 4)) # padding to section size
|
2020-10-26 23:40:14 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2020-10-26 23:40:25 +00:00
|
|
|
def testCompressImage(self):
|
|
|
|
"""Test compression of the entire image"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb(
|
|
|
|
'182_compress_image.dts', use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['offset', 'image-pos', 'size',
|
|
|
|
'uncomp-size'])
|
|
|
|
orig = self._decompress(data)
|
|
|
|
self.assertEquals(COMPRESS_DATA + U_BOOT_DATA, orig)
|
|
|
|
|
|
|
|
# Do a sanity check on various fields
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(2, len(entries))
|
|
|
|
|
|
|
|
entry = entries['blob']
|
|
|
|
self.assertEqual(COMPRESS_DATA, entry.data)
|
|
|
|
self.assertEqual(len(COMPRESS_DATA), entry.size)
|
|
|
|
|
|
|
|
entry = entries['u-boot']
|
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
|
|
|
self.assertEqual(len(data), image.size)
|
|
|
|
self.assertEqual(COMPRESS_DATA + U_BOOT_DATA, image.uncomp_data)
|
|
|
|
self.assertEqual(len(COMPRESS_DATA + U_BOOT_DATA), image.uncomp_size)
|
|
|
|
orig = self._decompress(image.data)
|
|
|
|
self.assertEqual(orig, image.uncomp_data)
|
|
|
|
|
|
|
|
expected = {
|
|
|
|
'blob:offset': 0,
|
|
|
|
'blob:size': len(COMPRESS_DATA),
|
|
|
|
'u-boot:offset': len(COMPRESS_DATA),
|
|
|
|
'u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'uncomp-size': len(COMPRESS_DATA + U_BOOT_DATA),
|
|
|
|
'offset': 0,
|
|
|
|
'image-pos': 0,
|
|
|
|
'size': len(data),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
|
|
|
def testCompressImageLess(self):
|
|
|
|
"""Test compression where compression reduces the image size"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb(
|
|
|
|
'183_compress_image_less.dts', use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['offset', 'image-pos', 'size',
|
|
|
|
'uncomp-size'])
|
|
|
|
orig = self._decompress(data)
|
|
|
|
|
|
|
|
self.assertEquals(COMPRESS_DATA + COMPRESS_DATA + U_BOOT_DATA, orig)
|
|
|
|
|
|
|
|
# Do a sanity check on various fields
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(2, len(entries))
|
|
|
|
|
|
|
|
entry = entries['blob']
|
|
|
|
self.assertEqual(COMPRESS_DATA_BIG, entry.data)
|
|
|
|
self.assertEqual(len(COMPRESS_DATA_BIG), entry.size)
|
|
|
|
|
|
|
|
entry = entries['u-boot']
|
|
|
|
self.assertEqual(U_BOOT_DATA, entry.data)
|
|
|
|
self.assertEqual(len(U_BOOT_DATA), entry.size)
|
|
|
|
|
|
|
|
self.assertEqual(len(data), image.size)
|
|
|
|
self.assertEqual(COMPRESS_DATA_BIG + U_BOOT_DATA, image.uncomp_data)
|
|
|
|
self.assertEqual(len(COMPRESS_DATA_BIG + U_BOOT_DATA),
|
|
|
|
image.uncomp_size)
|
|
|
|
orig = self._decompress(image.data)
|
|
|
|
self.assertEqual(orig, image.uncomp_data)
|
|
|
|
|
|
|
|
expected = {
|
|
|
|
'blob:offset': 0,
|
|
|
|
'blob:size': len(COMPRESS_DATA_BIG),
|
|
|
|
'u-boot:offset': len(COMPRESS_DATA_BIG),
|
|
|
|
'u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'uncomp-size': len(COMPRESS_DATA_BIG + U_BOOT_DATA),
|
|
|
|
'offset': 0,
|
|
|
|
'image-pos': 0,
|
|
|
|
'size': len(data),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
|
|
|
def testCompressSectionSize(self):
|
|
|
|
"""Test compression of a section with a fixed size"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb(
|
|
|
|
'184_compress_section_size.dts', use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['offset', 'image-pos', 'size',
|
|
|
|
'uncomp-size'])
|
|
|
|
orig = self._decompress(data)
|
|
|
|
self.assertEquals(COMPRESS_DATA + U_BOOT_DATA, orig)
|
|
|
|
expected = {
|
|
|
|
'section/blob:offset': 0,
|
|
|
|
'section/blob:size': len(COMPRESS_DATA),
|
|
|
|
'section/u-boot:offset': len(COMPRESS_DATA),
|
|
|
|
'section/u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'section:offset': 0,
|
|
|
|
'section:image-pos': 0,
|
|
|
|
'section:uncomp-size': len(COMPRESS_DATA + U_BOOT_DATA),
|
|
|
|
'section:size': 0x30,
|
|
|
|
'offset': 0,
|
|
|
|
'image-pos': 0,
|
|
|
|
'size': 0x30,
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
|
|
|
def testCompressSection(self):
|
|
|
|
"""Test compression of a section with no fixed size"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb(
|
|
|
|
'185_compress_section.dts', use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['offset', 'image-pos', 'size',
|
|
|
|
'uncomp-size'])
|
|
|
|
orig = self._decompress(data)
|
|
|
|
self.assertEquals(COMPRESS_DATA + U_BOOT_DATA, orig)
|
|
|
|
expected = {
|
|
|
|
'section/blob:offset': 0,
|
|
|
|
'section/blob:size': len(COMPRESS_DATA),
|
|
|
|
'section/u-boot:offset': len(COMPRESS_DATA),
|
|
|
|
'section/u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'section:offset': 0,
|
|
|
|
'section:image-pos': 0,
|
|
|
|
'section:uncomp-size': len(COMPRESS_DATA + U_BOOT_DATA),
|
|
|
|
'section:size': len(data),
|
|
|
|
'offset': 0,
|
|
|
|
'image-pos': 0,
|
|
|
|
'size': len(data),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
2022-08-19 14:25:31 +00:00
|
|
|
def testLz4Missing(self):
|
|
|
|
"""Test that binman still produces an image if lz4 is missing"""
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoTestFile('185_compress_section.dts',
|
|
|
|
force_missing_bintools='lz4')
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing bintools.*: lz4")
|
|
|
|
|
2020-10-26 23:40:25 +00:00
|
|
|
def testCompressExtra(self):
|
|
|
|
"""Test compression of a section with no fixed size"""
|
|
|
|
self._CheckLz4()
|
|
|
|
data, _, _, out_dtb_fname = self._DoReadFileDtb(
|
|
|
|
'186_compress_extra.dts', use_real_dtb=True, update_dtb=True)
|
|
|
|
dtb = fdt.Fdt(out_dtb_fname)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['offset', 'image-pos', 'size',
|
|
|
|
'uncomp-size'])
|
|
|
|
|
|
|
|
base = data[len(U_BOOT_DATA):]
|
|
|
|
self.assertEquals(U_BOOT_DATA, base[:len(U_BOOT_DATA)])
|
|
|
|
rest = base[len(U_BOOT_DATA):]
|
|
|
|
|
|
|
|
# Check compressed data
|
2022-08-19 14:25:28 +00:00
|
|
|
bintool = self.comp_bintools['lz4']
|
|
|
|
expect1 = bintool.compress(COMPRESS_DATA + U_BOOT_DATA)
|
2022-08-19 14:25:23 +00:00
|
|
|
data1 = rest[:len(expect1)]
|
|
|
|
section1 = self._decompress(data1)
|
|
|
|
self.assertEquals(expect1, data1)
|
2020-10-26 23:40:25 +00:00
|
|
|
self.assertEquals(COMPRESS_DATA + U_BOOT_DATA, section1)
|
|
|
|
rest1 = rest[len(expect1):]
|
|
|
|
|
2022-08-19 14:25:28 +00:00
|
|
|
expect2 = bintool.compress(COMPRESS_DATA + COMPRESS_DATA)
|
2022-08-19 14:25:23 +00:00
|
|
|
data2 = rest1[:len(expect2)]
|
|
|
|
section2 = self._decompress(data2)
|
|
|
|
self.assertEquals(expect2, data2)
|
2020-10-26 23:40:25 +00:00
|
|
|
self.assertEquals(COMPRESS_DATA + COMPRESS_DATA, section2)
|
|
|
|
rest2 = rest1[len(expect2):]
|
|
|
|
|
|
|
|
expect_size = (len(U_BOOT_DATA) + len(U_BOOT_DATA) + len(expect1) +
|
|
|
|
len(expect2) + len(U_BOOT_DATA))
|
|
|
|
#self.assertEquals(expect_size, len(data))
|
|
|
|
|
|
|
|
#self.assertEquals(U_BOOT_DATA, rest2)
|
|
|
|
|
|
|
|
self.maxDiff = None
|
|
|
|
expected = {
|
|
|
|
'u-boot:offset': 0,
|
|
|
|
'u-boot:image-pos': 0,
|
|
|
|
'u-boot:size': len(U_BOOT_DATA),
|
|
|
|
|
|
|
|
'base:offset': len(U_BOOT_DATA),
|
|
|
|
'base:image-pos': len(U_BOOT_DATA),
|
|
|
|
'base:size': len(data) - len(U_BOOT_DATA),
|
|
|
|
'base/u-boot:offset': 0,
|
|
|
|
'base/u-boot:image-pos': len(U_BOOT_DATA),
|
|
|
|
'base/u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'base/u-boot2:offset': len(U_BOOT_DATA) + len(expect1) +
|
|
|
|
len(expect2),
|
|
|
|
'base/u-boot2:image-pos': len(U_BOOT_DATA) * 2 + len(expect1) +
|
|
|
|
len(expect2),
|
|
|
|
'base/u-boot2:size': len(U_BOOT_DATA),
|
|
|
|
|
|
|
|
'base/section:offset': len(U_BOOT_DATA),
|
|
|
|
'base/section:image-pos': len(U_BOOT_DATA) * 2,
|
|
|
|
'base/section:size': len(expect1),
|
|
|
|
'base/section:uncomp-size': len(COMPRESS_DATA + U_BOOT_DATA),
|
|
|
|
'base/section/blob:offset': 0,
|
|
|
|
'base/section/blob:size': len(COMPRESS_DATA),
|
|
|
|
'base/section/u-boot:offset': len(COMPRESS_DATA),
|
|
|
|
'base/section/u-boot:size': len(U_BOOT_DATA),
|
|
|
|
|
|
|
|
'base/section2:offset': len(U_BOOT_DATA) + len(expect1),
|
|
|
|
'base/section2:image-pos': len(U_BOOT_DATA) * 2 + len(expect1),
|
|
|
|
'base/section2:size': len(expect2),
|
|
|
|
'base/section2:uncomp-size': len(COMPRESS_DATA + COMPRESS_DATA),
|
|
|
|
'base/section2/blob:offset': 0,
|
|
|
|
'base/section2/blob:size': len(COMPRESS_DATA),
|
|
|
|
'base/section2/blob2:offset': len(COMPRESS_DATA),
|
|
|
|
'base/section2/blob2:size': len(COMPRESS_DATA),
|
|
|
|
|
|
|
|
'offset': 0,
|
|
|
|
'image-pos': 0,
|
|
|
|
'size': len(data),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
2021-01-07 04:35:15 +00:00
|
|
|
def testSymbolsSubsection(self):
|
|
|
|
"""Test binman can assign symbols from a subsection"""
|
spl: binman: Check at runtime if binman symbols were filled in
Binman lets us declare symbols in SPL/TPL that refer to other entries in
the same binman image as them. These symbols are filled in with the
correct values while binman assembles the images, but this is done
in-memory only. Symbols marked as optional can be filled with
BINMAN_SYM_MISSING as an error value if their referred entry is missing.
However, the unmodified SPL/TPL binaries are still available on disk,
and can be used by people. For these files, nothing ensures that the
symbols are set to this error value, and they will be considered valid
when they are not.
Empirically, all symbols show up as zero in a sandbox_vpl build when we
run e.g. tpl/u-boot-tpl directly. On the other hand, zero is a perfectly
fine value for a binman-written symbol, so we cannot say the symbols
have wrong values based on that.
Declare a magic symbol that binman always fills in with a fixed value.
Check this value as an indicator that symbols were filled in correctly.
Return the error value for all symbols when this magic symbol has the
wrong value.
For binman tests, we need to make room for the new symbol in the mocked
SPL/TPL data by extending them by four bytes. This messes up some test
image layouts. Fix the affected values, and check the magic symbol
wherever it makes sense.
Signed-off-by: Alper Nebi Yasak <alpernebiyasak@gmail.com>
2022-06-18 12:13:11 +00:00
|
|
|
self.checkSymbols('187_symbols_sub.dts', U_BOOT_SPL_DATA, 0x1c)
|
2021-01-07 04:35:15 +00:00
|
|
|
|
2021-01-07 04:35:16 +00:00
|
|
|
def testReadImageEntryArg(self):
|
|
|
|
"""Test reading an image that would need an entry arg to generate"""
|
|
|
|
entry_args = {
|
|
|
|
'cros-ec-rw-path': 'ecrw.bin',
|
|
|
|
}
|
|
|
|
data = self.data = self._DoReadFileDtb(
|
|
|
|
'188_image_entryarg.dts',use_real_dtb=True, update_dtb=True,
|
|
|
|
entry_args=entry_args)
|
|
|
|
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2021-01-07 04:35:16 +00:00
|
|
|
orig_image = control.images['image']
|
|
|
|
|
|
|
|
# This should not generate an error about the missing 'cros-ec-rw-path'
|
|
|
|
# since we are reading the image from a file. Compare with
|
|
|
|
# testEntryArgsRequired()
|
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
self.assertEqual(orig_image.GetEntries().keys(),
|
|
|
|
image.GetEntries().keys())
|
|
|
|
|
2021-01-07 04:35:18 +00:00
|
|
|
def testFilesAlign(self):
|
|
|
|
"""Test alignment with files"""
|
|
|
|
data = self._DoReadFile('190_files_align.dts')
|
|
|
|
|
|
|
|
# The first string is 15 bytes so will align to 16
|
|
|
|
expect = FILES_DATA[:15] + b'\0' + FILES_DATA[15:]
|
|
|
|
self.assertEqual(expect, data)
|
|
|
|
|
2021-01-07 04:35:19 +00:00
|
|
|
def testReadImageSkip(self):
|
|
|
|
"""Test reading an image and accessing its FDT map"""
|
|
|
|
data = self.data = self._DoReadFileRealDtb('191_read_image_skip.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2021-01-07 04:35:19 +00:00
|
|
|
orig_image = control.images['image']
|
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
self.assertEqual(orig_image.GetEntries().keys(),
|
|
|
|
image.GetEntries().keys())
|
|
|
|
|
|
|
|
orig_entry = orig_image.GetEntries()['fdtmap']
|
|
|
|
entry = image.GetEntries()['fdtmap']
|
|
|
|
self.assertEqual(orig_entry.offset, entry.offset)
|
|
|
|
self.assertEqual(orig_entry.size, entry.size)
|
|
|
|
self.assertEqual(16, entry.image_pos)
|
|
|
|
|
|
|
|
u_boot = image.GetEntries()['section'].GetEntries()['u-boot']
|
|
|
|
|
|
|
|
self.assertEquals(U_BOOT_DATA, u_boot.ReadData())
|
|
|
|
|
2021-03-18 07:24:57 +00:00
|
|
|
def testTplNoDtb(self):
|
|
|
|
"""Test that an image with tpl/u-boot-tpl-nodtb.bin can be created"""
|
2021-04-24 20:39:32 +00:00
|
|
|
self._SetupTplElf()
|
2021-03-18 07:24:57 +00:00
|
|
|
data = self._DoReadFile('192_u_boot_tpl_nodtb.dts')
|
|
|
|
self.assertEqual(U_BOOT_TPL_NODTB_DATA,
|
|
|
|
data[:len(U_BOOT_TPL_NODTB_DATA)])
|
|
|
|
|
2021-03-18 07:24:58 +00:00
|
|
|
def testTplBssPad(self):
|
|
|
|
"""Test that we can pad TPL's BSS with zeros"""
|
|
|
|
# ELF file with a '__bss_size' symbol
|
|
|
|
self._SetupTplElf()
|
|
|
|
data = self._DoReadFile('193_tpl_bss_pad.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(U_BOOT_TPL_DATA + tools.get_bytes(0, 10) + U_BOOT_DATA,
|
2021-03-18 07:24:58 +00:00
|
|
|
data)
|
|
|
|
|
|
|
|
def testTplBssPadMissing(self):
|
|
|
|
"""Test that a missing symbol is detected"""
|
|
|
|
self._SetupTplElf('u_boot_ucode_ptr')
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('193_tpl_bss_pad.dts')
|
|
|
|
self.assertIn('Expected __bss_size symbol in tpl/u-boot-tpl',
|
|
|
|
str(e.exception))
|
|
|
|
|
2021-03-18 07:25:07 +00:00
|
|
|
def checkDtbSizes(self, data, pad_len, start):
|
|
|
|
"""Check the size arguments in a dtb embedded in an image
|
|
|
|
|
|
|
|
Args:
|
|
|
|
data: The image data
|
|
|
|
pad_len: Length of the pad section in the image, in bytes
|
|
|
|
start: Start offset of the devicetree to examine, within the image
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Size of the devicetree in bytes
|
|
|
|
"""
|
|
|
|
dtb_data = data[start:]
|
|
|
|
dtb = fdt.Fdt.FromData(dtb_data)
|
|
|
|
fdt_size = dtb.GetFdtObj().totalsize()
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, 'size')
|
|
|
|
self.assertEqual({
|
|
|
|
'size': len(data),
|
|
|
|
'u-boot-spl/u-boot-spl-bss-pad:size': pad_len,
|
|
|
|
'u-boot-spl/u-boot-spl-dtb:size': 801,
|
|
|
|
'u-boot-spl/u-boot-spl-nodtb:size': len(U_BOOT_SPL_NODTB_DATA),
|
|
|
|
'u-boot-spl:size': 860,
|
|
|
|
'u-boot-tpl:size': len(U_BOOT_TPL_DATA),
|
|
|
|
'u-boot/u-boot-dtb:size': 781,
|
|
|
|
'u-boot/u-boot-nodtb:size': len(U_BOOT_NODTB_DATA),
|
|
|
|
'u-boot:size': 827,
|
|
|
|
}, props)
|
|
|
|
return fdt_size
|
|
|
|
|
|
|
|
def testExpanded(self):
|
|
|
|
"""Test that an expanded entry type is selected when needed"""
|
|
|
|
self._SetupSplElf()
|
|
|
|
self._SetupTplElf()
|
|
|
|
|
|
|
|
# SPL has a devicetree, TPL does not
|
|
|
|
entry_args = {
|
|
|
|
'spl-dtb': '1',
|
|
|
|
'spl-bss-pad': 'y',
|
|
|
|
'tpl-dtb': '',
|
|
|
|
}
|
|
|
|
self._DoReadFileDtb('194_fdt_incl.dts', use_expanded=True,
|
|
|
|
entry_args=entry_args)
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(3, len(entries))
|
|
|
|
|
|
|
|
# First, u-boot, which should be expanded into u-boot-nodtb and dtb
|
|
|
|
self.assertIn('u-boot', entries)
|
|
|
|
entry = entries['u-boot']
|
|
|
|
self.assertEqual('u-boot-expanded', entry.etype)
|
|
|
|
subent = entry.GetEntries()
|
|
|
|
self.assertEqual(2, len(subent))
|
|
|
|
self.assertIn('u-boot-nodtb', subent)
|
|
|
|
self.assertIn('u-boot-dtb', subent)
|
|
|
|
|
|
|
|
# Second, u-boot-spl, which should be expanded into three parts
|
|
|
|
self.assertIn('u-boot-spl', entries)
|
|
|
|
entry = entries['u-boot-spl']
|
|
|
|
self.assertEqual('u-boot-spl-expanded', entry.etype)
|
|
|
|
subent = entry.GetEntries()
|
|
|
|
self.assertEqual(3, len(subent))
|
|
|
|
self.assertIn('u-boot-spl-nodtb', subent)
|
|
|
|
self.assertIn('u-boot-spl-bss-pad', subent)
|
|
|
|
self.assertIn('u-boot-spl-dtb', subent)
|
|
|
|
|
|
|
|
# Third, u-boot-tpl, which should be not be expanded, since TPL has no
|
|
|
|
# devicetree
|
|
|
|
self.assertIn('u-boot-tpl', entries)
|
|
|
|
entry = entries['u-boot-tpl']
|
|
|
|
self.assertEqual('u-boot-tpl', entry.etype)
|
|
|
|
self.assertEqual(None, entry.GetEntries())
|
|
|
|
|
|
|
|
def testExpandedTpl(self):
|
|
|
|
"""Test that an expanded entry type is selected for TPL when needed"""
|
|
|
|
self._SetupTplElf()
|
|
|
|
|
|
|
|
entry_args = {
|
|
|
|
'tpl-bss-pad': 'y',
|
|
|
|
'tpl-dtb': 'y',
|
|
|
|
}
|
|
|
|
self._DoReadFileDtb('195_fdt_incl_tpl.dts', use_expanded=True,
|
|
|
|
entry_args=entry_args)
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(1, len(entries))
|
|
|
|
|
|
|
|
# We only have u-boot-tpl, which be expanded
|
|
|
|
self.assertIn('u-boot-tpl', entries)
|
|
|
|
entry = entries['u-boot-tpl']
|
|
|
|
self.assertEqual('u-boot-tpl-expanded', entry.etype)
|
|
|
|
subent = entry.GetEntries()
|
|
|
|
self.assertEqual(3, len(subent))
|
|
|
|
self.assertIn('u-boot-tpl-nodtb', subent)
|
|
|
|
self.assertIn('u-boot-tpl-bss-pad', subent)
|
|
|
|
self.assertIn('u-boot-tpl-dtb', subent)
|
|
|
|
|
|
|
|
def testExpandedNoPad(self):
|
|
|
|
"""Test an expanded entry without BSS pad enabled"""
|
|
|
|
self._SetupSplElf()
|
|
|
|
self._SetupTplElf()
|
|
|
|
|
|
|
|
# SPL has a devicetree, TPL does not
|
|
|
|
entry_args = {
|
|
|
|
'spl-dtb': 'something',
|
|
|
|
'spl-bss-pad': 'n',
|
|
|
|
'tpl-dtb': '',
|
|
|
|
}
|
|
|
|
self._DoReadFileDtb('194_fdt_incl.dts', use_expanded=True,
|
|
|
|
entry_args=entry_args)
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
|
|
|
|
# Just check u-boot-spl, which should be expanded into two parts
|
|
|
|
self.assertIn('u-boot-spl', entries)
|
|
|
|
entry = entries['u-boot-spl']
|
|
|
|
self.assertEqual('u-boot-spl-expanded', entry.etype)
|
|
|
|
subent = entry.GetEntries()
|
|
|
|
self.assertEqual(2, len(subent))
|
|
|
|
self.assertIn('u-boot-spl-nodtb', subent)
|
|
|
|
self.assertIn('u-boot-spl-dtb', subent)
|
|
|
|
|
|
|
|
def testExpandedTplNoPad(self):
|
|
|
|
"""Test that an expanded entry type with padding disabled in TPL"""
|
|
|
|
self._SetupTplElf()
|
|
|
|
|
|
|
|
entry_args = {
|
|
|
|
'tpl-bss-pad': '',
|
|
|
|
'tpl-dtb': 'y',
|
|
|
|
}
|
|
|
|
self._DoReadFileDtb('195_fdt_incl_tpl.dts', use_expanded=True,
|
|
|
|
entry_args=entry_args)
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(1, len(entries))
|
|
|
|
|
|
|
|
# We only have u-boot-tpl, which be expanded
|
|
|
|
self.assertIn('u-boot-tpl', entries)
|
|
|
|
entry = entries['u-boot-tpl']
|
|
|
|
self.assertEqual('u-boot-tpl-expanded', entry.etype)
|
|
|
|
subent = entry.GetEntries()
|
|
|
|
self.assertEqual(2, len(subent))
|
|
|
|
self.assertIn('u-boot-tpl-nodtb', subent)
|
|
|
|
self.assertIn('u-boot-tpl-dtb', subent)
|
|
|
|
|
|
|
|
def testFdtInclude(self):
|
|
|
|
"""Test that an Fdt is update within all binaries"""
|
|
|
|
self._SetupSplElf()
|
|
|
|
self._SetupTplElf()
|
|
|
|
|
|
|
|
# SPL has a devicetree, TPL does not
|
|
|
|
self.maxDiff = None
|
|
|
|
entry_args = {
|
|
|
|
'spl-dtb': '1',
|
|
|
|
'spl-bss-pad': 'y',
|
|
|
|
'tpl-dtb': '',
|
|
|
|
}
|
|
|
|
# Build the image. It includes two separate devicetree binaries, each
|
|
|
|
# with their own contents, but all contain the binman definition.
|
|
|
|
data = self._DoReadFileDtb(
|
|
|
|
'194_fdt_incl.dts', use_real_dtb=True, use_expanded=True,
|
|
|
|
update_dtb=True, entry_args=entry_args)[0]
|
|
|
|
pad_len = 10
|
|
|
|
|
|
|
|
# Check the U-Boot dtb
|
|
|
|
start = len(U_BOOT_NODTB_DATA)
|
|
|
|
fdt_size = self.checkDtbSizes(data, pad_len, start)
|
|
|
|
|
|
|
|
# Now check SPL
|
|
|
|
start += fdt_size + len(U_BOOT_SPL_NODTB_DATA) + pad_len
|
|
|
|
fdt_size = self.checkDtbSizes(data, pad_len, start)
|
|
|
|
|
|
|
|
# TPL has no devicetree
|
|
|
|
start += fdt_size + len(U_BOOT_TPL_DATA)
|
|
|
|
self.assertEqual(len(data), start)
|
2020-10-26 23:40:14 +00:00
|
|
|
|
2021-03-21 05:24:30 +00:00
|
|
|
def testSymbolsExpanded(self):
|
|
|
|
"""Test binman can assign symbols in expanded entries"""
|
|
|
|
entry_args = {
|
|
|
|
'spl-dtb': '1',
|
|
|
|
}
|
|
|
|
self.checkSymbols('197_symbols_expand.dts', U_BOOT_SPL_NODTB_DATA +
|
|
|
|
U_BOOT_SPL_DTB_DATA, 0x38,
|
|
|
|
entry_args=entry_args, use_expanded=True)
|
|
|
|
|
2021-03-21 05:24:31 +00:00
|
|
|
def testCollection(self):
|
|
|
|
"""Test a collection"""
|
|
|
|
data = self._DoReadFile('198_collection.dts')
|
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA + U_BOOT_DTB_DATA +
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.get_bytes(0xff, 2) + U_BOOT_NODTB_DATA +
|
|
|
|
tools.get_bytes(0xfe, 3) + U_BOOT_DTB_DATA,
|
2021-03-21 05:24:31 +00:00
|
|
|
data)
|
|
|
|
|
binman: Support obtaining section contents immediately
Generally the content of sections is not built until the final assembly
of the image. This is partly to avoid wasting time, since the entries
within sections may change multiple times as binman works through its
various stages. This works quite well since sections exist in a strict
hierarchy, so they can be processed in a depth-first manner.
However the 'collection' entry type does not have this luxury. If it
contains a section within its 'content' list, then it must produce the
section contents, if available. That section is typically a sibling
node, i.e. not part oc the collection's hierarchy.
Add a new 'required' argument to section.GetData() to support this. When
required is True, any referenced sections are immediately built. If this
is not possible (because one of the subentries does not have its data yet)
then an error is produced.
The test for this uses a 'collection' entry type, referencing a section as
its first member. This forces a call to _BuildSectionData() with required
set to False, at first, then True later, when the image is assembled.
Signed-off-by: Simon Glass <sjg@chromium.org>
2021-03-21 05:24:32 +00:00
|
|
|
def testCollectionSection(self):
|
|
|
|
"""Test a collection where a section must be built first"""
|
|
|
|
# Sections never have their contents when GetData() is called, but when
|
2021-11-23 18:03:47 +00:00
|
|
|
# BuildSectionData() is called with required=True, a section will force
|
binman: Support obtaining section contents immediately
Generally the content of sections is not built until the final assembly
of the image. This is partly to avoid wasting time, since the entries
within sections may change multiple times as binman works through its
various stages. This works quite well since sections exist in a strict
hierarchy, so they can be processed in a depth-first manner.
However the 'collection' entry type does not have this luxury. If it
contains a section within its 'content' list, then it must produce the
section contents, if available. That section is typically a sibling
node, i.e. not part oc the collection's hierarchy.
Add a new 'required' argument to section.GetData() to support this. When
required is True, any referenced sections are immediately built. If this
is not possible (because one of the subentries does not have its data yet)
then an error is produced.
The test for this uses a 'collection' entry type, referencing a section as
its first member. This forces a call to _BuildSectionData() with required
set to False, at first, then True later, when the image is assembled.
Signed-off-by: Simon Glass <sjg@chromium.org>
2021-03-21 05:24:32 +00:00
|
|
|
# building the contents, producing an error is anything is still
|
|
|
|
# missing.
|
|
|
|
data = self._DoReadFile('199_collection_section.dts')
|
|
|
|
section = U_BOOT_NODTB_DATA + U_BOOT_DTB_DATA
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(section + U_BOOT_DATA + tools.get_bytes(0xff, 2) +
|
|
|
|
section + tools.get_bytes(0xfe, 3) + U_BOOT_DATA,
|
binman: Support obtaining section contents immediately
Generally the content of sections is not built until the final assembly
of the image. This is partly to avoid wasting time, since the entries
within sections may change multiple times as binman works through its
various stages. This works quite well since sections exist in a strict
hierarchy, so they can be processed in a depth-first manner.
However the 'collection' entry type does not have this luxury. If it
contains a section within its 'content' list, then it must produce the
section contents, if available. That section is typically a sibling
node, i.e. not part oc the collection's hierarchy.
Add a new 'required' argument to section.GetData() to support this. When
required is True, any referenced sections are immediately built. If this
is not possible (because one of the subentries does not have its data yet)
then an error is produced.
The test for this uses a 'collection' entry type, referencing a section as
its first member. This forces a call to _BuildSectionData() with required
set to False, at first, then True later, when the image is assembled.
Signed-off-by: Simon Glass <sjg@chromium.org>
2021-03-21 05:24:32 +00:00
|
|
|
data)
|
|
|
|
|
2021-03-21 05:24:33 +00:00
|
|
|
def testAlignDefault(self):
|
|
|
|
"""Test that default alignment works on sections"""
|
|
|
|
data = self._DoReadFile('200_align_default.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (U_BOOT_DATA + tools.get_bytes(0, 8 - len(U_BOOT_DATA)) +
|
2021-03-21 05:24:33 +00:00
|
|
|
U_BOOT_DATA)
|
|
|
|
# Special alignment for section
|
2022-01-29 21:14:04 +00:00
|
|
|
expected += tools.get_bytes(0, 32 - len(expected))
|
2021-03-21 05:24:33 +00:00
|
|
|
# No alignment within the nested section
|
|
|
|
expected += U_BOOT_DATA + U_BOOT_NODTB_DATA;
|
|
|
|
# Now the final piece, which should be default-aligned
|
2022-01-29 21:14:04 +00:00
|
|
|
expected += tools.get_bytes(0, 88 - len(expected)) + U_BOOT_NODTB_DATA
|
2021-03-21 05:24:33 +00:00
|
|
|
self.assertEqual(expected, data)
|
binman: Support obtaining section contents immediately
Generally the content of sections is not built until the final assembly
of the image. This is partly to avoid wasting time, since the entries
within sections may change multiple times as binman works through its
various stages. This works quite well since sections exist in a strict
hierarchy, so they can be processed in a depth-first manner.
However the 'collection' entry type does not have this luxury. If it
contains a section within its 'content' list, then it must produce the
section contents, if available. That section is typically a sibling
node, i.e. not part oc the collection's hierarchy.
Add a new 'required' argument to section.GetData() to support this. When
required is True, any referenced sections are immediately built. If this
is not possible (because one of the subentries does not have its data yet)
then an error is produced.
The test for this uses a 'collection' entry type, referencing a section as
its first member. This forces a call to _BuildSectionData() with required
set to False, at first, then True later, when the image is assembled.
Signed-off-by: Simon Glass <sjg@chromium.org>
2021-03-21 05:24:32 +00:00
|
|
|
|
2021-05-10 12:23:33 +00:00
|
|
|
def testPackOpenSBI(self):
|
|
|
|
"""Test that an image with an OpenSBI binary can be created"""
|
|
|
|
data = self._DoReadFile('201_opensbi.dts')
|
|
|
|
self.assertEqual(OPENSBI_DATA, data[:len(OPENSBI_DATA)])
|
|
|
|
|
2021-07-06 16:36:37 +00:00
|
|
|
def testSectionsSingleThread(self):
|
|
|
|
"""Test sections without multithreading"""
|
|
|
|
data = self._DoReadFileDtb('055_sections.dts', threads=0)[0]
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = (U_BOOT_DATA + tools.get_bytes(ord('!'), 12) +
|
|
|
|
U_BOOT_DATA + tools.get_bytes(ord('a'), 12) +
|
|
|
|
U_BOOT_DATA + tools.get_bytes(ord('&'), 4))
|
2021-07-06 16:36:37 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testThreadTimeout(self):
|
|
|
|
"""Test handling a thread that takes too long"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoTestFile('202_section_timeout.dts',
|
|
|
|
test_section_timeout=True)
|
2021-10-18 18:13:15 +00:00
|
|
|
self.assertIn("Timed out obtaining contents", str(e.exception))
|
2021-07-06 16:36:37 +00:00
|
|
|
|
2021-07-06 16:36:41 +00:00
|
|
|
def testTiming(self):
|
|
|
|
"""Test output of timing information"""
|
|
|
|
data = self._DoReadFile('055_sections.dts')
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
state.TimingShow()
|
|
|
|
self.assertIn('read:', stdout.getvalue())
|
|
|
|
self.assertIn('compress:', stdout.getvalue())
|
|
|
|
|
2021-11-04 03:09:18 +00:00
|
|
|
def testUpdateFdtInElf(self):
|
|
|
|
"""Test that we can update the devicetree in an ELF file"""
|
2022-08-19 14:25:18 +00:00
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
2021-11-04 03:09:18 +00:00
|
|
|
infile = elf_fname = self.ElfTestFile('u_boot_binman_embed')
|
|
|
|
outfile = os.path.join(self._indir, 'u-boot.out')
|
|
|
|
begin_sym = 'dtb_embed_begin'
|
|
|
|
end_sym = 'dtb_embed_end'
|
|
|
|
retcode = self._DoTestFile(
|
|
|
|
'060_fdt_update.dts', update_dtb=True,
|
|
|
|
update_fdt_in_elf=','.join([infile,outfile,begin_sym,end_sym]))
|
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
|
|
|
|
# Check that the output file does in fact contact a dtb with the binman
|
|
|
|
# definition in the correct place
|
|
|
|
syms = elf.GetSymbolFileOffset(infile,
|
|
|
|
['dtb_embed_begin', 'dtb_embed_end'])
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(outfile)
|
2021-11-04 03:09:18 +00:00
|
|
|
dtb_data = data[syms['dtb_embed_begin'].offset:
|
|
|
|
syms['dtb_embed_end'].offset]
|
|
|
|
|
|
|
|
dtb = fdt.Fdt.FromData(dtb_data)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS + REPACK_DTB_PROPS)
|
|
|
|
self.assertEqual({
|
|
|
|
'image-pos': 0,
|
|
|
|
'offset': 0,
|
|
|
|
'_testing:offset': 32,
|
|
|
|
'_testing:size': 2,
|
|
|
|
'_testing:image-pos': 32,
|
|
|
|
'section@0/u-boot:offset': 0,
|
|
|
|
'section@0/u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'section@0/u-boot:image-pos': 0,
|
|
|
|
'section@0:offset': 0,
|
|
|
|
'section@0:size': 16,
|
|
|
|
'section@0:image-pos': 0,
|
|
|
|
|
|
|
|
'section@1/u-boot:offset': 0,
|
|
|
|
'section@1/u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'section@1/u-boot:image-pos': 16,
|
|
|
|
'section@1:offset': 16,
|
|
|
|
'section@1:size': 16,
|
|
|
|
'section@1:image-pos': 16,
|
|
|
|
'size': 40
|
|
|
|
}, props)
|
|
|
|
|
|
|
|
def testUpdateFdtInElfInvalid(self):
|
|
|
|
"""Test that invalid args are detected with --update-fdt-in-elf"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoTestFile('060_fdt_update.dts', update_fdt_in_elf='fred')
|
|
|
|
self.assertIn("Invalid args ['fred'] to --update-fdt-in-elf",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testUpdateFdtInElfNoSyms(self):
|
|
|
|
"""Test that missing symbols are detected with --update-fdt-in-elf"""
|
2022-08-19 14:25:18 +00:00
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
2021-11-04 03:09:18 +00:00
|
|
|
infile = elf_fname = self.ElfTestFile('u_boot_binman_embed')
|
|
|
|
outfile = ''
|
|
|
|
begin_sym = 'wrong_begin'
|
|
|
|
end_sym = 'wrong_end'
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoTestFile(
|
|
|
|
'060_fdt_update.dts',
|
|
|
|
update_fdt_in_elf=','.join([infile,outfile,begin_sym,end_sym]))
|
|
|
|
self.assertIn("Expected two symbols 'wrong_begin' and 'wrong_end': got 0:",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testUpdateFdtInElfTooSmall(self):
|
|
|
|
"""Test that an over-large dtb is detected with --update-fdt-in-elf"""
|
2022-08-19 14:25:18 +00:00
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
2021-11-04 03:09:18 +00:00
|
|
|
infile = elf_fname = self.ElfTestFile('u_boot_binman_embed_sm')
|
|
|
|
outfile = os.path.join(self._indir, 'u-boot.out')
|
|
|
|
begin_sym = 'dtb_embed_begin'
|
|
|
|
end_sym = 'dtb_embed_end'
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoTestFile(
|
|
|
|
'060_fdt_update.dts', update_dtb=True,
|
|
|
|
update_fdt_in_elf=','.join([infile,outfile,begin_sym,end_sym]))
|
|
|
|
self.assertRegex(
|
|
|
|
str(e.exception),
|
|
|
|
"Not enough space in '.*u_boot_binman_embed_sm' for data length.*")
|
|
|
|
|
2021-11-23 18:03:42 +00:00
|
|
|
def testVersion(self):
|
|
|
|
"""Test we can get the binman version"""
|
|
|
|
version = '(unreleased)'
|
|
|
|
self.assertEqual(version, state.GetVersion(self._indir))
|
|
|
|
|
|
|
|
with self.assertRaises(SystemExit):
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoBinman('-V')
|
|
|
|
self.assertEqual('Binman %s\n' % version, stderr.getvalue())
|
|
|
|
|
|
|
|
# Try running the tool too, just to be safe
|
|
|
|
result = self._RunBinman('-V')
|
|
|
|
self.assertEqual('Binman %s\n' % version, result.stderr)
|
|
|
|
|
|
|
|
# Set up a version file to make sure that works
|
|
|
|
version = 'v2025.01-rc2'
|
2022-01-29 21:14:04 +00:00
|
|
|
tools.write_file(os.path.join(self._indir, 'version'), version,
|
2021-11-23 18:03:42 +00:00
|
|
|
binary=False)
|
|
|
|
self.assertEqual(version, state.GetVersion(self._indir))
|
|
|
|
|
2021-11-24 04:09:50 +00:00
|
|
|
def testAltFormat(self):
|
|
|
|
"""Test that alternative formats can be used to extract"""
|
|
|
|
self._DoReadFileRealDtb('213_fdtmap_alt_format.dts')
|
|
|
|
|
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
with test_util.capture_sys_output() as (stdout, _):
|
|
|
|
self._DoBinman('extract', '-i', updated_fname, '-F', 'list')
|
|
|
|
self.assertEqual(
|
|
|
|
'''Flag (-F) Entry type Description
|
|
|
|
fdt fdtmap Extract the devicetree blob from the fdtmap
|
|
|
|
''',
|
|
|
|
stdout.getvalue())
|
|
|
|
|
|
|
|
dtb = os.path.join(tmpdir, 'fdt.dtb')
|
|
|
|
self._DoBinman('extract', '-i', updated_fname, '-F', 'fdt', '-f',
|
|
|
|
dtb, 'fdtmap')
|
|
|
|
|
|
|
|
# Check that we can read it and it can be scanning, meaning it does
|
|
|
|
# not have a 16-byte fdtmap header
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(dtb)
|
2021-11-24 04:09:50 +00:00
|
|
|
dtb = fdt.Fdt.FromData(data)
|
|
|
|
dtb.Scan()
|
|
|
|
|
|
|
|
# Now check u-boot which has no alt_format
|
|
|
|
fname = os.path.join(tmpdir, 'fdt.dtb')
|
|
|
|
self._DoBinman('extract', '-i', updated_fname, '-F', 'dummy',
|
|
|
|
'-f', fname, 'u-boot')
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(fname)
|
2021-11-24 04:09:50 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
|
|
|
|
2021-11-24 04:09:52 +00:00
|
|
|
def testExtblobList(self):
|
|
|
|
"""Test an image with an external blob list"""
|
|
|
|
data = self._DoReadFile('215_blob_ext_list.dts')
|
|
|
|
self.assertEqual(REFCODE_DATA + FSP_M_DATA, data)
|
|
|
|
|
|
|
|
def testExtblobListMissing(self):
|
|
|
|
"""Test an image with a missing external blob"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('216_blob_ext_list_missing.dts')
|
|
|
|
self.assertIn("Filename 'missing-file' not found in input path",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testExtblobListMissingOk(self):
|
|
|
|
"""Test an image with an missing external blob that is allowed"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('216_blob_ext_list_missing.dts',
|
|
|
|
allow_missing=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err, "Image 'main-section'.*missing.*: blob-ext")
|
|
|
|
|
2021-11-24 04:08:59 +00:00
|
|
|
def testFip(self):
|
|
|
|
"""Basic test of generation of an ARM Firmware Image Package (FIP)"""
|
|
|
|
data = self._DoReadFile('203_fip.dts')
|
|
|
|
hdr, fents = fip_util.decode_fip(data)
|
|
|
|
self.assertEqual(fip_util.HEADER_MAGIC, hdr.name)
|
|
|
|
self.assertEqual(fip_util.HEADER_SERIAL, hdr.serial)
|
|
|
|
self.assertEqual(0x123, hdr.flags)
|
|
|
|
|
|
|
|
self.assertEqual(2, len(fents))
|
|
|
|
|
|
|
|
fent = fents[0]
|
|
|
|
self.assertEqual(
|
|
|
|
bytes([0x47, 0xd4, 0x08, 0x6d, 0x4c, 0xfe, 0x98, 0x46,
|
|
|
|
0x9b, 0x95, 0x29, 0x50, 0xcb, 0xbd, 0x5a, 0x0]), fent.uuid)
|
|
|
|
self.assertEqual('soc-fw', fent.fip_type)
|
|
|
|
self.assertEqual(0x88, fent.offset)
|
|
|
|
self.assertEqual(len(ATF_BL31_DATA), fent.size)
|
|
|
|
self.assertEqual(0x123456789abcdef, fent.flags)
|
|
|
|
self.assertEqual(ATF_BL31_DATA, fent.data)
|
|
|
|
self.assertEqual(True, fent.valid)
|
|
|
|
|
|
|
|
fent = fents[1]
|
|
|
|
self.assertEqual(
|
|
|
|
bytes([0x65, 0x92, 0x27, 0x03, 0x2f, 0x74, 0xe6, 0x44,
|
|
|
|
0x8d, 0xff, 0x57, 0x9a, 0xc1, 0xff, 0x06, 0x10]), fent.uuid)
|
|
|
|
self.assertEqual('scp-fwu-cfg', fent.fip_type)
|
|
|
|
self.assertEqual(0x8c, fent.offset)
|
|
|
|
self.assertEqual(len(ATF_BL31_DATA), fent.size)
|
|
|
|
self.assertEqual(0, fent.flags)
|
|
|
|
self.assertEqual(ATF_BL2U_DATA, fent.data)
|
|
|
|
self.assertEqual(True, fent.valid)
|
|
|
|
|
|
|
|
def testFipOther(self):
|
|
|
|
"""Basic FIP with something that isn't a external blob"""
|
|
|
|
data = self._DoReadFile('204_fip_other.dts')
|
|
|
|
hdr, fents = fip_util.decode_fip(data)
|
|
|
|
|
|
|
|
self.assertEqual(2, len(fents))
|
|
|
|
fent = fents[1]
|
|
|
|
self.assertEqual('rot-cert', fent.fip_type)
|
|
|
|
self.assertEqual(b'aa', fent.data)
|
|
|
|
|
|
|
|
def testFipNoType(self):
|
|
|
|
"""FIP with an entry of an unknown type"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('205_fip_no_type.dts')
|
|
|
|
self.assertIn("Must provide a fip-type (node name 'u-boot' is not a known FIP type)",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testFipUuid(self):
|
|
|
|
"""Basic FIP with a manual uuid"""
|
|
|
|
data = self._DoReadFile('206_fip_uuid.dts')
|
|
|
|
hdr, fents = fip_util.decode_fip(data)
|
|
|
|
|
|
|
|
self.assertEqual(2, len(fents))
|
|
|
|
fent = fents[1]
|
|
|
|
self.assertEqual(None, fent.fip_type)
|
|
|
|
self.assertEqual(
|
|
|
|
bytes([0xfc, 0x65, 0x13, 0x92, 0x4a, 0x5b, 0x11, 0xec,
|
|
|
|
0x94, 0x35, 0xff, 0x2d, 0x1c, 0xfc, 0x79, 0x9c]),
|
|
|
|
fent.uuid)
|
|
|
|
self.assertEqual(U_BOOT_DATA, fent.data)
|
|
|
|
|
|
|
|
def testFipLs(self):
|
|
|
|
"""Test listing a FIP"""
|
|
|
|
data = self._DoReadFileRealDtb('207_fip_ls.dts')
|
|
|
|
hdr, fents = fip_util.decode_fip(data)
|
|
|
|
|
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoBinman('ls', '-i', updated_fname)
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
|
|
|
lines = stdout.getvalue().splitlines()
|
|
|
|
expected = [
|
|
|
|
'Name Image-pos Size Entry-type Offset Uncomp-size',
|
|
|
|
'----------------------------------------------------------------',
|
|
|
|
'main-section 0 2d3 section 0',
|
|
|
|
' atf-fip 0 90 atf-fip 0',
|
|
|
|
' soc-fw 88 4 blob-ext 88',
|
|
|
|
' u-boot 8c 4 u-boot 8c',
|
|
|
|
' fdtmap 90 243 fdtmap 90',
|
|
|
|
]
|
|
|
|
self.assertEqual(expected, lines)
|
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
fdtmap = entries['fdtmap']
|
|
|
|
|
|
|
|
fdtmap_data = data[fdtmap.image_pos:fdtmap.image_pos + fdtmap.size]
|
|
|
|
magic = fdtmap_data[:8]
|
|
|
|
self.assertEqual(b'_FDTMAP_', magic)
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0, 8), fdtmap_data[8:16])
|
2021-11-24 04:08:59 +00:00
|
|
|
|
|
|
|
fdt_data = fdtmap_data[16:]
|
|
|
|
dtb = fdt.Fdt.FromData(fdt_data)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, BASE_DTB_PROPS, prefix='/')
|
|
|
|
self.assertEqual({
|
|
|
|
'atf-fip/soc-fw:image-pos': 136,
|
|
|
|
'atf-fip/soc-fw:offset': 136,
|
|
|
|
'atf-fip/soc-fw:size': 4,
|
|
|
|
'atf-fip/u-boot:image-pos': 140,
|
|
|
|
'atf-fip/u-boot:offset': 140,
|
|
|
|
'atf-fip/u-boot:size': 4,
|
|
|
|
'atf-fip:image-pos': 0,
|
|
|
|
'atf-fip:offset': 0,
|
|
|
|
'atf-fip:size': 144,
|
|
|
|
'image-pos': 0,
|
|
|
|
'offset': 0,
|
|
|
|
'fdtmap:image-pos': fdtmap.image_pos,
|
|
|
|
'fdtmap:offset': fdtmap.offset,
|
|
|
|
'fdtmap:size': len(fdtmap_data),
|
|
|
|
'size': len(data),
|
|
|
|
}, props)
|
|
|
|
|
|
|
|
def testFipExtractOneEntry(self):
|
|
|
|
"""Test extracting a single entry fron an FIP"""
|
|
|
|
self._DoReadFileRealDtb('207_fip_ls.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
2021-11-24 04:08:59 +00:00
|
|
|
fname = os.path.join(self._indir, 'output.extact')
|
|
|
|
control.ExtractEntries(image_fname, fname, None, ['atf-fip/u-boot'])
|
2022-01-29 21:14:04 +00:00
|
|
|
data = tools.read_file(fname)
|
2021-11-24 04:08:59 +00:00
|
|
|
self.assertEqual(U_BOOT_DATA, data)
|
|
|
|
|
|
|
|
def testFipReplace(self):
|
|
|
|
"""Test replacing a single file in a FIP"""
|
2022-01-29 21:14:04 +00:00
|
|
|
expected = U_BOOT_DATA + tools.get_bytes(0x78, 50)
|
2021-11-24 04:08:59 +00:00
|
|
|
data = self._DoReadFileRealDtb('208_fip_replace.dts')
|
2022-01-29 21:14:04 +00:00
|
|
|
updated_fname = tools.get_output_filename('image-updated.bin')
|
|
|
|
tools.write_file(updated_fname, data)
|
2021-11-24 04:08:59 +00:00
|
|
|
entry_name = 'atf-fip/u-boot'
|
|
|
|
control.WriteEntry(updated_fname, entry_name, expected,
|
|
|
|
allow_resize=True)
|
|
|
|
actual = control.ReadEntry(updated_fname, entry_name)
|
|
|
|
self.assertEqual(expected, actual)
|
|
|
|
|
2022-01-29 21:14:04 +00:00
|
|
|
new_data = tools.read_file(updated_fname)
|
2021-11-24 04:08:59 +00:00
|
|
|
hdr, fents = fip_util.decode_fip(new_data)
|
|
|
|
|
|
|
|
self.assertEqual(2, len(fents))
|
|
|
|
|
|
|
|
# Check that the FIP entry is updated
|
|
|
|
fent = fents[1]
|
|
|
|
self.assertEqual(0x8c, fent.offset)
|
|
|
|
self.assertEqual(len(expected), fent.size)
|
|
|
|
self.assertEqual(0, fent.flags)
|
|
|
|
self.assertEqual(expected, fent.data)
|
|
|
|
self.assertEqual(True, fent.valid)
|
|
|
|
|
|
|
|
def testFipMissing(self):
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('209_fip_missing.dts', allow_missing=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err, "Image 'main-section'.*missing.*: rmm-fw")
|
|
|
|
|
|
|
|
def testFipSize(self):
|
|
|
|
"""Test a FIP with a size property"""
|
|
|
|
data = self._DoReadFile('210_fip_size.dts')
|
|
|
|
self.assertEqual(0x100 + len(U_BOOT_DATA), len(data))
|
|
|
|
hdr, fents = fip_util.decode_fip(data)
|
|
|
|
self.assertEqual(fip_util.HEADER_MAGIC, hdr.name)
|
|
|
|
self.assertEqual(fip_util.HEADER_SERIAL, hdr.serial)
|
|
|
|
|
|
|
|
self.assertEqual(1, len(fents))
|
|
|
|
|
|
|
|
fent = fents[0]
|
|
|
|
self.assertEqual('soc-fw', fent.fip_type)
|
|
|
|
self.assertEqual(0x60, fent.offset)
|
|
|
|
self.assertEqual(len(ATF_BL31_DATA), fent.size)
|
|
|
|
self.assertEqual(ATF_BL31_DATA, fent.data)
|
|
|
|
self.assertEqual(True, fent.valid)
|
|
|
|
|
|
|
|
rest = data[0x60 + len(ATF_BL31_DATA):0x100]
|
2022-01-29 21:14:04 +00:00
|
|
|
self.assertEqual(tools.get_bytes(0xff, len(rest)), rest)
|
2021-11-24 04:08:59 +00:00
|
|
|
|
|
|
|
def testFipBadAlign(self):
|
|
|
|
"""Test that an invalid alignment value in a FIP is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoTestFile('211_fip_bad_align.dts')
|
|
|
|
self.assertIn(
|
|
|
|
"Node \'/binman/atf-fip\': FIP alignment 31 must be a power of two",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
def testFipCollection(self):
|
|
|
|
"""Test using a FIP in a collection"""
|
|
|
|
data = self._DoReadFile('212_fip_collection.dts')
|
|
|
|
entry1 = control.images['image'].GetEntries()['collection']
|
|
|
|
data1 = data[:entry1.size]
|
|
|
|
hdr1, fents2 = fip_util.decode_fip(data1)
|
|
|
|
|
|
|
|
entry2 = control.images['image'].GetEntries()['atf-fip']
|
|
|
|
data2 = data[entry2.offset:entry2.offset + entry2.size]
|
|
|
|
hdr1, fents2 = fip_util.decode_fip(data2)
|
|
|
|
|
|
|
|
# The 'collection' entry should have U-Boot included at the end
|
|
|
|
self.assertEqual(entry1.size - len(U_BOOT_DATA), entry2.size)
|
|
|
|
self.assertEqual(data1, data2 + U_BOOT_DATA)
|
|
|
|
self.assertEqual(U_BOOT_DATA, data1[-4:])
|
|
|
|
|
|
|
|
# There should be a U-Boot after the final FIP
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[-4:])
|
2021-07-06 16:36:37 +00:00
|
|
|
|
2022-01-12 20:10:35 +00:00
|
|
|
def testFakeBlob(self):
|
|
|
|
"""Test handling of faking an external blob"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('217_fake_blob.dts', allow_missing=True,
|
|
|
|
allow_fake_blobs=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(
|
|
|
|
err,
|
|
|
|
"Image '.*' has faked external blobs and is non-functional: .*")
|
|
|
|
|
2022-01-10 03:13:46 +00:00
|
|
|
def testExtblobListFaked(self):
|
|
|
|
"""Test an extblob with missing external blob that are faked"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('216_blob_ext_list_missing.dts',
|
|
|
|
allow_fake_blobs=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err, "Image 'main-section'.*faked.*: blob-ext-list")
|
|
|
|
|
2022-01-10 03:13:57 +00:00
|
|
|
def testListBintools(self):
|
|
|
|
args = ['tool', '--list']
|
|
|
|
with test_util.capture_sys_output() as (stdout, _):
|
|
|
|
self._DoBinman(*args)
|
|
|
|
out = stdout.getvalue().splitlines()
|
|
|
|
self.assertTrue(len(out) >= 2)
|
|
|
|
|
|
|
|
def testFetchBintools(self):
|
|
|
|
def fail_download(url):
|
2022-01-29 21:14:04 +00:00
|
|
|
"""Take the tools.download() function by raising an exception"""
|
2022-01-10 03:13:57 +00:00
|
|
|
raise urllib.error.URLError('my error')
|
|
|
|
|
|
|
|
args = ['tool']
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoBinman(*args)
|
|
|
|
self.assertIn("Invalid arguments to 'tool' subcommand",
|
|
|
|
str(e.exception))
|
|
|
|
|
|
|
|
args = ['tool', '--fetch']
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoBinman(*args)
|
|
|
|
self.assertIn('Please specify bintools to fetch', str(e.exception))
|
|
|
|
|
|
|
|
args = ['tool', '--fetch', '_testing']
|
2022-01-29 21:14:04 +00:00
|
|
|
with unittest.mock.patch.object(tools, 'download',
|
2022-01-10 03:13:57 +00:00
|
|
|
side_effect=fail_download):
|
|
|
|
with test_util.capture_sys_output() as (stdout, _):
|
|
|
|
self._DoBinman(*args)
|
|
|
|
self.assertIn('failed to fetch with all methods', stdout.getvalue())
|
|
|
|
|
2022-01-10 03:14:11 +00:00
|
|
|
def testBintoolDocs(self):
|
|
|
|
"""Test for creation of bintool documentation"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
control.write_bintool_docs(control.bintool.Bintool.get_tool_list())
|
|
|
|
self.assertTrue(len(stdout.getvalue()) > 0)
|
|
|
|
|
|
|
|
def testBintoolDocsMissing(self):
|
|
|
|
"""Test handling of missing bintool documentation"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
control.write_bintool_docs(
|
|
|
|
control.bintool.Bintool.get_tool_list(), 'mkimage')
|
|
|
|
self.assertIn('Documentation is missing for modules: mkimage',
|
|
|
|
str(e.exception))
|
|
|
|
|
2022-01-28 19:37:53 +00:00
|
|
|
def testListWithGenNode(self):
|
|
|
|
"""Check handling of an FDT map when the section cannot be found"""
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
}
|
|
|
|
data = self._DoReadFileDtb(
|
|
|
|
'219_fit_gennode.dts',
|
|
|
|
entry_args=entry_args,
|
|
|
|
use_real_dtb=True,
|
|
|
|
extra_indirs=[os.path.join(self._indir, TEST_FDT_SUBDIR)])
|
|
|
|
|
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._RunBinman('ls', '-i', updated_fname)
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
|
|
|
|
2022-02-07 22:08:05 +00:00
|
|
|
def testFitSubentryUsesBintool(self):
|
|
|
|
"""Test that binman FIT subentries can use bintools"""
|
|
|
|
command.test_result = self._HandleGbbCommand
|
|
|
|
entry_args = {
|
|
|
|
'keydir': 'devkeys',
|
|
|
|
'bmpblk': 'bmpblk.bin',
|
|
|
|
}
|
|
|
|
data, _, _, _ = self._DoReadFileDtb('220_fit_subentry_bintool.dts',
|
|
|
|
entry_args=entry_args)
|
|
|
|
|
2022-02-07 22:08:07 +00:00
|
|
|
expected = (GBB_DATA + GBB_DATA + tools.get_bytes(0, 8) +
|
|
|
|
tools.get_bytes(0, 0x2180 - 16))
|
2022-02-07 22:08:05 +00:00
|
|
|
self.assertIn(expected, data)
|
|
|
|
|
|
|
|
def testFitSubentryMissingBintool(self):
|
|
|
|
"""Test that binman reports missing bintools for FIT subentries"""
|
|
|
|
entry_args = {
|
|
|
|
'keydir': 'devkeys',
|
|
|
|
}
|
|
|
|
with test_util.capture_sys_output() as (_, stderr):
|
|
|
|
self._DoTestFile('220_fit_subentry_bintool.dts',
|
|
|
|
force_missing_bintools='futility', entry_args=entry_args)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(err,
|
|
|
|
"Image 'main-section'.*missing bintools.*: futility")
|
2022-01-12 20:10:35 +00:00
|
|
|
|
2022-02-09 19:02:35 +00:00
|
|
|
def testFitSubentryHashSubnode(self):
|
|
|
|
"""Test an image with a FIT inside"""
|
|
|
|
data, _, _, out_dtb_name = self._DoReadFileDtb(
|
|
|
|
'221_fit_subentry_hash.dts', use_real_dtb=True, update_dtb=True)
|
|
|
|
|
|
|
|
mkimage_dtb = fdt.Fdt.FromData(data)
|
|
|
|
mkimage_dtb.Scan()
|
|
|
|
binman_dtb = fdt.Fdt(out_dtb_name)
|
|
|
|
binman_dtb.Scan()
|
|
|
|
|
|
|
|
# Check that binman didn't add hash values
|
|
|
|
fnode = binman_dtb.GetNode('/binman/fit/images/kernel/hash')
|
|
|
|
self.assertNotIn('value', fnode.props)
|
|
|
|
|
|
|
|
fnode = binman_dtb.GetNode('/binman/fit/images/fdt-1/hash')
|
|
|
|
self.assertNotIn('value', fnode.props)
|
|
|
|
|
|
|
|
# Check that mkimage added hash values
|
|
|
|
fnode = mkimage_dtb.GetNode('/images/kernel/hash')
|
|
|
|
self.assertIn('value', fnode.props)
|
|
|
|
|
|
|
|
fnode = mkimage_dtb.GetNode('/images/fdt-1/hash')
|
|
|
|
self.assertIn('value', fnode.props)
|
|
|
|
|
2022-02-19 18:50:04 +00:00
|
|
|
def testPackTeeOs(self):
|
|
|
|
"""Test that an image with an TEE binary can be created"""
|
|
|
|
data = self._DoReadFile('222_tee_os.dts')
|
|
|
|
self.assertEqual(TEE_OS_DATA, data[:len(TEE_OS_DATA)])
|
|
|
|
|
2022-02-08 18:50:03 +00:00
|
|
|
def testFitFdtOper(self):
|
|
|
|
"""Check handling of a specified FIT operation"""
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
'default-dt': 'test-fdt2',
|
|
|
|
}
|
|
|
|
self._DoReadFileDtb(
|
|
|
|
'223_fit_fdt_oper.dts',
|
|
|
|
entry_args=entry_args,
|
|
|
|
extra_indirs=[os.path.join(self._indir, TEST_FDT_SUBDIR)])[0]
|
|
|
|
|
|
|
|
def testFitFdtBadOper(self):
|
|
|
|
"""Check handling of an FDT map when the section cannot be found"""
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self._DoReadFileDtb('224_fit_bad_oper.dts')
|
2022-03-06 03:19:09 +00:00
|
|
|
self.assertIn("Node '/binman/fit': subnode 'images/@fdt-SEQ': Unknown operation 'unknown'",
|
2022-02-08 18:50:03 +00:00
|
|
|
str(exc.exception))
|
|
|
|
|
2022-03-06 03:18:59 +00:00
|
|
|
def test_uses_expand_size(self):
|
|
|
|
"""Test that the 'expand-size' property cannot be used anymore"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
data = self._DoReadFile('225_expand_size_bad.dts')
|
|
|
|
self.assertIn(
|
|
|
|
"Node '/binman/u-boot': Please use 'extend-size' instead of 'expand-size'",
|
|
|
|
str(e.exception))
|
|
|
|
|
2022-03-06 03:19:12 +00:00
|
|
|
def testFitSplitElf(self):
|
|
|
|
"""Test an image with an FIT with an split-elf operation"""
|
2022-08-19 14:25:18 +00:00
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
2022-03-06 03:19:12 +00:00
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
'default-dt': 'test-fdt2',
|
|
|
|
'atf-bl31-path': 'bl31.elf',
|
|
|
|
'tee-os-path': 'tee.elf',
|
|
|
|
}
|
|
|
|
test_subdir = os.path.join(self._indir, TEST_FDT_SUBDIR)
|
|
|
|
data = self._DoReadFileDtb(
|
|
|
|
'226_fit_split_elf.dts',
|
|
|
|
entry_args=entry_args,
|
|
|
|
extra_indirs=[test_subdir])[0]
|
|
|
|
|
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA, data[-len(U_BOOT_NODTB_DATA):])
|
|
|
|
fit_data = data[len(U_BOOT_DATA):-len(U_BOOT_NODTB_DATA)]
|
|
|
|
|
|
|
|
base_keys = {'description', 'type', 'arch', 'os', 'compression',
|
|
|
|
'data', 'load'}
|
|
|
|
dtb = fdt.Fdt.FromData(fit_data)
|
|
|
|
dtb.Scan()
|
|
|
|
|
|
|
|
elf_data = tools.read_file(os.path.join(self._indir, 'bl31.elf'))
|
|
|
|
segments, entry = elf.read_loadable_segments(elf_data)
|
|
|
|
|
|
|
|
# We assume there are two segments
|
|
|
|
self.assertEquals(2, len(segments))
|
|
|
|
|
|
|
|
atf1 = dtb.GetNode('/images/atf-1')
|
|
|
|
_, start, data = segments[0]
|
|
|
|
self.assertEqual(base_keys | {'entry'}, atf1.props.keys())
|
|
|
|
self.assertEqual(entry,
|
|
|
|
fdt_util.fdt32_to_cpu(atf1.props['entry'].value))
|
|
|
|
self.assertEqual(start,
|
|
|
|
fdt_util.fdt32_to_cpu(atf1.props['load'].value))
|
|
|
|
self.assertEqual(data, atf1.props['data'].bytes)
|
|
|
|
|
|
|
|
atf2 = dtb.GetNode('/images/atf-2')
|
|
|
|
self.assertEqual(base_keys, atf2.props.keys())
|
|
|
|
_, start, data = segments[1]
|
|
|
|
self.assertEqual(start,
|
|
|
|
fdt_util.fdt32_to_cpu(atf2.props['load'].value))
|
|
|
|
self.assertEqual(data, atf2.props['data'].bytes)
|
|
|
|
|
|
|
|
conf = dtb.GetNode('/configurations')
|
|
|
|
self.assertEqual({'default'}, conf.props.keys())
|
|
|
|
|
|
|
|
for subnode in conf.subnodes:
|
|
|
|
self.assertEqual({'description', 'fdt', 'loadables'},
|
|
|
|
subnode.props.keys())
|
|
|
|
self.assertEqual(
|
|
|
|
['atf-1', 'atf-2', 'tee-1', 'tee-2'],
|
|
|
|
fdt_util.GetStringList(subnode, 'loadables'))
|
|
|
|
|
|
|
|
def _check_bad_fit(self, dts):
|
|
|
|
"""Check a bad FIT
|
|
|
|
|
|
|
|
This runs with the given dts and returns the assertion raised
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dts (str): dts filename to use
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
str: Assertion string raised
|
|
|
|
"""
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
'default-dt': 'test-fdt2',
|
|
|
|
'atf-bl31-path': 'bl31.elf',
|
|
|
|
'tee-os-path': 'tee.elf',
|
|
|
|
}
|
|
|
|
test_subdir = os.path.join(self._indir, TEST_FDT_SUBDIR)
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self._DoReadFileDtb(dts, entry_args=entry_args,
|
|
|
|
extra_indirs=[test_subdir])[0]
|
|
|
|
return str(exc.exception)
|
|
|
|
|
|
|
|
def testFitSplitElfBadElf(self):
|
|
|
|
"""Test a FIT split-elf operation with an invalid ELF file"""
|
2022-08-19 14:25:18 +00:00
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
2022-03-06 03:19:12 +00:00
|
|
|
TestFunctional._MakeInputFile('bad.elf', tools.get_bytes(100, 100))
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
'default-dt': 'test-fdt2',
|
|
|
|
'atf-bl31-path': 'bad.elf',
|
|
|
|
'tee-os-path': 'tee.elf',
|
|
|
|
}
|
|
|
|
test_subdir = os.path.join(self._indir, TEST_FDT_SUBDIR)
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self._DoReadFileDtb(
|
|
|
|
'226_fit_split_elf.dts',
|
|
|
|
entry_args=entry_args,
|
|
|
|
extra_indirs=[test_subdir])[0]
|
|
|
|
self.assertIn(
|
|
|
|
"Node '/binman/fit': subnode 'images/@atf-SEQ': Failed to read ELF file: Magic number does not match",
|
|
|
|
str(exc.exception))
|
|
|
|
|
|
|
|
def checkFitSplitElf(self, **kwargs):
|
2022-08-07 15:46:46 +00:00
|
|
|
"""Test an split-elf FIT with a missing ELF file
|
|
|
|
|
|
|
|
Args:
|
|
|
|
kwargs (dict of str): Arguments to pass to _DoTestFile()
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
tuple:
|
|
|
|
str: stdout result
|
|
|
|
str: stderr result
|
|
|
|
"""
|
2022-03-06 03:19:12 +00:00
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
'default-dt': 'test-fdt2',
|
|
|
|
'atf-bl31-path': 'bl31.elf',
|
|
|
|
'tee-os-path': 'missing.elf',
|
|
|
|
}
|
|
|
|
test_subdir = os.path.join(self._indir, TEST_FDT_SUBDIR)
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile(
|
|
|
|
'226_fit_split_elf.dts', entry_args=entry_args,
|
2022-08-07 15:46:46 +00:00
|
|
|
extra_indirs=[test_subdir], verbosity=3, **kwargs)
|
|
|
|
out = stdout.getvalue()
|
|
|
|
err = stderr.getvalue()
|
|
|
|
return out, err
|
2022-03-06 03:19:12 +00:00
|
|
|
|
2022-08-23 10:46:09 +00:00
|
|
|
def testFitSplitElfBadDirective(self):
|
|
|
|
"""Test a FIT split-elf invalid fit,xxx directive in an image node"""
|
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
|
|
|
err = self._check_bad_fit('227_fit_bad_dir.dts')
|
|
|
|
self.assertIn(
|
|
|
|
"Node '/binman/fit': subnode 'images/@atf-SEQ': Unknown directive 'fit,something'",
|
|
|
|
err)
|
|
|
|
|
|
|
|
def testFitSplitElfBadDirectiveConfig(self):
|
|
|
|
"""Test a FIT split-elf with invalid fit,xxx directive in config"""
|
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
|
|
|
err = self._check_bad_fit('228_fit_bad_dir_config.dts')
|
|
|
|
self.assertEqual(
|
|
|
|
"Node '/binman/fit': subnode 'configurations/@config-SEQ': Unknown directive 'fit,config'",
|
|
|
|
err)
|
|
|
|
|
|
|
|
|
2022-03-06 03:19:12 +00:00
|
|
|
def testFitSplitElfMissing(self):
|
|
|
|
"""Test an split-elf FIT with a missing ELF file"""
|
2022-08-19 14:25:18 +00:00
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
2022-08-07 15:46:46 +00:00
|
|
|
out, err = self.checkFitSplitElf(allow_missing=True)
|
2022-03-06 03:19:12 +00:00
|
|
|
self.assertRegex(
|
|
|
|
err,
|
|
|
|
"Image '.*' is missing external blobs and is non-functional: .*")
|
2022-08-07 15:46:46 +00:00
|
|
|
self.assertNotRegex(out, '.*Faked blob.*')
|
|
|
|
fname = tools.get_output_filename('binman-fake/missing.elf')
|
|
|
|
self.assertFalse(os.path.exists(fname))
|
2022-03-06 03:19:12 +00:00
|
|
|
|
|
|
|
def testFitSplitElfFaked(self):
|
|
|
|
"""Test an split-elf FIT with faked ELF file"""
|
2022-08-19 14:25:18 +00:00
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
2022-08-07 15:46:46 +00:00
|
|
|
out, err = self.checkFitSplitElf(allow_missing=True, allow_fake_blobs=True)
|
2022-03-06 03:19:12 +00:00
|
|
|
self.assertRegex(
|
|
|
|
err,
|
|
|
|
"Image '.*' is missing external blobs and is non-functional: .*")
|
2022-08-07 15:46:46 +00:00
|
|
|
self.assertRegex(
|
|
|
|
out,
|
|
|
|
"Entry '/binman/fit/images/@tee-SEQ/tee-os': Faked blob '.*binman-fake/missing.elf")
|
|
|
|
fname = tools.get_output_filename('binman-fake/missing.elf')
|
|
|
|
self.assertTrue(os.path.exists(fname))
|
2022-03-06 03:19:12 +00:00
|
|
|
|
2022-08-23 10:46:09 +00:00
|
|
|
def testMkimageMissingBlob(self):
|
|
|
|
"""Test using mkimage to build an image"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
self._DoTestFile('229_mkimage_missing.dts', allow_missing=True,
|
|
|
|
allow_fake_blobs=True)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(
|
|
|
|
err,
|
|
|
|
"Image '.*' has faked external blobs and is non-functional: .*")
|
|
|
|
|
2022-03-28 20:57:04 +00:00
|
|
|
def testPreLoad(self):
|
|
|
|
"""Test an image with a pre-load header"""
|
|
|
|
entry_args = {
|
|
|
|
'pre-load-key-path': '.',
|
|
|
|
}
|
2022-08-23 10:46:09 +00:00
|
|
|
data, _, _, _ = self._DoReadFileDtb('230_pre_load.dts',
|
2022-03-28 20:57:04 +00:00
|
|
|
entry_args=entry_args)
|
|
|
|
self.assertEqual(PRE_LOAD_MAGIC, data[:len(PRE_LOAD_MAGIC)])
|
|
|
|
self.assertEqual(PRE_LOAD_VERSION, data[4:4 + len(PRE_LOAD_VERSION)])
|
|
|
|
self.assertEqual(PRE_LOAD_HDR_SIZE, data[8:8 + len(PRE_LOAD_HDR_SIZE)])
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('230_pre_load.dts')
|
2022-03-28 20:57:04 +00:00
|
|
|
self.assertEqual(PRE_LOAD_MAGIC, data[:len(PRE_LOAD_MAGIC)])
|
|
|
|
self.assertEqual(PRE_LOAD_VERSION, data[4:4 + len(PRE_LOAD_VERSION)])
|
|
|
|
self.assertEqual(PRE_LOAD_HDR_SIZE, data[8:8 + len(PRE_LOAD_HDR_SIZE)])
|
|
|
|
|
|
|
|
def testPreLoadPkcs(self):
|
|
|
|
"""Test an image with a pre-load header with padding pkcs"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('231_pre_load_pkcs.dts')
|
2022-03-28 20:57:04 +00:00
|
|
|
self.assertEqual(PRE_LOAD_MAGIC, data[:len(PRE_LOAD_MAGIC)])
|
|
|
|
self.assertEqual(PRE_LOAD_VERSION, data[4:4 + len(PRE_LOAD_VERSION)])
|
|
|
|
self.assertEqual(PRE_LOAD_HDR_SIZE, data[8:8 + len(PRE_LOAD_HDR_SIZE)])
|
|
|
|
|
|
|
|
def testPreLoadPss(self):
|
|
|
|
"""Test an image with a pre-load header with padding pss"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('232_pre_load_pss.dts')
|
2022-03-28 20:57:04 +00:00
|
|
|
self.assertEqual(PRE_LOAD_MAGIC, data[:len(PRE_LOAD_MAGIC)])
|
|
|
|
self.assertEqual(PRE_LOAD_VERSION, data[4:4 + len(PRE_LOAD_VERSION)])
|
|
|
|
self.assertEqual(PRE_LOAD_HDR_SIZE, data[8:8 + len(PRE_LOAD_HDR_SIZE)])
|
|
|
|
|
|
|
|
def testPreLoadInvalidPadding(self):
|
|
|
|
"""Test an image with a pre-load header with an invalid padding"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('233_pre_load_invalid_padding.dts')
|
2022-03-28 20:57:04 +00:00
|
|
|
|
|
|
|
def testPreLoadInvalidSha(self):
|
|
|
|
"""Test an image with a pre-load header with an invalid hash"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('234_pre_load_invalid_sha.dts')
|
2022-03-28 20:57:04 +00:00
|
|
|
|
|
|
|
def testPreLoadInvalidAlgo(self):
|
|
|
|
"""Test an image with a pre-load header with an invalid algo"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('235_pre_load_invalid_algo.dts')
|
2022-03-28 20:57:04 +00:00
|
|
|
|
|
|
|
def testPreLoadInvalidKey(self):
|
|
|
|
"""Test an image with a pre-load header with an invalid key"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('236_pre_load_invalid_key.dts')
|
2022-02-19 18:50:04 +00:00
|
|
|
|
2022-03-27 15:31:44 +00:00
|
|
|
def _CheckSafeUniqueNames(self, *images):
|
|
|
|
"""Check all entries of given images for unsafe unique names"""
|
|
|
|
for image in images:
|
|
|
|
entries = {}
|
|
|
|
image._CollectEntries(entries, {}, image)
|
|
|
|
for entry in entries.values():
|
|
|
|
uniq = entry.GetUniqueName()
|
|
|
|
|
|
|
|
# Used as part of a filename, so must not be absolute paths.
|
|
|
|
self.assertFalse(os.path.isabs(uniq))
|
|
|
|
|
|
|
|
def testSafeUniqueNames(self):
|
|
|
|
"""Test entry unique names are safe in single image configuration"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFileRealDtb('237_unique_names.dts')
|
2022-03-27 15:31:44 +00:00
|
|
|
|
|
|
|
orig_image = control.images['image']
|
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
|
|
|
|
self._CheckSafeUniqueNames(orig_image, image)
|
|
|
|
|
|
|
|
def testSafeUniqueNamesMulti(self):
|
|
|
|
"""Test entry unique names are safe with multiple images"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFileRealDtb('238_unique_names_multi.dts')
|
2022-03-27 15:31:44 +00:00
|
|
|
|
|
|
|
orig_image = control.images['image']
|
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
image = Image.FromFile(image_fname)
|
|
|
|
|
|
|
|
self._CheckSafeUniqueNames(orig_image, image)
|
|
|
|
|
2022-03-27 15:31:45 +00:00
|
|
|
def testReplaceCmdWithBintool(self):
|
|
|
|
"""Test replacing an entry that needs a bintool to pack"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFileRealDtb('239_replace_with_bintool.dts')
|
2022-03-27 15:31:45 +00:00
|
|
|
expected = U_BOOT_DATA + b'aa'
|
|
|
|
self.assertEqual(expected, data[:len(expected)])
|
|
|
|
|
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
fname = os.path.join(tmpdir, 'update-testing.bin')
|
|
|
|
tools.write_file(fname, b'zz')
|
|
|
|
self._DoBinman('replace', '-i', updated_fname,
|
|
|
|
'_testing', '-f', fname)
|
|
|
|
|
|
|
|
data = tools.read_file(updated_fname)
|
|
|
|
expected = U_BOOT_DATA + b'zz'
|
|
|
|
self.assertEqual(expected, data[:len(expected)])
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
|
|
|
|
|
|
|
def testReplaceCmdOtherWithBintool(self):
|
|
|
|
"""Test replacing an entry when another needs a bintool to pack"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFileRealDtb('239_replace_with_bintool.dts')
|
2022-03-27 15:31:45 +00:00
|
|
|
expected = U_BOOT_DATA + b'aa'
|
|
|
|
self.assertEqual(expected, data[:len(expected)])
|
|
|
|
|
|
|
|
try:
|
|
|
|
tmpdir, updated_fname = self._SetupImageInTmpdir()
|
|
|
|
fname = os.path.join(tmpdir, 'update-u-boot.bin')
|
|
|
|
tools.write_file(fname, b'x' * len(U_BOOT_DATA))
|
|
|
|
self._DoBinman('replace', '-i', updated_fname,
|
|
|
|
'u-boot', '-f', fname)
|
|
|
|
|
|
|
|
data = tools.read_file(updated_fname)
|
|
|
|
expected = b'x' * len(U_BOOT_DATA) + b'aa'
|
|
|
|
self.assertEqual(expected, data[:len(expected)])
|
|
|
|
finally:
|
|
|
|
shutil.rmtree(tmpdir)
|
|
|
|
|
2022-03-27 15:31:46 +00:00
|
|
|
def testReplaceResizeNoRepackSameSize(self):
|
|
|
|
"""Test replacing entries with same-size data without repacking"""
|
|
|
|
expected = b'x' * len(U_BOOT_DATA)
|
|
|
|
data, expected_fdtmap, _ = self._RunReplaceCmd('u-boot', expected)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
path, fdtmap = state.GetFdtContents('fdtmap')
|
|
|
|
self.assertIsNotNone(path)
|
|
|
|
self.assertEqual(expected_fdtmap, fdtmap)
|
|
|
|
|
|
|
|
def testReplaceResizeNoRepackSmallerSize(self):
|
|
|
|
"""Test replacing entries with smaller-size data without repacking"""
|
|
|
|
new_data = b'x'
|
|
|
|
data, expected_fdtmap, _ = self._RunReplaceCmd('u-boot', new_data)
|
|
|
|
expected = new_data.ljust(len(U_BOOT_DATA), b'\0')
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
path, fdtmap = state.GetFdtContents('fdtmap')
|
|
|
|
self.assertIsNotNone(path)
|
|
|
|
self.assertEqual(expected_fdtmap, fdtmap)
|
|
|
|
|
2022-03-27 15:31:48 +00:00
|
|
|
def testExtractFit(self):
|
|
|
|
"""Test extracting a FIT section"""
|
2022-08-23 10:46:09 +00:00
|
|
|
self._DoReadFileRealDtb('240_fit_extract_replace.dts')
|
2022-03-27 15:31:48 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
|
|
|
|
fit_data = control.ReadEntry(image_fname, 'fit')
|
|
|
|
fit = fdt.Fdt.FromData(fit_data)
|
|
|
|
fit.Scan()
|
|
|
|
|
|
|
|
# Check subentry data inside the extracted fit
|
|
|
|
for node_path, expected in [
|
|
|
|
('/images/kernel', U_BOOT_DATA),
|
|
|
|
('/images/fdt-1', U_BOOT_NODTB_DATA),
|
|
|
|
('/images/scr-1', COMPRESS_DATA),
|
|
|
|
]:
|
|
|
|
node = fit.GetNode(node_path)
|
|
|
|
data = fit.GetProps(node)['data'].bytes
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
def testExtractFitSubentries(self):
|
|
|
|
"""Test extracting FIT section subentries"""
|
2022-08-23 10:46:09 +00:00
|
|
|
self._DoReadFileRealDtb('240_fit_extract_replace.dts')
|
2022-03-27 15:31:48 +00:00
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
|
|
|
|
for entry_path, expected in [
|
|
|
|
('fit/kernel', U_BOOT_DATA),
|
|
|
|
('fit/kernel/u-boot', U_BOOT_DATA),
|
|
|
|
('fit/fdt-1', U_BOOT_NODTB_DATA),
|
|
|
|
('fit/fdt-1/u-boot-nodtb', U_BOOT_NODTB_DATA),
|
|
|
|
('fit/scr-1', COMPRESS_DATA),
|
|
|
|
('fit/scr-1/blob', COMPRESS_DATA),
|
|
|
|
]:
|
|
|
|
data = control.ReadEntry(image_fname, entry_path)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2022-03-27 15:31:49 +00:00
|
|
|
def testReplaceFitSubentryLeafSameSize(self):
|
|
|
|
"""Test replacing a FIT leaf subentry with same-size data"""
|
|
|
|
new_data = b'x' * len(U_BOOT_DATA)
|
|
|
|
data, expected_fdtmap, _ = self._RunReplaceCmd(
|
|
|
|
'fit/kernel/u-boot', new_data,
|
2022-08-23 10:46:09 +00:00
|
|
|
dts='240_fit_extract_replace.dts')
|
2022-03-27 15:31:49 +00:00
|
|
|
self.assertEqual(new_data, data)
|
|
|
|
|
|
|
|
path, fdtmap = state.GetFdtContents('fdtmap')
|
|
|
|
self.assertIsNotNone(path)
|
|
|
|
self.assertEqual(expected_fdtmap, fdtmap)
|
|
|
|
|
|
|
|
def testReplaceFitSubentryLeafBiggerSize(self):
|
|
|
|
"""Test replacing a FIT leaf subentry with bigger-size data"""
|
|
|
|
new_data = b'ub' * len(U_BOOT_NODTB_DATA)
|
|
|
|
data, expected_fdtmap, _ = self._RunReplaceCmd(
|
|
|
|
'fit/fdt-1/u-boot-nodtb', new_data,
|
2022-08-23 10:46:09 +00:00
|
|
|
dts='240_fit_extract_replace.dts')
|
2022-03-27 15:31:49 +00:00
|
|
|
self.assertEqual(new_data, data)
|
|
|
|
|
|
|
|
# Will be repacked, so fdtmap must change
|
|
|
|
path, fdtmap = state.GetFdtContents('fdtmap')
|
|
|
|
self.assertIsNotNone(path)
|
|
|
|
self.assertNotEqual(expected_fdtmap, fdtmap)
|
|
|
|
|
|
|
|
def testReplaceFitSubentryLeafSmallerSize(self):
|
|
|
|
"""Test replacing a FIT leaf subentry with smaller-size data"""
|
|
|
|
new_data = b'x'
|
|
|
|
expected = new_data.ljust(len(U_BOOT_NODTB_DATA), b'\0')
|
|
|
|
data, expected_fdtmap, _ = self._RunReplaceCmd(
|
|
|
|
'fit/fdt-1/u-boot-nodtb', new_data,
|
2022-08-23 10:46:09 +00:00
|
|
|
dts='240_fit_extract_replace.dts')
|
2022-03-27 15:31:49 +00:00
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
path, fdtmap = state.GetFdtContents('fdtmap')
|
|
|
|
self.assertIsNotNone(path)
|
|
|
|
self.assertEqual(expected_fdtmap, fdtmap)
|
|
|
|
|
2022-03-27 15:31:50 +00:00
|
|
|
def testReplaceSectionSimple(self):
|
|
|
|
"""Test replacing a simple section with arbitrary data"""
|
|
|
|
new_data = b'w' * len(COMPRESS_DATA + U_BOOT_DATA)
|
2022-08-13 17:40:46 +00:00
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self._RunReplaceCmd('section', new_data,
|
2022-08-23 10:46:09 +00:00
|
|
|
dts='241_replace_section_simple.dts')
|
2022-08-13 17:40:46 +00:00
|
|
|
self.assertIn(
|
|
|
|
"Node '/section': Replacing sections is not implemented yet",
|
|
|
|
str(exc.exception))
|
2022-03-27 15:31:50 +00:00
|
|
|
|
2022-08-13 17:40:48 +00:00
|
|
|
def testMkimageImagename(self):
|
|
|
|
"""Test using mkimage with -n holding the data too"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('242_mkimage_name.dts')
|
2022-08-13 17:40:48 +00:00
|
|
|
|
|
|
|
# Check that the data appears in the file somewhere
|
|
|
|
self.assertIn(U_BOOT_SPL_DATA, data)
|
|
|
|
|
2022-09-07 02:26:52 +00:00
|
|
|
# Get struct legacy_img_hdr -> ih_name
|
2022-08-13 17:40:48 +00:00
|
|
|
name = data[0x20:0x40]
|
|
|
|
|
|
|
|
# Build the filename that we expect to be placed in there, by virtue of
|
|
|
|
# the -n paraameter
|
|
|
|
expect = os.path.join(tools.get_output_dir(), 'mkimage.mkimage')
|
|
|
|
|
|
|
|
# Check that the image name is set to the temporary filename used
|
|
|
|
self.assertEqual(expect.encode('utf-8')[:0x20], name)
|
|
|
|
|
2022-08-13 17:40:49 +00:00
|
|
|
def testMkimageImage(self):
|
|
|
|
"""Test using mkimage with -n holding the data too"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('243_mkimage_image.dts')
|
2022-08-13 17:40:49 +00:00
|
|
|
|
|
|
|
# Check that the data appears in the file somewhere
|
|
|
|
self.assertIn(U_BOOT_SPL_DATA, data)
|
|
|
|
|
2022-09-07 02:26:52 +00:00
|
|
|
# Get struct legacy_img_hdr -> ih_name
|
2022-08-13 17:40:49 +00:00
|
|
|
name = data[0x20:0x40]
|
|
|
|
|
|
|
|
# Build the filename that we expect to be placed in there, by virtue of
|
|
|
|
# the -n paraameter
|
|
|
|
expect = os.path.join(tools.get_output_dir(), 'mkimage-n.mkimage')
|
|
|
|
|
|
|
|
# Check that the image name is set to the temporary filename used
|
|
|
|
self.assertEqual(expect.encode('utf-8')[:0x20], name)
|
|
|
|
|
|
|
|
# Check the corect data is in the imagename file
|
|
|
|
self.assertEqual(U_BOOT_DATA, tools.read_file(expect))
|
|
|
|
|
|
|
|
def testMkimageImageNoContent(self):
|
|
|
|
"""Test using mkimage with -n and no data"""
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
2022-08-23 10:46:09 +00:00
|
|
|
self._DoReadFile('244_mkimage_image_no_content.dts')
|
2022-08-13 17:40:49 +00:00
|
|
|
self.assertIn('Could not complete processing of contents',
|
|
|
|
str(exc.exception))
|
|
|
|
|
|
|
|
def testMkimageImageBad(self):
|
|
|
|
"""Test using mkimage with imagename node and data-to-imagename"""
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
2022-08-23 10:46:09 +00:00
|
|
|
self._DoReadFile('245_mkimage_image_bad.dts')
|
2022-08-13 17:40:49 +00:00
|
|
|
self.assertIn('Cannot use both imagename node and data-to-imagename',
|
|
|
|
str(exc.exception))
|
|
|
|
|
2022-08-13 17:40:50 +00:00
|
|
|
def testCollectionOther(self):
|
|
|
|
"""Test a collection where the data comes from another section"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('246_collection_other.dts')
|
2022-08-13 17:40:50 +00:00
|
|
|
self.assertEqual(U_BOOT_NODTB_DATA + U_BOOT_DTB_DATA +
|
|
|
|
tools.get_bytes(0xff, 2) + U_BOOT_NODTB_DATA +
|
|
|
|
tools.get_bytes(0xfe, 3) + U_BOOT_DTB_DATA,
|
|
|
|
data)
|
|
|
|
|
|
|
|
def testMkimageCollection(self):
|
|
|
|
"""Test using a collection referring to an entry in a mkimage entry"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFile('247_mkimage_coll.dts')
|
2022-08-13 17:40:50 +00:00
|
|
|
expect = U_BOOT_SPL_DATA + U_BOOT_DATA
|
|
|
|
self.assertEqual(expect, data[:len(expect)])
|
|
|
|
|
2022-08-19 14:25:25 +00:00
|
|
|
def testCompressDtbPrependInvalid(self):
|
|
|
|
"""Test that invalid header is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-08-23 10:46:09 +00:00
|
|
|
self._DoReadFileDtb('248_compress_dtb_prepend_invalid.dts')
|
2022-08-19 14:25:25 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-dtb': Invalid prepend in "
|
|
|
|
"'u-boot-dtb': 'invalid'", str(e.exception))
|
|
|
|
|
|
|
|
def testCompressDtbPrependLength(self):
|
|
|
|
"""Test that compress with length header works as expected"""
|
2022-08-23 10:46:09 +00:00
|
|
|
data = self._DoReadFileRealDtb('249_compress_dtb_prepend_length.dts')
|
2022-08-19 14:25:25 +00:00
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertIn('u-boot-dtb', entries)
|
|
|
|
u_boot_dtb = entries['u-boot-dtb']
|
|
|
|
self.assertIn('fdtmap', entries)
|
|
|
|
fdtmap = entries['fdtmap']
|
|
|
|
|
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
orig = control.ReadEntry(image_fname, 'u-boot-dtb')
|
|
|
|
dtb = fdt.Fdt.FromData(orig)
|
|
|
|
dtb.Scan()
|
|
|
|
props = self._GetPropTree(dtb, ['size', 'uncomp-size'])
|
|
|
|
expected = {
|
|
|
|
'u-boot:size': len(U_BOOT_DATA),
|
|
|
|
'u-boot-dtb:uncomp-size': len(orig),
|
|
|
|
'u-boot-dtb:size': u_boot_dtb.size,
|
|
|
|
'fdtmap:size': fdtmap.size,
|
|
|
|
'size': len(data),
|
|
|
|
}
|
|
|
|
self.assertEqual(expected, props)
|
|
|
|
|
|
|
|
# Check implementation
|
|
|
|
self.assertEqual(U_BOOT_DATA, data[:len(U_BOOT_DATA)])
|
|
|
|
rest = data[len(U_BOOT_DATA):]
|
|
|
|
comp_data_len = struct.unpack('<I', rest[:4])[0]
|
|
|
|
comp_data = rest[4:4 + comp_data_len]
|
|
|
|
orig2 = self._decompress(comp_data)
|
|
|
|
self.assertEqual(orig, orig2)
|
|
|
|
|
2022-08-19 14:25:30 +00:00
|
|
|
def testInvalidCompress(self):
|
|
|
|
"""Test that invalid compress algorithm is detected"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-08-23 10:46:09 +00:00
|
|
|
self._DoTestFile('250_compress_dtb_invalid.dts')
|
2022-08-19 14:25:30 +00:00
|
|
|
self.assertIn("Unknown algorithm 'invalid'", str(e.exception))
|
|
|
|
|
2022-08-19 14:25:32 +00:00
|
|
|
def testCompUtilCompressions(self):
|
|
|
|
"""Test compression algorithms"""
|
|
|
|
for bintool in self.comp_bintools.values():
|
|
|
|
self._CheckBintool(bintool)
|
|
|
|
data = bintool.compress(COMPRESS_DATA)
|
|
|
|
self.assertNotEqual(COMPRESS_DATA, data)
|
|
|
|
orig = bintool.decompress(data)
|
|
|
|
self.assertEquals(COMPRESS_DATA, orig)
|
|
|
|
|
|
|
|
def testCompUtilVersions(self):
|
|
|
|
"""Test tool version of compression algorithms"""
|
|
|
|
for bintool in self.comp_bintools.values():
|
|
|
|
self._CheckBintool(bintool)
|
|
|
|
version = bintool.version()
|
|
|
|
self.assertRegex(version, '^v?[0-9]+[0-9.]*')
|
|
|
|
|
|
|
|
def testCompUtilPadding(self):
|
|
|
|
"""Test padding of compression algorithms"""
|
2022-08-19 14:25:38 +00:00
|
|
|
# Skip zstd because it doesn't support padding
|
|
|
|
for bintool in [v for k,v in self.comp_bintools.items() if k != 'zstd']:
|
2022-08-19 14:25:32 +00:00
|
|
|
self._CheckBintool(bintool)
|
|
|
|
data = bintool.compress(COMPRESS_DATA)
|
|
|
|
self.assertNotEqual(COMPRESS_DATA, data)
|
|
|
|
data += tools.get_bytes(0, 64)
|
|
|
|
orig = bintool.decompress(data)
|
|
|
|
self.assertEquals(COMPRESS_DATA, orig)
|
|
|
|
|
2022-08-19 14:25:38 +00:00
|
|
|
def testCompressDtbZstd(self):
|
|
|
|
"""Test that zstd compress of device-tree files failed"""
|
|
|
|
with self.assertRaises(ValueError) as e:
|
2022-08-23 10:46:09 +00:00
|
|
|
self._DoTestFile('251_compress_dtb_zstd.dts')
|
2022-08-19 14:25:38 +00:00
|
|
|
self.assertIn("Node '/binman/u-boot-dtb': The zstd compression "
|
|
|
|
"requires a length header", str(e.exception))
|
|
|
|
|
2022-09-02 13:10:48 +00:00
|
|
|
def testMkimageMultipleDataFiles(self):
|
|
|
|
"""Test passing multiple files to mkimage in a mkimage entry"""
|
|
|
|
data = self._DoReadFile('252_mkimage_mult_data.dts')
|
|
|
|
# Size of files are packed in their 4B big-endian format
|
|
|
|
expect = struct.pack('>I', len(U_BOOT_TPL_DATA))
|
|
|
|
expect += struct.pack('>I', len(U_BOOT_SPL_DATA))
|
|
|
|
# Size info is always followed by a 4B zero value.
|
|
|
|
expect += tools.get_bytes(0, 4)
|
|
|
|
expect += U_BOOT_TPL_DATA
|
|
|
|
# All but last files are 4B-aligned
|
|
|
|
align_pad = len(U_BOOT_TPL_DATA) % 4
|
|
|
|
if align_pad:
|
|
|
|
expect += tools.get_bytes(0, align_pad)
|
|
|
|
expect += U_BOOT_SPL_DATA
|
|
|
|
self.assertEqual(expect, data[-len(expect):])
|
|
|
|
|
|
|
|
def testMkimageMultipleNoContent(self):
|
|
|
|
"""Test passing multiple data files to mkimage with one data file having no content"""
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self._DoReadFile('253_mkimage_mult_no_content.dts')
|
|
|
|
self.assertIn('Could not complete processing of contents',
|
|
|
|
str(exc.exception))
|
|
|
|
|
2022-09-02 13:10:49 +00:00
|
|
|
def testMkimageFilename(self):
|
|
|
|
"""Test using mkimage to build a binary with a filename"""
|
|
|
|
retcode = self._DoTestFile('254_mkimage_filename.dts')
|
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
fname = tools.get_output_filename('mkimage-test.bin')
|
|
|
|
self.assertTrue(os.path.exists(fname))
|
|
|
|
|
2022-02-28 14:16:54 +00:00
|
|
|
def testVpl(self):
|
|
|
|
"""Test that an image with VPL and its device tree can be created"""
|
|
|
|
# ELF file with a '__bss_size' symbol
|
|
|
|
self._SetupVplElf()
|
|
|
|
data = self._DoReadFile('255_u_boot_vpl.dts')
|
|
|
|
self.assertEqual(U_BOOT_VPL_DATA + U_BOOT_VPL_DTB_DATA, data)
|
|
|
|
|
|
|
|
def testVplNoDtb(self):
|
|
|
|
"""Test that an image with vpl/u-boot-vpl-nodtb.bin can be created"""
|
|
|
|
self._SetupVplElf()
|
|
|
|
data = self._DoReadFile('256_u_boot_vpl_nodtb.dts')
|
|
|
|
self.assertEqual(U_BOOT_VPL_NODTB_DATA,
|
|
|
|
data[:len(U_BOOT_VPL_NODTB_DATA)])
|
|
|
|
|
|
|
|
def testExpandedVpl(self):
|
|
|
|
"""Test that an expanded entry type is selected for TPL when needed"""
|
|
|
|
self._SetupVplElf()
|
|
|
|
|
|
|
|
entry_args = {
|
|
|
|
'vpl-bss-pad': 'y',
|
|
|
|
'vpl-dtb': 'y',
|
|
|
|
}
|
|
|
|
self._DoReadFileDtb('257_fdt_incl_vpl.dts', use_expanded=True,
|
|
|
|
entry_args=entry_args)
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
self.assertEqual(1, len(entries))
|
|
|
|
|
|
|
|
# We only have u-boot-vpl, which be expanded
|
|
|
|
self.assertIn('u-boot-vpl', entries)
|
|
|
|
entry = entries['u-boot-vpl']
|
|
|
|
self.assertEqual('u-boot-vpl-expanded', entry.etype)
|
|
|
|
subent = entry.GetEntries()
|
|
|
|
self.assertEqual(3, len(subent))
|
|
|
|
self.assertIn('u-boot-vpl-nodtb', subent)
|
|
|
|
self.assertIn('u-boot-vpl-bss-pad', subent)
|
|
|
|
self.assertIn('u-boot-vpl-dtb', subent)
|
|
|
|
|
|
|
|
def testVplBssPadMissing(self):
|
|
|
|
"""Test that a missing symbol is detected"""
|
|
|
|
self._SetupVplElf('u_boot_ucode_ptr')
|
|
|
|
with self.assertRaises(ValueError) as e:
|
|
|
|
self._DoReadFile('258_vpl_bss_pad.dts')
|
|
|
|
self.assertIn('Expected __bss_size symbol in vpl/u-boot-vpl',
|
|
|
|
str(e.exception))
|
|
|
|
|
2022-10-17 11:06:25 +00:00
|
|
|
def testSymlink(self):
|
|
|
|
"""Test that image files can be named"""
|
|
|
|
retcode = self._DoTestFile('259_symlink.dts', debug=True, map=True)
|
|
|
|
self.assertEqual(0, retcode)
|
|
|
|
image = control.images['test_image']
|
|
|
|
fname = tools.get_output_filename('test_image.bin')
|
|
|
|
sname = tools.get_output_filename('symlink_to_test.bin')
|
|
|
|
self.assertTrue(os.path.islink(sname))
|
|
|
|
self.assertEqual(os.readlink(sname), fname)
|
2022-03-27 15:31:45 +00:00
|
|
|
|
2022-10-21 00:22:47 +00:00
|
|
|
def testSymbolsElf(self):
|
|
|
|
"""Test binman can assign symbols embedded in an ELF file"""
|
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
|
|
|
self._SetupTplElf('u_boot_binman_syms')
|
|
|
|
self._SetupVplElf('u_boot_binman_syms')
|
|
|
|
self._SetupSplElf('u_boot_binman_syms')
|
|
|
|
data = self._DoReadFileDtb('260_symbols_elf.dts')[0]
|
|
|
|
image_fname = tools.get_output_filename('image.bin')
|
|
|
|
|
|
|
|
image = control.images['image']
|
|
|
|
entries = image.GetEntries()
|
|
|
|
|
|
|
|
for entry in entries.values():
|
|
|
|
# No symbols in u-boot and it has faked contents anyway
|
|
|
|
if entry.name == 'u-boot':
|
|
|
|
continue
|
|
|
|
edata = data[entry.image_pos:entry.image_pos + entry.size]
|
|
|
|
efname = tools.get_output_filename(f'edata-{entry.name}')
|
|
|
|
tools.write_file(efname, edata)
|
|
|
|
|
|
|
|
syms = elf.GetSymbolFileOffset(efname, ['_binman_u_boot'])
|
|
|
|
re_name = re.compile('_binman_(u_boot_(.*))_prop_(.*)')
|
|
|
|
for name, sym in syms.items():
|
|
|
|
msg = 'test'
|
|
|
|
val = elf.GetSymbolValue(sym, edata, msg)
|
|
|
|
entry_m = re_name.match(name)
|
|
|
|
if entry_m:
|
|
|
|
ename, prop = entry_m.group(1), entry_m.group(3)
|
|
|
|
entry, entry_name, prop_name = image.LookupEntry(entries,
|
|
|
|
name, msg)
|
|
|
|
if prop_name == 'offset':
|
|
|
|
expect_val = entry.offset
|
|
|
|
elif prop_name == 'image_pos':
|
|
|
|
expect_val = entry.image_pos
|
|
|
|
elif prop_name == 'size':
|
|
|
|
expect_val = entry.size
|
|
|
|
self.assertEqual(expect_val, val)
|
|
|
|
|
|
|
|
def testSymbolsElfBad(self):
|
|
|
|
"""Check error when trying to write symbols without the elftools lib"""
|
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
|
|
|
self._SetupTplElf('u_boot_binman_syms')
|
|
|
|
self._SetupVplElf('u_boot_binman_syms')
|
|
|
|
self._SetupSplElf('u_boot_binman_syms')
|
|
|
|
try:
|
|
|
|
elf.ELF_TOOLS = False
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self._DoReadFileDtb('260_symbols_elf.dts')
|
|
|
|
finally:
|
|
|
|
elf.ELF_TOOLS = True
|
|
|
|
self.assertIn(
|
|
|
|
"Section '/binman': entry '/binman/u-boot-spl-elf': "
|
|
|
|
'Cannot write symbols to an ELF file without Python elftools',
|
|
|
|
str(exc.exception))
|
|
|
|
|
2023-01-07 21:07:08 +00:00
|
|
|
def testSectionFilename(self):
|
|
|
|
"""Check writing of section contents to a file"""
|
|
|
|
data = self._DoReadFile('261_section_fname.dts')
|
|
|
|
expected = (b'&&' + U_BOOT_DATA + b'&&&' +
|
|
|
|
tools.get_bytes(ord('!'), 7) +
|
|
|
|
U_BOOT_DATA + tools.get_bytes(ord('&'), 12))
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
|
|
|
sect_fname = tools.get_output_filename('outfile.bin')
|
|
|
|
self.assertTrue(os.path.exists(sect_fname))
|
|
|
|
sect_data = tools.read_file(sect_fname)
|
|
|
|
self.assertEqual(U_BOOT_DATA, sect_data)
|
|
|
|
|
2023-01-07 21:07:12 +00:00
|
|
|
def testAbsent(self):
|
|
|
|
"""Check handling of absent entries"""
|
|
|
|
data = self._DoReadFile('262_absent.dts')
|
|
|
|
self.assertEqual(U_BOOT_DATA + U_BOOT_IMG_DATA, data)
|
|
|
|
|
2023-01-07 21:07:14 +00:00
|
|
|
def testPackTeeOsOptional(self):
|
|
|
|
"""Test that an image with an optional TEE binary can be created"""
|
|
|
|
entry_args = {
|
|
|
|
'tee-os-path': 'tee.elf',
|
|
|
|
}
|
|
|
|
data = self._DoReadFileDtb('263_tee_os_opt.dts',
|
|
|
|
entry_args=entry_args)[0]
|
|
|
|
self.assertEqual(U_BOOT_DATA + U_BOOT_IMG_DATA, data)
|
|
|
|
|
|
|
|
def checkFitTee(self, dts, tee_fname):
|
|
|
|
"""Check that a tee-os entry works and returns data
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dts (str): Device tree filename to use
|
|
|
|
tee_fname (str): filename containing tee-os
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
bytes: Image contents
|
|
|
|
"""
|
|
|
|
if not elf.ELF_TOOLS:
|
|
|
|
self.skipTest('Python elftools not available')
|
|
|
|
entry_args = {
|
|
|
|
'of-list': 'test-fdt1 test-fdt2',
|
|
|
|
'default-dt': 'test-fdt2',
|
|
|
|
'tee-os-path': tee_fname,
|
|
|
|
}
|
|
|
|
test_subdir = os.path.join(self._indir, TEST_FDT_SUBDIR)
|
|
|
|
data = self._DoReadFileDtb(dts, entry_args=entry_args,
|
|
|
|
extra_indirs=[test_subdir])[0]
|
|
|
|
return data
|
|
|
|
|
|
|
|
def testFitTeeOsOptionalFit(self):
|
|
|
|
"""Test an image with a FIT with an optional OP-TEE binary"""
|
|
|
|
data = self.checkFitTee('264_tee_os_opt_fit.dts', 'tee.bin')
|
|
|
|
|
|
|
|
# There should be only one node, holding the data set up in SetUpClass()
|
|
|
|
# for tee.bin
|
|
|
|
dtb = fdt.Fdt.FromData(data)
|
|
|
|
dtb.Scan()
|
|
|
|
node = dtb.GetNode('/images/tee-1')
|
|
|
|
self.assertEqual(TEE_ADDR,
|
|
|
|
fdt_util.fdt32_to_cpu(node.props['load'].value))
|
|
|
|
self.assertEqual(TEE_ADDR,
|
|
|
|
fdt_util.fdt32_to_cpu(node.props['entry'].value))
|
|
|
|
self.assertEqual(U_BOOT_DATA, node.props['data'].bytes)
|
|
|
|
|
|
|
|
def testFitTeeOsOptionalFitBad(self):
|
|
|
|
"""Test an image with a FIT with an optional OP-TEE binary"""
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self.checkFitTee('265_tee_os_opt_fit_bad.dts', 'tee.bin')
|
|
|
|
self.assertIn(
|
|
|
|
"Node '/binman/fit': subnode 'images/@tee-SEQ': Failed to read ELF file: Magic number does not match",
|
|
|
|
str(exc.exception))
|
|
|
|
|
|
|
|
def testFitTeeOsBad(self):
|
|
|
|
"""Test an OP-TEE binary with wrong formats"""
|
|
|
|
self.make_tee_bin('tee.bad1', 123)
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self.checkFitTee('264_tee_os_opt_fit.dts', 'tee.bad1')
|
|
|
|
self.assertIn(
|
|
|
|
"Node '/binman/fit/images/@tee-SEQ/tee-os': OP-TEE paged mode not supported",
|
|
|
|
str(exc.exception))
|
|
|
|
|
|
|
|
self.make_tee_bin('tee.bad2', 0, b'extra data')
|
|
|
|
with self.assertRaises(ValueError) as exc:
|
|
|
|
self.checkFitTee('264_tee_os_opt_fit.dts', 'tee.bad2')
|
|
|
|
self.assertIn(
|
|
|
|
"Node '/binman/fit/images/@tee-SEQ/tee-os': Invalid OP-TEE file: size mismatch (expected 0x4, have 0xe)",
|
|
|
|
str(exc.exception))
|
|
|
|
|
2023-01-07 21:07:15 +00:00
|
|
|
def testExtblobOptional(self):
|
|
|
|
"""Test an image with an external blob that is optional"""
|
|
|
|
with test_util.capture_sys_output() as (stdout, stderr):
|
|
|
|
data = self._DoReadFile('266_blob_ext_opt.dts')
|
|
|
|
self.assertEqual(REFCODE_DATA, data)
|
|
|
|
err = stderr.getvalue()
|
|
|
|
self.assertRegex(
|
|
|
|
err,
|
|
|
|
"Image '.*' is missing external blobs but is still functional: missing")
|
|
|
|
|
2023-01-11 23:10:12 +00:00
|
|
|
def testSectionInner(self):
|
|
|
|
"""Test an inner section with a size"""
|
|
|
|
data = self._DoReadFile('267_section_inner.dts')
|
|
|
|
expected = U_BOOT_DATA + tools.get_bytes(0, 12)
|
|
|
|
self.assertEqual(expected, data)
|
|
|
|
|
2023-01-11 23:10:14 +00:00
|
|
|
def testNull(self):
|
|
|
|
"""Test an image with a null entry"""
|
|
|
|
data = self._DoReadFile('268_null.dts')
|
|
|
|
self.assertEqual(U_BOOT_DATA + b'\xff\xff\xff\xff' + U_BOOT_IMG_DATA, data)
|
|
|
|
|
2023-01-07 21:07:08 +00:00
|
|
|
|
2017-11-13 04:52:22 +00:00
|
|
|
if __name__ == "__main__":
|
|
|
|
unittest.main()
|