2017-06-19 04:08:58 +00:00
|
|
|
#!/usr/bin/python
|
|
|
|
#
|
|
|
|
# Copyright (C) 2017 Google, Inc
|
|
|
|
# Written by Simon Glass <sjg@chromium.org>
|
|
|
|
#
|
|
|
|
# SPDX-License-Identifier: GPL-2.0+
|
|
|
|
#
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
"""Device tree to platform data class
|
|
|
|
|
|
|
|
This supports converting device tree data to C structures definitions and
|
|
|
|
static data.
|
|
|
|
"""
|
|
|
|
|
2017-06-19 04:08:58 +00:00
|
|
|
import copy
|
2017-06-19 04:08:59 +00:00
|
|
|
import sys
|
2017-06-19 04:08:58 +00:00
|
|
|
|
|
|
|
import fdt
|
|
|
|
import fdt_util
|
|
|
|
|
|
|
|
# When we see these properties we ignore them - i.e. do not create a structure member
|
|
|
|
PROP_IGNORE_LIST = [
|
|
|
|
'#address-cells',
|
|
|
|
'#gpio-cells',
|
|
|
|
'#size-cells',
|
|
|
|
'compatible',
|
|
|
|
'linux,phandle',
|
|
|
|
"status",
|
|
|
|
'phandle',
|
|
|
|
'u-boot,dm-pre-reloc',
|
|
|
|
'u-boot,dm-tpl',
|
|
|
|
'u-boot,dm-spl',
|
|
|
|
]
|
|
|
|
|
|
|
|
# C type declarations for the tyues we support
|
|
|
|
TYPE_NAMES = {
|
|
|
|
fdt.TYPE_INT: 'fdt32_t',
|
|
|
|
fdt.TYPE_BYTE: 'unsigned char',
|
|
|
|
fdt.TYPE_STRING: 'const char *',
|
|
|
|
fdt.TYPE_BOOL: 'bool',
|
2017-06-19 04:08:59 +00:00
|
|
|
}
|
2017-06-19 04:08:58 +00:00
|
|
|
|
|
|
|
STRUCT_PREFIX = 'dtd_'
|
|
|
|
VAL_PREFIX = 'dtv_'
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def conv_name_to_c(name):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Convert a device-tree name to a C identifier
|
|
|
|
|
2017-06-19 04:09:04 +00:00
|
|
|
This uses multiple replace() calls instead of re.sub() since it is faster
|
|
|
|
(400ms for 1m calls versus 1000ms for the 're' version).
|
|
|
|
|
2017-06-19 04:08:58 +00:00
|
|
|
Args:
|
|
|
|
name: Name to convert
|
|
|
|
Return:
|
|
|
|
String containing the C version of this name
|
|
|
|
"""
|
2017-06-19 04:08:59 +00:00
|
|
|
new = name.replace('@', '_at_')
|
|
|
|
new = new.replace('-', '_')
|
|
|
|
new = new.replace(',', '_')
|
|
|
|
new = new.replace('.', '_')
|
|
|
|
return new
|
|
|
|
|
|
|
|
def tab_to(num_tabs, line):
|
|
|
|
"""Append tabs to a line of text to reach a tab stop.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
num_tabs: Tab stop to obtain (0 = column 0, 1 = column 8, etc.)
|
|
|
|
line: Line of text to append to
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
line with the correct number of tabs appeneded. If the line already
|
|
|
|
extends past that tab stop then a single space is appended.
|
|
|
|
"""
|
|
|
|
if len(line) >= num_tabs * 8:
|
|
|
|
return line + ' '
|
|
|
|
return line + '\t' * (num_tabs - len(line) // 8)
|
|
|
|
|
2017-06-19 04:09:02 +00:00
|
|
|
def get_value(ftype, value):
|
|
|
|
"""Get a value as a C expression
|
|
|
|
|
|
|
|
For integers this returns a byte-swapped (little-endian) hex string
|
|
|
|
For bytes this returns a hex string, e.g. 0x12
|
|
|
|
For strings this returns a literal string enclosed in quotes
|
|
|
|
For booleans this return 'true'
|
|
|
|
|
|
|
|
Args:
|
|
|
|
type: Data type (fdt_util)
|
|
|
|
value: Data value, as a string of bytes
|
|
|
|
"""
|
|
|
|
if ftype == fdt.TYPE_INT:
|
|
|
|
return '%#x' % fdt_util.fdt32_to_cpu(value)
|
|
|
|
elif ftype == fdt.TYPE_BYTE:
|
|
|
|
return '%#x' % ord(value[0])
|
|
|
|
elif ftype == fdt.TYPE_STRING:
|
|
|
|
return '"%s"' % value
|
|
|
|
elif ftype == fdt.TYPE_BOOL:
|
|
|
|
return 'true'
|
|
|
|
|
|
|
|
def get_compat_name(node):
|
|
|
|
"""Get a node's first compatible string as a C identifier
|
|
|
|
|
|
|
|
Args:
|
|
|
|
node: Node object to check
|
|
|
|
Return:
|
|
|
|
Tuple:
|
|
|
|
C identifier for the first compatible string
|
|
|
|
List of C identifiers for all the other compatible strings
|
|
|
|
(possibly empty)
|
|
|
|
"""
|
|
|
|
compat = node.props['compatible'].value
|
|
|
|
aliases = []
|
|
|
|
if isinstance(compat, list):
|
|
|
|
compat, aliases = compat[0], compat[1:]
|
|
|
|
return conv_name_to_c(compat), [conv_name_to_c(a) for a in aliases]
|
|
|
|
|
|
|
|
def is_phandle(prop):
|
|
|
|
"""Check if a node contains phandles
|
|
|
|
|
|
|
|
We have no reliable way of detecting whether a node uses a phandle
|
|
|
|
or not. As an interim measure, use a list of known property names.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
prop: Prop object to check
|
|
|
|
Return:
|
|
|
|
True if the object value contains phandles, else False
|
|
|
|
"""
|
|
|
|
if prop.name in ['clocks']:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
class DtbPlatdata(object):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Provide a means to convert device tree binary data to platform data
|
|
|
|
|
|
|
|
The output of this process is C structures which can be used in space-
|
|
|
|
constrained encvironments where the ~3KB code overhead of device tree
|
|
|
|
code is not affordable.
|
|
|
|
|
|
|
|
Properties:
|
2017-06-19 04:08:59 +00:00
|
|
|
_fdt: Fdt object, referencing the device tree
|
2017-06-19 04:08:58 +00:00
|
|
|
_dtb_fname: Filename of the input device tree binary file
|
|
|
|
_valid_nodes: A list of Node object with compatible strings
|
2017-06-19 04:09:01 +00:00
|
|
|
_include_disabled: true to include nodes marked status = "disabled"
|
2017-06-19 04:08:59 +00:00
|
|
|
_phandle_nodes: A dict of nodes indexed by phandle number (1, 2...)
|
2017-06-19 04:08:58 +00:00
|
|
|
_outfile: The current output file (sys.stdout or a real file)
|
|
|
|
_lines: Stashed list of output lines for outputting in the future
|
2017-06-19 04:08:59 +00:00
|
|
|
_phandle_nodes: A dict of Nodes indexed by phandle (an integer)
|
2017-06-19 04:08:58 +00:00
|
|
|
"""
|
2017-06-19 04:09:01 +00:00
|
|
|
def __init__(self, dtb_fname, include_disabled):
|
2017-06-19 04:08:59 +00:00
|
|
|
self._fdt = None
|
2017-06-19 04:08:58 +00:00
|
|
|
self._dtb_fname = dtb_fname
|
|
|
|
self._valid_nodes = None
|
2017-06-19 04:09:01 +00:00
|
|
|
self._include_disabled = include_disabled
|
2017-06-19 04:08:59 +00:00
|
|
|
self._phandle_nodes = {}
|
2017-06-19 04:08:58 +00:00
|
|
|
self._outfile = None
|
|
|
|
self._lines = []
|
|
|
|
self._aliases = {}
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def setup_output(self, fname):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Set up the output destination
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
Once this is done, future calls to self.out() will output to this
|
2017-06-19 04:08:58 +00:00
|
|
|
file.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
fname: Filename to send output to, or '-' for stdout
|
|
|
|
"""
|
|
|
|
if fname == '-':
|
|
|
|
self._outfile = sys.stdout
|
|
|
|
else:
|
|
|
|
self._outfile = open(fname, 'w')
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def out(self, line):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Output a string to the output file
|
|
|
|
|
|
|
|
Args:
|
2017-06-19 04:08:59 +00:00
|
|
|
line: String to output
|
2017-06-19 04:08:58 +00:00
|
|
|
"""
|
2017-06-19 04:08:59 +00:00
|
|
|
self._outfile.write(line)
|
2017-06-19 04:08:58 +00:00
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def buf(self, line):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Buffer up a string to send later
|
|
|
|
|
|
|
|
Args:
|
2017-06-19 04:08:59 +00:00
|
|
|
line: String to add to our 'buffer' list
|
2017-06-19 04:08:58 +00:00
|
|
|
"""
|
2017-06-19 04:08:59 +00:00
|
|
|
self._lines.append(line)
|
2017-06-19 04:08:58 +00:00
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def get_buf(self):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Get the contents of the output buffer, and clear it
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The output buffer, which is then cleared for future use
|
|
|
|
"""
|
|
|
|
lines = self._lines
|
|
|
|
self._lines = []
|
|
|
|
return lines
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def scan_dtb(self):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Scan the device tree to obtain a tree of notes and properties
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
Once this is done, self._fdt.GetRoot() can be called to obtain the
|
2017-06-19 04:08:58 +00:00
|
|
|
device tree root node, and progress from there.
|
|
|
|
"""
|
2017-06-19 04:08:59 +00:00
|
|
|
self._fdt = fdt.FdtScan(self._dtb_fname)
|
|
|
|
|
|
|
|
def scan_node(self, root):
|
|
|
|
"""Scan a node and subnodes to build a tree of node and phandle info
|
|
|
|
|
|
|
|
This adds each node to self._valid_nodes and each phandle to
|
|
|
|
self._phandle_nodes.
|
2017-06-19 04:08:58 +00:00
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
Args:
|
|
|
|
root: Root node for scan
|
|
|
|
"""
|
2017-06-19 04:08:58 +00:00
|
|
|
for node in root.subnodes:
|
|
|
|
if 'compatible' in node.props:
|
|
|
|
status = node.props.get('status')
|
2017-06-19 04:09:01 +00:00
|
|
|
if (not self._include_disabled and not status or
|
2017-06-19 04:08:59 +00:00
|
|
|
status.value != 'disabled'):
|
2017-06-19 04:08:58 +00:00
|
|
|
self._valid_nodes.append(node)
|
|
|
|
phandle_prop = node.props.get('phandle')
|
|
|
|
if phandle_prop:
|
|
|
|
phandle = phandle_prop.GetPhandle()
|
2017-06-19 04:08:59 +00:00
|
|
|
self._phandle_nodes[phandle] = node
|
2017-06-19 04:08:58 +00:00
|
|
|
|
|
|
|
# recurse to handle any subnodes
|
2017-06-19 04:08:59 +00:00
|
|
|
self.scan_node(node)
|
2017-06-19 04:08:58 +00:00
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def scan_tree(self):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Scan the device tree for useful information
|
|
|
|
|
|
|
|
This fills in the following properties:
|
2017-06-19 04:08:59 +00:00
|
|
|
_phandle_nodes: A dict of Nodes indexed by phandle (an integer)
|
2017-06-19 04:08:58 +00:00
|
|
|
_valid_nodes: A list of nodes we wish to consider include in the
|
|
|
|
platform data
|
|
|
|
"""
|
2017-06-19 04:08:59 +00:00
|
|
|
self._phandle_nodes = {}
|
2017-06-19 04:08:58 +00:00
|
|
|
self._valid_nodes = []
|
2017-06-19 04:08:59 +00:00
|
|
|
return self.scan_node(self._fdt.GetRoot())
|
2017-06-19 04:08:58 +00:00
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def scan_structs(self):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Scan the device tree building up the C structures we will use.
|
|
|
|
|
|
|
|
Build a dict keyed by C struct name containing a dict of Prop
|
|
|
|
object for each struct field (keyed by property name). Where the
|
|
|
|
same struct appears multiple times, try to use the 'widest'
|
|
|
|
property, i.e. the one with a type which can express all others.
|
|
|
|
|
|
|
|
Once the widest property is determined, all other properties are
|
|
|
|
updated to match that width.
|
|
|
|
"""
|
|
|
|
structs = {}
|
|
|
|
for node in self._valid_nodes:
|
2017-06-19 04:09:02 +00:00
|
|
|
node_name, _ = get_compat_name(node)
|
2017-06-19 04:08:58 +00:00
|
|
|
fields = {}
|
|
|
|
|
|
|
|
# Get a list of all the valid properties in this node.
|
|
|
|
for name, prop in node.props.items():
|
|
|
|
if name not in PROP_IGNORE_LIST and name[0] != '#':
|
|
|
|
fields[name] = copy.deepcopy(prop)
|
|
|
|
|
|
|
|
# If we've seen this node_name before, update the existing struct.
|
|
|
|
if node_name in structs:
|
|
|
|
struct = structs[node_name]
|
|
|
|
for name, prop in fields.items():
|
|
|
|
oldprop = struct.get(name)
|
|
|
|
if oldprop:
|
|
|
|
oldprop.Widen(prop)
|
|
|
|
else:
|
|
|
|
struct[name] = prop
|
|
|
|
|
|
|
|
# Otherwise store this as a new struct.
|
|
|
|
else:
|
|
|
|
structs[node_name] = fields
|
|
|
|
|
|
|
|
upto = 0
|
|
|
|
for node in self._valid_nodes:
|
2017-06-19 04:09:02 +00:00
|
|
|
node_name, _ = get_compat_name(node)
|
2017-06-19 04:08:58 +00:00
|
|
|
struct = structs[node_name]
|
|
|
|
for name, prop in node.props.items():
|
|
|
|
if name not in PROP_IGNORE_LIST and name[0] != '#':
|
|
|
|
prop.Widen(struct[name])
|
|
|
|
upto += 1
|
|
|
|
|
2017-06-19 04:09:02 +00:00
|
|
|
struct_name, aliases = get_compat_name(node)
|
2017-06-19 04:08:58 +00:00
|
|
|
for alias in aliases:
|
|
|
|
self._aliases[alias] = struct_name
|
|
|
|
|
|
|
|
return structs
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def scan_phandles(self):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Figure out what phandles each node uses
|
|
|
|
|
|
|
|
We need to be careful when outputing nodes that use phandles since
|
|
|
|
they must come after the declaration of the phandles in the C file.
|
|
|
|
Otherwise we get a compiler error since the phandle struct is not yet
|
|
|
|
declared.
|
|
|
|
|
|
|
|
This function adds to each node a list of phandle nodes that the node
|
|
|
|
depends on. This allows us to output things in the right order.
|
|
|
|
"""
|
|
|
|
for node in self._valid_nodes:
|
|
|
|
node.phandles = set()
|
|
|
|
for pname, prop in node.props.items():
|
|
|
|
if pname in PROP_IGNORE_LIST or pname[0] == '#':
|
|
|
|
continue
|
2017-06-19 04:08:59 +00:00
|
|
|
if isinstance(prop.value, list):
|
2017-06-19 04:09:02 +00:00
|
|
|
if is_phandle(prop):
|
2017-06-19 04:08:58 +00:00
|
|
|
# Process the list as pairs of (phandle, id)
|
2017-06-19 04:08:59 +00:00
|
|
|
value_it = iter(prop.value)
|
|
|
|
for phandle_cell, _ in zip(value_it, value_it):
|
2017-06-19 04:08:58 +00:00
|
|
|
phandle = fdt_util.fdt32_to_cpu(phandle_cell)
|
2017-06-19 04:08:59 +00:00
|
|
|
target_node = self._phandle_nodes[phandle]
|
2017-06-19 04:08:58 +00:00
|
|
|
node.phandles.add(target_node)
|
|
|
|
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def generate_structs(self, structs):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Generate struct defintions for the platform data
|
|
|
|
|
|
|
|
This writes out the body of a header file consisting of structure
|
|
|
|
definitions for node in self._valid_nodes. See the documentation in
|
|
|
|
README.of-plat for more information.
|
|
|
|
"""
|
2017-06-19 04:08:59 +00:00
|
|
|
self.out('#include <stdbool.h>\n')
|
|
|
|
self.out('#include <libfdt.h>\n')
|
2017-06-19 04:08:58 +00:00
|
|
|
|
|
|
|
# Output the struct definition
|
|
|
|
for name in sorted(structs):
|
2017-06-19 04:08:59 +00:00
|
|
|
self.out('struct %s%s {\n' % (STRUCT_PREFIX, name))
|
2017-06-19 04:08:58 +00:00
|
|
|
for pname in sorted(structs[name]):
|
|
|
|
prop = structs[name][pname]
|
2017-06-19 04:09:02 +00:00
|
|
|
if is_phandle(prop):
|
2017-06-19 04:08:58 +00:00
|
|
|
# For phandles, include a reference to the target
|
2017-06-19 04:08:59 +00:00
|
|
|
self.out('\t%s%s[%d]' % (tab_to(2, 'struct phandle_2_cell'),
|
|
|
|
conv_name_to_c(prop.name),
|
2017-06-19 04:08:58 +00:00
|
|
|
len(prop.value) / 2))
|
|
|
|
else:
|
|
|
|
ptype = TYPE_NAMES[prop.type]
|
2017-06-19 04:08:59 +00:00
|
|
|
self.out('\t%s%s' % (tab_to(2, ptype),
|
|
|
|
conv_name_to_c(prop.name)))
|
|
|
|
if isinstance(prop.value, list):
|
|
|
|
self.out('[%d]' % len(prop.value))
|
|
|
|
self.out(';\n')
|
|
|
|
self.out('};\n')
|
2017-06-19 04:08:58 +00:00
|
|
|
|
|
|
|
for alias, struct_name in self._aliases.iteritems():
|
2017-06-19 04:08:59 +00:00
|
|
|
self.out('#define %s%s %s%s\n'% (STRUCT_PREFIX, alias,
|
2017-06-19 04:08:58 +00:00
|
|
|
STRUCT_PREFIX, struct_name))
|
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def output_node(self, node):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Output the C code for a node
|
|
|
|
|
|
|
|
Args:
|
|
|
|
node: node to output
|
|
|
|
"""
|
2017-06-19 04:09:02 +00:00
|
|
|
struct_name, _ = get_compat_name(node)
|
2017-06-19 04:08:59 +00:00
|
|
|
var_name = conv_name_to_c(node.name)
|
|
|
|
self.buf('static struct %s%s %s%s = {\n' %
|
|
|
|
(STRUCT_PREFIX, struct_name, VAL_PREFIX, var_name))
|
2017-06-19 04:08:58 +00:00
|
|
|
for pname, prop in node.props.items():
|
|
|
|
if pname in PROP_IGNORE_LIST or pname[0] == '#':
|
|
|
|
continue
|
2017-06-19 04:08:59 +00:00
|
|
|
member_name = conv_name_to_c(prop.name)
|
|
|
|
self.buf('\t%s= ' % tab_to(3, '.' + member_name))
|
2017-06-19 04:08:58 +00:00
|
|
|
|
|
|
|
# Special handling for lists
|
2017-06-19 04:08:59 +00:00
|
|
|
if isinstance(prop.value, list):
|
|
|
|
self.buf('{')
|
2017-06-19 04:08:58 +00:00
|
|
|
vals = []
|
|
|
|
# For phandles, output a reference to the platform data
|
|
|
|
# of the target node.
|
2017-06-19 04:09:02 +00:00
|
|
|
if is_phandle(prop):
|
2017-06-19 04:08:58 +00:00
|
|
|
# Process the list as pairs of (phandle, id)
|
2017-06-19 04:08:59 +00:00
|
|
|
value_it = iter(prop.value)
|
|
|
|
for phandle_cell, id_cell in zip(value_it, value_it):
|
2017-06-19 04:08:58 +00:00
|
|
|
phandle = fdt_util.fdt32_to_cpu(phandle_cell)
|
2017-06-19 04:08:59 +00:00
|
|
|
id_num = fdt_util.fdt32_to_cpu(id_cell)
|
|
|
|
target_node = self._phandle_nodes[phandle]
|
|
|
|
name = conv_name_to_c(target_node.name)
|
|
|
|
vals.append('{&%s%s, %d}' % (VAL_PREFIX, name, id_num))
|
2017-06-19 04:08:58 +00:00
|
|
|
else:
|
|
|
|
for val in prop.value:
|
2017-06-19 04:09:02 +00:00
|
|
|
vals.append(get_value(prop.type, val))
|
2017-06-19 04:08:59 +00:00
|
|
|
self.buf(', '.join(vals))
|
|
|
|
self.buf('}')
|
2017-06-19 04:08:58 +00:00
|
|
|
else:
|
2017-06-19 04:09:02 +00:00
|
|
|
self.buf(get_value(prop.type, prop.value))
|
2017-06-19 04:08:59 +00:00
|
|
|
self.buf(',\n')
|
|
|
|
self.buf('};\n')
|
2017-06-19 04:08:58 +00:00
|
|
|
|
|
|
|
# Add a device declaration
|
2017-06-19 04:08:59 +00:00
|
|
|
self.buf('U_BOOT_DEVICE(%s) = {\n' % var_name)
|
|
|
|
self.buf('\t.name\t\t= "%s",\n' % struct_name)
|
|
|
|
self.buf('\t.platdata\t= &%s%s,\n' % (VAL_PREFIX, var_name))
|
|
|
|
self.buf('\t.platdata_size\t= sizeof(%s%s),\n' % (VAL_PREFIX, var_name))
|
|
|
|
self.buf('};\n')
|
|
|
|
self.buf('\n')
|
2017-06-19 04:08:58 +00:00
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
self.out(''.join(self.get_buf()))
|
2017-06-19 04:08:58 +00:00
|
|
|
|
2017-06-19 04:08:59 +00:00
|
|
|
def generate_tables(self):
|
2017-06-19 04:08:58 +00:00
|
|
|
"""Generate device defintions for the platform data
|
|
|
|
|
|
|
|
This writes out C platform data initialisation data and
|
|
|
|
U_BOOT_DEVICE() declarations for each valid node. Where a node has
|
|
|
|
multiple compatible strings, a #define is used to make them equivalent.
|
|
|
|
|
|
|
|
See the documentation in doc/driver-model/of-plat.txt for more
|
|
|
|
information.
|
|
|
|
"""
|
2017-06-19 04:08:59 +00:00
|
|
|
self.out('#include <common.h>\n')
|
|
|
|
self.out('#include <dm.h>\n')
|
|
|
|
self.out('#include <dt-structs.h>\n')
|
|
|
|
self.out('\n')
|
2017-06-19 04:08:58 +00:00
|
|
|
nodes_to_output = list(self._valid_nodes)
|
|
|
|
|
|
|
|
# Keep outputing nodes until there is none left
|
|
|
|
while nodes_to_output:
|
|
|
|
node = nodes_to_output[0]
|
|
|
|
# Output all the node's dependencies first
|
|
|
|
for req_node in node.phandles:
|
|
|
|
if req_node in nodes_to_output:
|
2017-06-19 04:08:59 +00:00
|
|
|
self.output_node(req_node)
|
2017-06-19 04:08:58 +00:00
|
|
|
nodes_to_output.remove(req_node)
|
2017-06-19 04:08:59 +00:00
|
|
|
self.output_node(node)
|
2017-06-19 04:08:58 +00:00
|
|
|
nodes_to_output.remove(node)
|
2017-06-19 04:09:03 +00:00
|
|
|
|
|
|
|
|
|
|
|
def run_steps(args, dtb_file, include_disabled, output):
|
|
|
|
"""Run all the steps of the dtoc tool
|
|
|
|
|
|
|
|
Args:
|
|
|
|
args: List of non-option arguments provided to the problem
|
|
|
|
dtb_file: Filename of dtb file to process
|
|
|
|
include_disabled: True to include disabled nodes
|
|
|
|
output: Name of output file
|
|
|
|
"""
|
|
|
|
if not args:
|
|
|
|
raise ValueError('Please specify a command: struct, platdata')
|
|
|
|
|
|
|
|
plat = DtbPlatdata(dtb_file, include_disabled)
|
|
|
|
plat.scan_dtb()
|
|
|
|
plat.scan_tree()
|
|
|
|
plat.setup_output(output)
|
|
|
|
structs = plat.scan_structs()
|
|
|
|
plat.scan_phandles()
|
|
|
|
|
|
|
|
for cmd in args[0].split(','):
|
|
|
|
if cmd == 'struct':
|
|
|
|
plat.generate_structs(structs)
|
|
|
|
elif cmd == 'platdata':
|
|
|
|
plat.generate_tables()
|
|
|
|
else:
|
|
|
|
raise ValueError("Unknown command '%s': (use: struct, platdata)" %
|
|
|
|
cmd)
|