158 lines
		
	
	
		
			4.6 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			158 lines
		
	
	
		
			4.6 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
	
	
	
#!/usr/bin/env python2
 | 
						|
# SPDX-License-Identifier: GPL-2.0+
 | 
						|
 | 
						|
# Copyright (c) 2016 Google, Inc
 | 
						|
# Written by Simon Glass <sjg@chromium.org>
 | 
						|
#
 | 
						|
# Creates binary images from input files controlled by a description
 | 
						|
#
 | 
						|
 | 
						|
"""See README for more information"""
 | 
						|
 | 
						|
import glob
 | 
						|
import multiprocessing
 | 
						|
import os
 | 
						|
import sys
 | 
						|
import traceback
 | 
						|
import unittest
 | 
						|
 | 
						|
# Bring in the patman and dtoc libraries
 | 
						|
our_path = os.path.dirname(os.path.realpath(__file__))
 | 
						|
for dirname in ['../patman', '../dtoc', '..', '../concurrencytest']:
 | 
						|
    sys.path.insert(0, os.path.join(our_path, dirname))
 | 
						|
 | 
						|
# Bring in the libfdt module
 | 
						|
sys.path.insert(0, 'scripts/dtc/pylibfdt')
 | 
						|
sys.path.insert(0, os.path.join(our_path,
 | 
						|
                '../../build-sandbox_spl/scripts/dtc/pylibfdt'))
 | 
						|
 | 
						|
import cmdline
 | 
						|
import command
 | 
						|
use_concurrent = True
 | 
						|
try:
 | 
						|
    from concurrencytest import ConcurrentTestSuite, fork_for_tests
 | 
						|
except:
 | 
						|
    use_concurrent = False
 | 
						|
import control
 | 
						|
import test_util
 | 
						|
 | 
						|
def RunTests(debug, processes, args):
 | 
						|
    """Run the functional tests and any embedded doctests
 | 
						|
 | 
						|
    Args:
 | 
						|
        debug: True to enable debugging, which shows a full stack trace on error
 | 
						|
        args: List of positional args provided to binman. This can hold a test
 | 
						|
            name to execute (as in 'binman -t testSections', for example)
 | 
						|
        processes: Number of processes to use to run tests (None=same as #CPUs)
 | 
						|
    """
 | 
						|
    import elf_test
 | 
						|
    import entry_test
 | 
						|
    import fdt_test
 | 
						|
    import ftest
 | 
						|
    import image_test
 | 
						|
    import test
 | 
						|
    import doctest
 | 
						|
 | 
						|
    result = unittest.TestResult()
 | 
						|
    for module in []:
 | 
						|
        suite = doctest.DocTestSuite(module)
 | 
						|
        suite.run(result)
 | 
						|
 | 
						|
    sys.argv = [sys.argv[0]]
 | 
						|
    if debug:
 | 
						|
        sys.argv.append('-D')
 | 
						|
    if debug:
 | 
						|
        sys.argv.append('-D')
 | 
						|
 | 
						|
    # Run the entry tests first ,since these need to be the first to import the
 | 
						|
    # 'entry' module.
 | 
						|
    test_name = args and args[0] or None
 | 
						|
    suite = unittest.TestSuite()
 | 
						|
    loader = unittest.TestLoader()
 | 
						|
    for module in (entry_test.TestEntry, ftest.TestFunctional, fdt_test.TestFdt,
 | 
						|
                   elf_test.TestElf, image_test.TestImage):
 | 
						|
        if test_name:
 | 
						|
            try:
 | 
						|
                suite.addTests(loader.loadTestsFromName(test_name, module))
 | 
						|
            except AttributeError:
 | 
						|
                continue
 | 
						|
        else:
 | 
						|
            suite.addTests(loader.loadTestsFromTestCase(module))
 | 
						|
    if use_concurrent and processes != 1:
 | 
						|
        concurrent_suite = ConcurrentTestSuite(suite,
 | 
						|
                fork_for_tests(processes or multiprocessing.cpu_count()))
 | 
						|
        concurrent_suite.run(result)
 | 
						|
    else:
 | 
						|
        suite.run(result)
 | 
						|
 | 
						|
    print result
 | 
						|
    for test, err in result.errors:
 | 
						|
        print test.id(), err
 | 
						|
    for test, err in result.failures:
 | 
						|
        print err, result.failures
 | 
						|
    if result.errors or result.failures:
 | 
						|
      print 'binman tests FAILED'
 | 
						|
      return 1
 | 
						|
    return 0
 | 
						|
 | 
						|
def GetEntryModules(include_testing=True):
 | 
						|
    """Get a set of entry class implementations
 | 
						|
 | 
						|
    Returns:
 | 
						|
        Set of paths to entry class filenames
 | 
						|
    """
 | 
						|
    glob_list = glob.glob(os.path.join(our_path, 'etype/*.py'))
 | 
						|
    return set([os.path.splitext(os.path.basename(item))[0]
 | 
						|
                for item in glob_list
 | 
						|
                if include_testing or '_testing' not in item])
 | 
						|
 | 
						|
def RunTestCoverage():
 | 
						|
    """Run the tests and check that we get 100% coverage"""
 | 
						|
    glob_list = GetEntryModules(False)
 | 
						|
    all_set = set([os.path.splitext(os.path.basename(item))[0]
 | 
						|
                   for item in glob_list if '_testing' not in item])
 | 
						|
    test_util.RunTestCoverage('tools/binman/binman.py', None,
 | 
						|
            ['*test*', '*binman.py', 'tools/patman/*', 'tools/dtoc/*'],
 | 
						|
            options.build_dir, all_set)
 | 
						|
 | 
						|
def RunBinman(options, args):
 | 
						|
    """Main entry point to binman once arguments are parsed
 | 
						|
 | 
						|
    Args:
 | 
						|
        options: Command-line options
 | 
						|
        args: Non-option arguments
 | 
						|
    """
 | 
						|
    ret_code = 0
 | 
						|
 | 
						|
    # For testing: This enables full exception traces.
 | 
						|
    #options.debug = True
 | 
						|
 | 
						|
    if not options.debug:
 | 
						|
        sys.tracebacklimit = 0
 | 
						|
 | 
						|
    if options.test:
 | 
						|
        ret_code = RunTests(options.debug, options.processes, args[1:])
 | 
						|
 | 
						|
    elif options.test_coverage:
 | 
						|
        RunTestCoverage()
 | 
						|
 | 
						|
    elif options.entry_docs:
 | 
						|
        control.WriteEntryDocs(GetEntryModules())
 | 
						|
 | 
						|
    else:
 | 
						|
        try:
 | 
						|
            ret_code = control.Binman(options, args)
 | 
						|
        except Exception as e:
 | 
						|
            print 'binman: %s' % e
 | 
						|
            if options.debug:
 | 
						|
                print
 | 
						|
                traceback.print_exc()
 | 
						|
            ret_code = 1
 | 
						|
    return ret_code
 | 
						|
 | 
						|
 | 
						|
if __name__ == "__main__":
 | 
						|
    (options, args) = cmdline.ParseArgs(sys.argv)
 | 
						|
    ret_code = RunBinman(options, args)
 | 
						|
    sys.exit(ret_code)
 |