ardupilot/Tools/ardupilotwaf/ardupilotwaf.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

727 lines
21 KiB
Python
Raw Permalink Normal View History

2015-10-09 11:03:59 -03:00
# encoding: utf-8
from __future__ import print_function
from waflib import Build, ConfigSet, Configure, Context, Errors, Logs, Options, Utils, Task
from waflib.Configure import conf
from waflib.Scripting import run_command
from waflib.TaskGen import before_method, after_method, feature
import os.path, os
from pathlib import Path
from collections import OrderedDict
import subprocess
import ap_persistent
2015-10-09 11:03:59 -03:00
SOURCE_EXTS = [
'*.S',
'*.c',
'*.cpp',
]
COMMON_VEHICLE_DEPENDENT_CAN_LIBRARIES = [
'AP_CANManager',
'AP_KDECAN',
'AP_PiccoloCAN',
'AP_PiccoloCAN/piccolo_protocol',
]
2015-10-09 11:03:59 -03:00
COMMON_VEHICLE_DEPENDENT_LIBRARIES = [
'AP_Airspeed',
'AP_AccelCal',
2015-10-09 11:03:59 -03:00
'AP_ADC',
'AP_AHRS',
'AP_Airspeed',
'AP_Baro',
'AP_BattMonitor',
'AP_BoardConfig',
'AP_Camera',
2015-10-09 11:03:59 -03:00
'AP_Common',
'AP_Compass',
'AP_Declination',
'AP_GPS',
'AP_GSOF',
2015-10-09 11:03:59 -03:00
'AP_HAL',
'AP_HAL_Empty',
'AP_InertialSensor',
'AP_Math',
'AP_Mission',
2020-11-05 19:36:51 -04:00
'AP_DAL',
2019-08-28 01:08:59 -03:00
'AP_NavEKF',
2015-10-09 11:03:59 -03:00
'AP_NavEKF2',
2016-07-14 02:08:43 -03:00
'AP_NavEKF3',
2015-10-09 11:03:59 -03:00
'AP_Notify',
'AP_OpticalFlow',
'AP_Param',
'AP_Rally',
'AP_RangeFinder',
'AP_Scheduler',
'AP_SerialManager',
'AP_Terrain',
'AP_Vehicle',
'AP_InternalError',
'AP_Logger',
2015-10-09 11:03:59 -03:00
'Filter',
'GCS_MAVLink',
'RC_Channel',
2017-01-06 02:16:00 -04:00
'SRV_Channel',
2015-10-09 11:03:59 -03:00
'StorageManager',
2016-05-06 07:38:20 -03:00
'AP_Tuning',
2016-05-09 23:20:44 -03:00
'AP_RPM',
'AP_RSSI',
'AP_Mount',
2016-07-12 23:49:00 -03:00
'AP_Module',
2016-07-21 21:27:39 -03:00
'AP_Button',
2016-07-23 04:37:04 -03:00
'AP_ICEngine',
'AP_Networking',
2016-10-27 01:19:52 -03:00
'AP_Frsky_Telem',
'AP_IBus_Telem',
2016-11-17 05:19:01 -04:00
'AP_FlashStorage',
'AP_Relay',
'AP_ServoRelayEvents',
2017-11-02 23:35:14 -03:00
'AP_Volz_Protocol',
2017-11-22 13:36:55 -04:00
'AP_SBusOut',
'AP_IOMCU',
'AP_Parachute',
'AP_RAMTRON',
'AP_RCProtocol',
2018-01-05 04:56:49 -04:00
'AP_Radio',
2017-04-13 00:19:49 -03:00
'AP_TempCalibration',
'AP_VisualOdom',
2018-03-21 21:45:27 -03:00
'AP_BLHeli',
2018-04-14 05:30:54 -03:00
'AP_ROMFS',
'AP_Proximity',
2018-06-08 03:19:40 -03:00
'AP_Gripper',
'AP_RTC',
'AC_Sprayer',
'AC_Fence',
2018-10-26 02:39:00 -03:00
'AC_Avoidance',
'AP_LandingGear',
2018-12-18 23:28:04 -04:00
'AP_RobotisServo',
'AP_NMEA_Output',
'AP_OSD',
2019-08-01 08:17:38 -03:00
'AP_Filesystem',
'AP_ADSB',
'AP_ADSB/sagetech-sdk',
2019-11-02 00:11:23 -03:00
'AC_PID',
2019-10-30 07:14:17 -03:00
'AP_SerialLED',
2019-11-11 00:36:43 -04:00
'AP_EFI',
2020-01-02 23:55:52 -04:00
'AP_Hott_Telem',
'AP_ESC_Telem',
'AP_Servo_Telem',
2020-01-03 06:14:53 -04:00
'AP_Stats',
'AP_GyroFFT',
2020-03-28 19:41:57 -03:00
'AP_RCTelemetry',
2019-11-06 20:33:12 -04:00
'AP_Generator',
2020-08-04 17:46:16 -03:00
'AP_MSP',
2020-08-20 02:24:57 -03:00
'AP_OLC',
'AP_WheelEncoder',
2020-12-27 22:05:42 -04:00
'AP_ExternalAHRS',
2021-01-29 15:57:21 -04:00
'AP_VideoTX',
2021-02-05 11:37:53 -04:00
'AP_FETtecOneWire',
'AP_TemperatureSensor',
'AP_Torqeedo',
'AP_CustomRotations',
'AP_AIS',
'AP_OpenDroneID',
'AP_CheckFirmware',
'AP_ExternalControl',
2024-01-31 18:01:40 -04:00
'AP_JSON',
'AP_Beacon',
'AP_Arming',
'AP_RCMapper',
2024-11-17 18:52:40 -04:00
'AP_MultiHeap',
2015-10-09 11:03:59 -03:00
]
def get_legacy_defines(sketch_name, bld):
# If we are building heli, we adjust the build directory define so that
# we do not need to actually split heli and copter directories
if bld.cmd == 'heli' or 'heli' in bld.targets:
return [
'APM_BUILD_DIRECTORY=APM_BUILD_Heli',
'AP_BUILD_TARGET_NAME="' + sketch_name + '"',
]
2015-10-09 11:03:59 -03:00
return [
'APM_BUILD_DIRECTORY=APM_BUILD_' + sketch_name,
'AP_BUILD_TARGET_NAME="' + sketch_name + '"',
2015-10-09 11:03:59 -03:00
]
IGNORED_AP_LIBRARIES = [
'doc',
'AP_Scripting', # this gets explicitly included when it is needed and should otherwise never be globbed in
'AP_DDS',
2015-10-09 11:03:59 -03:00
]
def ap_autoconfigure(execute_method):
"""
Decorator that enables context commands to run *configure* as needed.
"""
def execute(self):
"""
Wraps :py:func:`waflib.Context.Context.execute` on the context class
"""
if 'tools/' in self.targets:
raise Errors.WafError('\"tools\" name has been replaced with \"tool\" for build please use that!')
if not Configure.autoconfig:
return execute_method(self)
# Disable autoconfig so waf's version doesn't run (and don't end up on loop of bad configure)
Configure.autoconfig = False
if self.variant == '':
raise Errors.WafError('The project is badly configured: run "waf configure" again!')
env = ConfigSet.ConfigSet()
do_config = False
try:
p = os.path.join(Context.out_dir, Build.CACHE_DIR, self.variant + Build.CACHE_SUFFIX)
env.load(p)
except EnvironmentError:
raise Errors.WafError('The project is not configured for board {0}: run "waf configure --board {0} [...]" first!'.format(self.variant))
lock_env = ConfigSet.ConfigSet()
try:
lock_env.load(os.path.join(Context.top_dir, Options.lockfile))
except EnvironmentError:
Logs.warn('Configuring the project')
do_config = True
else:
if lock_env.run_dir != Context.run_dir:
do_config = True
else:
h = 0
for f in env.CONFIGURE_FILES:
try:
h = Utils.h_list((h, Utils.readf(f, 'rb')))
except EnvironmentError:
do_config = True
break
else:
do_config = h != env.CONFIGURE_HASH
if do_config:
cmd = lock_env.config_cmd or 'configure'
tmp = Options.options.__dict__
if env.OPTIONS and sorted(env.OPTIONS.keys()) == sorted(tmp.keys()):
Options.options.__dict__ = env.OPTIONS
else:
raise Errors.WafError('The project configure options have changed: run "waf configure" again!')
try:
run_command(cmd)
finally:
Options.options.__dict__ = tmp
run_command(self.cmd)
else:
return execute_method(self)
return execute
def ap_configure_post_recurse():
post_recurse_orig = Configure.ConfigurationContext.post_recurse
def post_recurse(self, node):
post_recurse_orig(self, node)
self.all_envs[self.variant].CONFIGURE_FILES = self.files
self.all_envs[self.variant].CONFIGURE_HASH = self.hash
return post_recurse
@conf
def ap_get_all_libraries(bld):
if bld.env.BOOTLOADER:
# we don't need the full set of libraries for the bootloader build
return ['AP_HAL']
2015-10-09 11:03:59 -03:00
libraries = []
for lib_node in bld.srcnode.ant_glob('libraries/*', dir=True, src=False):
2015-10-09 11:03:59 -03:00
name = lib_node.name
if name in IGNORED_AP_LIBRARIES:
continue
if name.startswith('AP_HAL'):
continue
if name == 'SITL':
continue
2015-10-09 11:03:59 -03:00
libraries.append(name)
libraries.extend(['AP_HAL', 'AP_HAL_Empty'])
libraries.append('AP_PiccoloCAN/piccolo_protocol')
2015-10-09 11:03:59 -03:00
return libraries
@conf
def ap_common_vehicle_libraries(bld):
libraries = COMMON_VEHICLE_DEPENDENT_LIBRARIES
if bld.env.with_can or bld.env.HAL_NUM_CAN_IFACES:
libraries.extend(COMMON_VEHICLE_DEPENDENT_CAN_LIBRARIES)
return libraries
2016-01-22 08:26:08 -04:00
_grouped_programs = {}
class upload_fw_blueos(Task.Task):
def run(self):
# this is rarely used, so we import requests here to avoid the overhead
import requests
binary_path = self.inputs[0].abspath()
# check if .apj file exists for chibios builds
if Path(binary_path + ".apj").exists():
binary_path = binary_path + ".apj"
bld = self.generator.bld
board = bld.bldnode.name.capitalize()
print(f"Uploading {binary_path} to BlueOS at {bld.options.upload_blueos} for board {board}")
url = f'{bld.options.upload_blueos}/ardupilot-manager/v1.0/install_firmware_from_file?board_name={board}'
files = {
'binary': open(binary_path, 'rb')
}
response = requests.post(url, files=files, verify=False)
if response.status_code != 200:
raise Errors.WafError(f"Failed to upload firmware to BlueOS: {response.status_code}: {response.text}")
print("Upload complete")
def keyword(self):
return "Uploading to BlueOS"
class check_elf_symbols(Task.Task):
color='CYAN'
always_run = True
def keyword(self):
return "checking symbols"
def run(self):
'''
check for disallowed symbols in elf file, such as C++ exceptions
'''
elfpath = self.inputs[0].abspath()
if not self.env.CHECK_SYMBOLS:
# checking symbols disabled on this build
return
if not self.env.vehicle_binary or self.env.SIM_ENABLED:
# we only want to check symbols for vehicle binaries, allowing examples
# to use C++ exceptions. We also allow them in simulator builds
return
# we use string find on these symbols, so this catches all types of throw
# calls this should catch all uses of exceptions unless the compiler
# manages to inline them
blacklist = ['std::__throw',
'operator new[](unsigned int)',
'operator new[](unsigned long)',
'operator new(unsigned int)',
'operator new(unsigned long)']
nmout = subprocess.getoutput("%s -C %s" % (self.env.get_flat('NM'), elfpath))
for b in blacklist:
if nmout.find(b) != -1:
raise Errors.WafError("Disallowed symbol in %s: %s" % (elfpath, b))
@feature('post_link')
@after_method('process_source')
def post_link(self):
'''
setup tasks to run after link stage
'''
self.link_task.always_run = True
link_output = self.link_task.outputs[0]
check_elf_task = self.create_task('check_elf_symbols', src=link_output)
check_elf_task.set_run_after(self.link_task)
if self.bld.options.upload_blueos and self.env["BOARD_CLASS"] == "LINUX":
_upload_task = self.create_task('upload_fw_blueos', src=link_output)
_upload_task.set_run_after(self.link_task)
@conf
def ap_program(bld,
program_groups='bin',
program_dir=None,
use_legacy_defines=True,
program_name=None,
vehicle_binary=True,
**kw):
2015-10-09 11:03:59 -03:00
if 'target' in kw:
bld.fatal('Do not pass target for program')
if 'defines' not in kw:
kw['defines'] = []
if 'source' not in kw:
kw['source'] = bld.path.ant_glob(SOURCE_EXTS)
if not program_name:
program_name = bld.path.name
if use_legacy_defines:
kw['defines'].extend(get_legacy_defines(bld.path.name, bld))
2015-10-09 11:03:59 -03:00
kw['features'] = kw.get('features', []) + bld.env.AP_PROGRAM_FEATURES + ['post_link']
program_groups = Utils.to_list(program_groups)
if not program_dir:
program_dir = program_groups[0]
name = os.path.join(program_dir, program_name)
tg_constructor = bld.program
if bld.env.AP_PROGRAM_AS_STLIB:
tg_constructor = bld.stlib
else:
if bld.env.STATIC_LINKING:
kw['features'].append('static_linking')
tg = tg_constructor(
target='#%s' % name,
name=name,
program_name=program_name,
program_dir=program_dir,
2015-10-09 11:03:59 -03:00
**kw
)
tg.env.vehicle_binary = vehicle_binary
2018-12-30 20:55:27 -04:00
if 'use' in kw and bld.env.STATIC_LINKING:
# ensure we link against vehicle library
tg.env.STLIB += [kw['use']]
for group in program_groups:
_grouped_programs.setdefault(group, {}).update({tg.name : tg})
return tg
2015-10-09 11:03:59 -03:00
@conf
def ap_example(bld, **kw):
kw['program_groups'] = 'examples'
ap_program(bld, use_legacy_defines=False, vehicle_binary=False, **kw)
def unique_list(items):
'''remove duplicate elements from a list while maintaining ordering'''
return list(OrderedDict.fromkeys(items))
@conf
def ap_stlib(bld, **kw):
2015-10-09 11:03:59 -03:00
if 'name' not in kw:
bld.fatal('Missing name for ap_stlib')
if 'ap_vehicle' not in kw:
bld.fatal('Missing ap_vehicle for ap_stlib')
if 'ap_libraries' not in kw:
bld.fatal('Missing ap_libraries for ap_stlib')
2015-10-09 11:03:59 -03:00
kw['ap_libraries'] = unique_list(kw['ap_libraries'] + bld.env.AP_LIBRARIES)
for l in kw['ap_libraries']:
bld.ap_library(l, kw['ap_vehicle'])
2015-10-09 11:03:59 -03:00
if 'dynamic_source' not in kw:
kw['dynamic_source'] = 'modules/DroneCAN/libcanard/dsdlc_generated/src/**.c'
kw['features'] = kw.get('features', []) + ['cxx', 'cxxstlib']
kw['target'] = kw['name']
kw['source'] = []
bld.stlib(**kw)
_created_program_dirs = set()
@feature('cxxstlib', 'cxxprogram')
@before_method('process_rule')
def ap_create_program_dir(self):
if not hasattr(self, 'program_dir'):
return
if self.program_dir in _created_program_dirs:
return
self.bld.bldnode.make_node(self.program_dir).mkdir()
_created_program_dirs.add(self.program_dir)
@feature('cxxstlib')
@before_method('process_rule')
def ap_stlib_target(self):
if self.target.startswith('#'):
self.target = self.target[1:]
self.target = '#%s' % os.path.join('lib', self.target)
@conf
def ap_find_tests(bld, use=[], DOUBLE_PRECISION_SOURCES=[]):
if not bld.env.HAS_GTEST:
return
features = []
if bld.cmd == 'check':
features.append('test')
use = Utils.to_list(use)
use.append('GTEST')
includes = [bld.srcnode.abspath() + '/tests/']
for f in bld.path.ant_glob(incl='*.cpp'):
t = ap_program(
bld,
features=features,
includes=includes,
source=[f],
use=use,
program_name=f.change_ext('').name,
program_groups='tests',
use_legacy_defines=False,
vehicle_binary=False,
cxxflags=['-Wno-undef'],
)
filename = os.path.basename(f.abspath())
if filename in DOUBLE_PRECISION_SOURCES:
t.env.CXXFLAGS = t.env.CXXFLAGS[:]
single_precision_option='-fsingle-precision-constant'
if single_precision_option in t.env.CXXFLAGS:
t.env.CXXFLAGS.remove(single_precision_option)
single_precision_option='-cl-single-precision-constant'
if single_precision_option in t.env.CXXFLAGS:
t.env.CXXFLAGS.remove(single_precision_option)
t.env.CXXFLAGS.append("-DALLOW_DOUBLE_MATH_FUNCTIONS")
_versions = []
@conf
def ap_version_append_str(ctx, k, v):
ctx.env['AP_VERSION_ITEMS'] += [(k, '"{}"'.format(os.environ.get(k, v)))]
@conf
def ap_version_append_int(ctx, k, v):
ctx.env['AP_VERSION_ITEMS'] += [(k, '{}'.format(os.environ.get(k, v)))]
@conf
def write_version_header(ctx, tgt):
with open(tgt, 'w') as f:
print(
'''// auto-generated header, do not edit
#pragma once
#ifndef FORCE_VERSION_H_INCLUDE
#error ap_version.h should never be included directly. You probably want to include AP_Common/AP_FWVersion.h
#endif
''', file=f)
for k, v in ctx.env['AP_VERSION_ITEMS']:
print('#define {} {}'.format(k, v), file=f)
@conf
def ap_find_benchmarks(bld, use=[]):
if not bld.env.HAS_GBENCHMARK:
return
includes = [bld.srcnode.abspath() + '/benchmarks/']
to_remove = '-Werror=suggest-override'
if to_remove in bld.env.CXXFLAGS:
need_remove = True
else:
need_remove = False
if need_remove:
while to_remove in bld.env.CXXFLAGS:
bld.env.CXXFLAGS.remove(to_remove)
for f in bld.path.ant_glob(incl='*.cpp'):
ap_program(
bld,
features=['gbenchmark'],
includes=includes,
source=[f],
use=use,
vehicle_binary=False,
program_name=f.change_ext('').name,
program_groups='benchmarks',
use_legacy_defines=False,
)
def test_summary(bld):
from io import BytesIO
import sys
if not hasattr(bld, 'utest_results'):
Logs.info('check: no test run')
return
fails = []
for filename, exit_code, out, err in bld.utest_results:
Logs.pprint('GREEN' if exit_code == 0 else 'YELLOW',
' %s' % filename,
'returned %d' % exit_code)
if exit_code != 0:
fails.append(filename)
elif not bld.options.check_verbose:
continue
if len(out):
buf = BytesIO(out)
for line in buf:
print(" OUT: %s" % line.decode(), end='', file=sys.stderr)
print()
if len(err):
buf = BytesIO(err)
for line in buf:
print(" ERR: %s" % line.decode(), end='', file=sys.stderr)
print()
if not fails:
Logs.info('check: All %u tests passed!' % len(bld.utest_results))
return
Logs.error('check: %u of %u tests failed' %
(len(fails), len(bld.utest_results)))
for filename in fails:
Logs.error(' %s' % filename)
bld.fatal('check: some tests failed')
_build_commands = {}
def _process_build_command(bld):
if bld.cmd not in _build_commands:
return
params = _build_commands[bld.cmd]
targets = params['targets']
if targets:
if bld.targets:
bld.targets += ',' + targets
else:
bld.targets = targets
program_group_list = Utils.to_list(params['program_group_list'])
bld.options.program_group.extend(program_group_list)
def build_command(name,
targets=None,
program_group_list=[],
doc='build shortcut'):
_build_commands[name] = dict(
targets=targets,
program_group_list=program_group_list,
)
class context_class(Build.BuildContext):
cmd = name
context_class.__doc__ = doc
2016-01-22 08:26:08 -04:00
def _select_programs_from_group(bld):
groups = bld.options.program_group
if not groups:
if bld.targets:
groups = []
else:
groups = ['bin']
possible_groups = list(_grouped_programs.keys())
possible_groups.remove('bin') # Remove `bin` so as not to duplicate all items in bin
2016-01-22 08:26:08 -04:00
if 'all' in groups:
groups = possible_groups
2016-01-22 08:26:08 -04:00
for group in groups:
if group not in _grouped_programs:
bld.fatal(f'Group {group} not found, possible groups: {possible_groups}')
2016-01-22 08:26:08 -04:00
target_names = _grouped_programs[group].keys()
2016-01-22 08:26:08 -04:00
for name in target_names:
if bld.targets:
bld.targets += ',' + name
else:
bld.targets = name
2016-01-22 08:26:08 -04:00
def options(opt):
opt.ap_groups = {
'configure': opt.add_option_group('Ardupilot configure options'),
'linux': opt.add_option_group('Linux boards configure options'),
'build': opt.add_option_group('Ardupilot build options'),
'check': opt.add_option_group('Ardupilot check options'),
'clean': opt.add_option_group('Ardupilot clean options'),
}
g = opt.ap_groups['build']
2016-01-22 08:26:08 -04:00
g.add_option('--program-group',
action='append',
default=[],
help='''Select all programs that go in <PROGRAM_GROUP>/ for the build.
Example: `waf --program-group examples` builds all examples. The
special group "all" selects all programs.
''')
2016-01-22 08:26:08 -04:00
g.add_option('--upload',
action='store_true',
help='''Upload applicable targets to a connected device. Not all
platforms may support this. Example: `waf copter --upload` means "build
arducopter and upload it to my board".
''')
2019-09-25 08:32:32 -03:00
g.add_option('--upload-port',
action='store',
dest='upload_port',
default=None,
help='''Specify the port to be used with the --upload option. For example a port of /dev/ttyS10 indicates that serial port 10 shuld be used.
''')
g.add_option('--upload-blueos',
action='store',
dest='upload_blueos',
default=None,
help='''Automatically upload to a BlueOS device. The argument is the url for the device. http://blueos.local for example.
''')
g.add_option('--upload-force',
action='store_true',
help='''Override board type check and continue loading. Same as using uploader.py --force.
''')
g.add_option('--define',
action='append',
help='Add C++ define to build.')
g = opt.ap_groups['check']
g.add_option('--check-verbose',
action='store_true',
help='Output all test programs.')
g = opt.ap_groups['clean']
g.add_option('--clean-all-sigs',
action='store_true',
help='''Clean signatures for all tasks. By default, tasks that scan for
implicit dependencies (like the compilation tasks) keep the dependency
information across clean commands, so that that information is changed
only when really necessary. Also, some tasks that don't really produce
files persist their signature. This option avoids that behavior when
cleaning the build.
''')
g.add_option('--asan',
action='store_true',
help='''Build using the macOS clang Address Sanitizer. In order to run with
Address Sanitizer support llvm-symbolizer is required to be on the PATH.
This option is only supported on macOS versions of clang.
''')
g.add_option('--ubsan',
action='store_true',
help='''Build using the gcc undefined behaviour sanitizer''')
g.add_option('--ubsan-abort',
action='store_true',
help='''Build using the gcc undefined behaviour sanitizer and abort on error''')
2016-01-22 08:26:08 -04:00
def build(bld):
bld.add_pre_fun(_process_build_command)
2016-01-22 08:26:08 -04:00
bld.add_pre_fun(_select_programs_from_group)