# IceCream distributed compiling support # # Stages directories with symlinks from gcc/g++ to icecc, for both # native and cross compilers. Depending on each configure or compile, # the directories are added at the head of the PATH list and ICECC_CXX # and ICEC_CC are set. # # For the cross compiler, creates a tar.gz of our toolchain and sets # ICECC_VERSION accordingly. # # The class now handles all 3 different compile 'stages' (i.e native ,cross-kernel and target) creating the # necessary environment tar.gz file to be used by the remote machines. # It also supports meta-toolchain generation # # If ICECC_PATH is not set in local.conf then the class will try to locate it using 'bb.utils.which' # but nothing is sure ;) # # If ICECC_ENV_EXEC is set in local.conf, then it should point to the icecc-create-env script provided by the user # or the default one provided by icecc-create-env.bb will be used # (NOTE that this is a modified version of the script need it and *not the one that comes with icecc* # # User can specify if specific packages or packages belonging to class should not use icecc to distribute # compile jobs to remote machines, but handled locally, by defining ICECC_USER_CLASS_BL and ICECC_USER_PACKAGE_BL # with the appropriate values in local.conf. In addition the user can force to enable icecc for packages # which set an empty PARALLEL_MAKE variable by defining ICECC_USER_PACKAGE_WL. # ######################################################################################### #Error checking is kept to minimum so double check any parameters you pass to the class ########################################################################################### BB_HASHBASE_WHITELIST += "ICECC_PARALLEL_MAKE ICECC_DISABLED ICECC_USER_PACKAGE_BL \ ICECC_USER_CLASS_BL ICECC_USER_PACKAGE_WL ICECC_PATH ICECC_ENV_EXEC \ ICECC_CARET_WORKAROUND ICECC_CFLAGS ICECC_ENV_VERSION \ ICECC_DEBUG ICECC_LOGFILE ICECC_REPEAT_RATE ICECC_PREFERRED_HOST \ ICECC_CLANG_REMOTE_CPP ICECC_IGNORE_UNVERIFIED ICECC_TEST_SOCKET \ ICECC_ENV_DEBUG ICECC_SYSTEM_PACKAGE_BL ICECC_SYSTEM_CLASS_BL \ ICECC_REMOTE_CPP \ " ICECC_ENV_EXEC ?= "${STAGING_BINDIR_NATIVE}/icecc-create-env" HOSTTOOLS_NONFATAL += "icecc patchelf" # This version can be incremented when changes are made to the environment that # invalidate the version on the compile nodes. Changing it will cause a new # environment to be created. # # A useful thing to do for testing Icecream changes locally is to add a # subversion in local.conf: # ICECC_ENV_VERSION:append = "-my-ver-1" ICECC_ENV_VERSION = "2" # Default to disabling the caret workaround, If set to "1" in local.conf, icecc # will locally recompile any files that have warnings, which can adversely # affect performance. # # See: https://github.com/icecc/icecream/issues/190 export ICECC_CARET_WORKAROUND ??= "0" export ICECC_REMOTE_CPP ??= "0" ICECC_CFLAGS = "" CFLAGS += "${ICECC_CFLAGS}" CXXFLAGS += "${ICECC_CFLAGS}" # Debug flags when generating environments ICECC_ENV_DEBUG ??= "" # "system" recipe blacklist contains a list of packages that can not distribute # compile tasks for one reason or the other. When adding new entry, please # document why (how it failed) so that we can re-evaluate it later e.g. when # there is new version # # libgcc-initial - fails with CPP sanity check error if host sysroot contains # cross gcc built for another target tune/variant # pixman - prng_state: TLS reference mismatches non-TLS reference, possibly due to # pragma omp threadprivate(prng_state) # systemtap - _HelperSDT.c undefs macros and uses the identifiers in macros emitting # inline assembly # target-sdk-provides-dummy - ${HOST_PREFIX} is empty which triggers the "NULL # prefix" error. ICECC_SYSTEM_PACKAGE_BL += "\ libgcc-initial \ pixman \ systemtap \ target-sdk-provides-dummy \ " # "system" classes that should be blacklisted. When adding new entry, please # document why (how it failed) so that we can re-evaluate it later # # image - Image aren't compiling, but the testing framework for images captures # PARALLEL_MAKE as part of the test environment. Many tests won't use # icecream, but leaving the high level of parallelism can cause them to # consume an unnecessary amount of resources. ICECC_SYSTEM_CLASS_BL += "\ image \ " def get_icecc_dep(d): # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not # we need that built is the responsibility of the patch function / class, not # the application. if not d.getVar('INHIBIT_DEFAULT_DEPS'): return "icecc-create-env-native" return "" DEPENDS:prepend = "${@get_icecc_dep(d)} " get_cross_kernel_cc[vardepsexclude] += "KERNEL_CC" def get_cross_kernel_cc(bb,d): if not icecc_is_kernel(bb, d): return None # evaluate the expression by the shell if necessary kernel_cc = d.getVar('KERNEL_CC') if '`' in kernel_cc or '$(' in kernel_cc: import subprocess kernel_cc = subprocess.check_output("echo %s" % kernel_cc, shell=True).decode("utf-8")[:-1] kernel_cc = kernel_cc.replace('ccache', '').strip() kernel_cc = kernel_cc.split(' ')[0] kernel_cc = kernel_cc.strip() return kernel_cc def get_icecc(d): return d.getVar('ICECC_PATH') or bb.utils.which(os.getenv("PATH"), "icecc") def use_icecc(bb,d): if d.getVar('ICECC_DISABLED') == "1": # don't even try it, when explicitly disabled return "no" # allarch recipes don't use compiler if icecc_is_allarch(bb, d): return "no" if icecc_is_cross_canadian(bb, d): return "no" pn = d.getVar('PN') bpn = d.getVar('BPN') # Blacklist/whitelist checks are made against BPN, because there is a good # chance that if icecc should be skipped for a recipe, it should be skipped # for all the variants of that recipe. PN is still checked in case a user # specified a more specific recipe. check_pn = set([pn, bpn]) system_class_blacklist = (d.getVar('ICECC_SYSTEM_CLASS_BL') or "").split() user_class_blacklist = (d.getVar('ICECC_USER_CLASS_BL') or "none").split() package_class_blacklist = system_class_blacklist + user_class_blacklist for black in package_class_blacklist: if bb.data.inherits_class(black, d): bb.debug(1, "%s: class %s found in blacklist, disable icecc" % (pn, black)) return "no" system_package_blacklist = (d.getVar('ICECC_SYSTEM_PACKAGE_BL') or "").split() user_package_blacklist = (d.getVar('ICECC_USER_PACKAGE_BL') or "").split() user_package_whitelist = (d.getVar('ICECC_USER_PACKAGE_WL') or "").split() package_blacklist = system_package_blacklist + user_package_blacklist if check_pn & set(package_blacklist): bb.debug(1, "%s: found in blacklist, disable icecc" % pn) return "no" if check_pn & set(user_package_whitelist): bb.debug(1, "%s: found in whitelist, enable icecc" % pn) return "yes" if d.getVar('PARALLEL_MAKE') == "": bb.debug(1, "%s: has empty PARALLEL_MAKE, disable icecc" % pn) return "no" return "yes" def icecc_is_allarch(bb, d): return d.getVar("PACKAGE_ARCH") == "all" def icecc_is_kernel(bb, d): return \ bb.data.inherits_class("kernel", d); def icecc_is_native(bb, d): return \ bb.data.inherits_class("cross", d) or \ bb.data.inherits_class("native", d); def icecc_is_cross_canadian(bb, d): return bb.data.inherits_class("cross-canadian", d) def icecc_dir(bb, d): return d.expand('${TMPDIR}/work-shared/ice') # Don't pollute allarch signatures with TARGET_FPU icecc_version[vardepsexclude] += "TARGET_FPU" def icecc_version(bb, d): if use_icecc(bb, d) == "no": return "" parallel = d.getVar('ICECC_PARALLEL_MAKE') or "" if not d.getVar('PARALLEL_MAKE') == "" and parallel: d.setVar("PARALLEL_MAKE", parallel) # Disable showing the caret in the GCC compiler output if the workaround is # disabled if d.getVar('ICECC_CARET_WORKAROUND') == '0': d.setVar('ICECC_CFLAGS', '-fno-diagnostics-show-caret') if icecc_is_native(bb, d): archive_name = "local-host-env" elif d.expand('${HOST_PREFIX}') == "": bb.fatal(d.expand("${PN}"), " NULL prefix") else: prefix = d.expand('${HOST_PREFIX}' ) distro = d.expand('${DISTRO}') target_sys = d.expand('${TARGET_SYS}') float = d.getVar('TARGET_FPU') or "hard" archive_name = prefix + distro + "-" + target_sys + "-" + float if icecc_is_kernel(bb, d): archive_name += "-kernel" import socket ice_dir = icecc_dir(bb, d) tar_file = os.path.join(ice_dir, "{archive}-{version}-@VERSION@-{hostname}.tar.gz".format( archive=archive_name, version=d.getVar('ICECC_ENV_VERSION'), hostname=socket.gethostname() )) return tar_file def icecc_path(bb,d): if use_icecc(bb, d) == "no": # don't create unnecessary directories when icecc is disabled return staging = os.path.join(d.expand('${STAGING_BINDIR}'), "ice") if icecc_is_kernel(bb, d): staging += "-kernel" return staging def icecc_get_external_tool(bb, d, tool): external_toolchain_bindir = d.expand('${EXTERNAL_TOOLCHAIN}${bindir_cross}') target_prefix = d.expand('${TARGET_PREFIX}') return os.path.join(external_toolchain_bindir, '%s%s' % (target_prefix, tool)) def icecc_get_tool_link(tool, d): import subprocess try: return subprocess.check_output("readlink -f %s" % tool, shell=True).decode("utf-8")[:-1] except subprocess.CalledProcessError as e: bb.note("icecc: one of the tools probably disappeared during recipe parsing, cmd readlink -f %s returned %d:\n%s" % (tool, e.returncode, e.output.decode("utf-8"))) return tool def icecc_get_path_tool(tool, d): # This is a little ugly, but we want to make sure we add an actual # compiler to the toolchain, not ccache. Some distros (e.g. Fedora) # have ccache enabled by default using symlinks PATH, meaning ccache # would be found first when looking for the compiler. paths = os.getenv("PATH").split(':') while True: p, hist = bb.utils.which(':'.join(paths), tool, history=True) if not p or os.path.basename(icecc_get_tool_link(p, d)) != 'ccache': return p paths = paths[len(hist):] return "" # Don't pollute native signatures with target TUNE_PKGARCH through STAGING_BINDIR_TOOLCHAIN icecc_get_tool[vardepsexclude] += "STAGING_BINDIR_TOOLCHAIN" def icecc_get_tool(bb, d, tool): if icecc_is_native(bb, d): return icecc_get_path_tool(tool, d) elif icecc_is_kernel(bb, d): return icecc_get_path_tool(get_cross_kernel_cc(bb, d), d) else: ice_dir = d.expand('${STAGING_BINDIR_TOOLCHAIN}') target_sys = d.expand('${TARGET_SYS}') for p in ice_dir.split(':'): tool_bin = os.path.join(p, "%s-%s" % (target_sys, tool)) if os.path.isfile(tool_bin): return tool_bin external_tool_bin = icecc_get_external_tool(bb, d, tool) if os.path.isfile(external_tool_bin): return external_tool_bin return "" def icecc_get_and_check_tool(bb, d, tool): # Check that g++ or gcc is not a symbolic link to icecc binary in # PATH or icecc-create-env script will silently create an invalid # compiler environment package. t = icecc_get_tool(bb, d, tool) if t: link_path = icecc_get_tool_link(t, d) if link_path == get_icecc(d): bb.error("%s is a symlink to %s in PATH and this prevents icecc from working" % (t, link_path)) return "" else: return t else: return t wait_for_file() { local TIME_ELAPSED=0 local FILE_TO_TEST=$1 local TIMEOUT=$2 until [ -f "$FILE_TO_TEST" ] do TIME_ELAPSED=`expr $TIME_ELAPSED + 1` if [ $TIME_ELAPSED -gt $TIMEOUT ] then return 1 fi sleep 1 done } def set_icecc_env(): # dummy python version of set_icecc_env return set_icecc_env[vardepsexclude] += "KERNEL_CC" set_icecc_env() { if [ "${@use_icecc(bb, d)}" = "no" ] then return fi ICECC_VERSION="${@icecc_version(bb, d)}" if [ "x${ICECC_VERSION}" = "x" ] then bbwarn "Cannot use icecc: could not get ICECC_VERSION" return fi ICE_PATH="${@icecc_path(bb, d)}" if [ "x${ICE_PATH}" = "x" ] then bbwarn "Cannot use icecc: could not get ICE_PATH" return fi ICECC_BIN="${@get_icecc(d)}" if [ -z "${ICECC_BIN}" ]; then bbwarn "Cannot use icecc: icecc binary not found" return fi if [ -z "$(which patchelf patchelf-uninative)" ]; then bbwarn "Cannot use icecc: patchelf not found" return fi ICECC_CC="${@icecc_get_and_check_tool(bb, d, "gcc")}" ICECC_CXX="${@icecc_get_and_check_tool(bb, d, "g++")}" # cannot use icecc_get_and_check_tool here because it assumes as without target_sys prefix ICECC_WHICH_AS="${@bb.utils.which(os.getenv('PATH'), 'as')}" if [ ! -x "${ICECC_CC}" -o ! -x "${ICECC_CXX}" ] then bbnote "Cannot use icecc: could not get ICECC_CC or ICECC_CXX" return fi ICE_VERSION=`$ICECC_CC -dumpversion` ICECC_VERSION=`echo ${ICECC_VERSION} | sed -e "s/@VERSION@/$ICE_VERSION/g"` if [ ! -x "${ICECC_ENV_EXEC}" ] then bbwarn "Cannot use icecc: invalid ICECC_ENV_EXEC" return fi # Create symlinks to icecc and wrapper-scripts in the recipe-sysroot directory mkdir -p $ICE_PATH/symlinks if [ -n "${KERNEL_CC}" ]; then compilers="${@get_cross_kernel_cc(bb,d)}" else compilers="${HOST_PREFIX}gcc ${HOST_PREFIX}g++" fi for compiler in $compilers; do ln -sf $ICECC_BIN $ICE_PATH/symlinks/$compiler rm -f $ICE_PATH/$compiler cat <<-__EOF__ > $ICE_PATH/$compiler #!/bin/sh -e export ICECC_VERSION=$ICECC_VERSION export ICECC_CC=$ICECC_CC export ICECC_CXX=$ICECC_CXX $ICE_PATH/symlinks/$compiler "\$@" __EOF__ chmod 775 $ICE_PATH/$compiler done ICECC_AS="`${ICECC_CC} -print-prog-name=as`" # for target recipes should return something like: # /OE/tmp-eglibc/sysroots/x86_64-linux/usr/libexec/arm920tt-oe-linux-gnueabi/gcc/arm-oe-linux-gnueabi/4.8.2/as # and just "as" for native, if it returns "as" in current directory (for whatever reason) use "as" from PATH if [ "`dirname "${ICECC_AS}"`" = "." ] then ICECC_AS="${ICECC_WHICH_AS}" fi if [ ! -f "${ICECC_VERSION}.done" ] then mkdir -p "`dirname "${ICECC_VERSION}"`" # the ICECC_VERSION generation step must be locked by a mutex # in order to prevent race conditions if flock -n "${ICECC_VERSION}.lock" \ ${ICECC_ENV_EXEC} ${ICECC_ENV_DEBUG} "${ICECC_CC}" "${ICECC_CXX}" "${ICECC_AS}" "${ICECC_VERSION}" then touch "${ICECC_VERSION}.done" elif ! wait_for_file "${ICECC_VERSION}.done" 30 then # locking failed so wait for ${ICECC_VERSION}.done to appear bbwarn "Timeout waiting for ${ICECC_VERSION}.done" return fi fi # Don't let ccache find the icecream compiler links that have been created, otherwise # it can end up invoking icecream recursively. export CCACHE_PATH="$PATH" export CCACHE_DISABLE="1" export PATH="$ICE_PATH:$PATH" bbnote "Using icecc path: $ICE_PATH" bbnote "Using icecc tarball: $ICECC_VERSION" } do_configure:prepend() { set_icecc_env } do_compile:prepend() { set_icecc_env } do_compile_kernelmodules:prepend() { set_icecc_env } do_install:prepend() { set_icecc_env } # IceCream is not (currently) supported in the extensible SDK ICECC_SDK_HOST_TASK = "nativesdk-icecc-toolchain" ICECC_SDK_HOST_TASK:task-populate-sdk-ext = "" # Don't include IceCream in uninative tarball ICECC_SDK_HOST_TASK:pn-uninative-tarball = "" # Add the toolchain scripts to the SDK TOOLCHAIN_HOST_TASK:append = " ${ICECC_SDK_HOST_TASK}"