From patchwork Tue Jul 1 21:18:42 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Khem Raj X-Patchwork-Id: 66037 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from aws-us-west-2-korg-lkml-1.web.codeaurora.org (localhost.localdomain [127.0.0.1]) by smtp.lore.kernel.org (Postfix) with ESMTP id 9FE2AC7EE30 for ; Tue, 1 Jul 2025 21:18:58 +0000 (UTC) Received: from mail-pf1-f180.google.com (mail-pf1-f180.google.com [209.85.210.180]) by mx.groups.io with SMTP id smtpd.web10.8582.1751404729849302901 for ; Tue, 01 Jul 2025 14:18:49 -0700 Authentication-Results: mx.groups.io; dkim=pass header.i=@gmail.com header.s=20230601 header.b=LENMlCgo; spf=pass (domain: gmail.com, ip: 209.85.210.180, mailfrom: raj.khem@gmail.com) Received: by mail-pf1-f180.google.com with SMTP id d2e1a72fcca58-7481600130eso4865539b3a.3 for ; Tue, 01 Jul 2025 14:18:49 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20230601; t=1751404729; x=1752009529; darn=lists.openembedded.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=nLHX7QsjEYrCC/TW+FrCqJ6dzQaEn4LCTb+VjAFcM0k=; b=LENMlCgolRqoCk34FrE34ghoj4UqbWDWZEbfHqOPPgcYGMavBkiSnxdIh8wB0GR/q+ LOsTYtcShBLkfcN+Lr7a/+vR2XQ7xSUWHnQoDKEZ3obVomjiVRyP/wkpvs3teYnmohtQ lmN4ZpoPFIV8gCLIrPz1MEgx1TUgDKzMW294ul8y51yN2VqbN3r1YUMXj2STVseGBlHL yQj3y8vGTxOb81IqfhFG5zVLWW3jdxqHH1xdaB6TtDgYhcyaD4kI9KOjes/k4rzo1hl5 n7VctD2MJU5dTTg/7u6TJtPzp3xmEoLQhSazZI5vX0PjacoxFj0kur/YU7jmnGLbeZEo cAvg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1751404729; x=1752009529; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=nLHX7QsjEYrCC/TW+FrCqJ6dzQaEn4LCTb+VjAFcM0k=; b=qT2PrYQGVDoCCC0QJouZIDpi66NCSuTYwiogQ+ttieOZjV8gA8YVCbc4EJhJFK10Yp Acuwi32n36jMGziw8W5c+qO7nacarjkexG/gmPqRK5kLhCq3XmYUi3lnbafCTMj3lZIq wA4OsUZyCI5rwFFsK+qZZiFqsUsGx2wDHSvDRHIYByB/0xYHv+SW0KyFb5lA9o5gqEqw vE/XXE8pvctLJ86tq9hkwIKhp3qBBopDkFGlakGiN0o7smhHPR5d+MfVbM5qHTU1MTLO CG6QrldIhPM8ABCajsfE8DH6ea+TOjVQ0BtL6w0ud9XxxhT/Rx/wy4ZnNho26/2QjVpj tBUg== X-Gm-Message-State: AOJu0YykQHUG87RKtST43h4X19nePDTK+zyWZTiMlbux1vztwK62/Gta yagiTEPt6V3JTbDzjLmnSuntEiH0l/ukH44vQZ/VM8eCIzZwYqXsniw5rDgbZ4Fu X-Gm-Gg: ASbGncuP3mocZeBAaG4j1mmMRByk8fldCr6nt+IMM15aDeT5sBRtX/B0gflgOfYAO9w 1L93EX4XknNUvycfu6mqIXxwi3ktehALn8oWH9bICykfaLuKL9P8j4V1hSZOQHHtRugY9vlUQIL +X0gHpIqZwv67uDzjfCQBq1bPqchFL/XRoqEsqRMXCokPQLnRAdMAy4DxZwKfQt67mOslAqlT6N P2U67tp30I1TBwTLIYI32t1BlEbt5Fwgg2gk8IHfgG4tyj3VFmRwQf4rze9gKrl4+tkx4fVqDsA gFZ2IspGelJ+t29Oj32pq2I0JWDiGPZkMBTngAoYkNmqdFxQVxI8jw== X-Google-Smtp-Source: AGHT+IHfKqd+KeyR92hAYd9ARZYK6+YU2YqMYuz6PJeDUzd3ziN56Ne0cD538uLZyVFxX2FVnc/xoA== X-Received: by 2002:a05:6a00:14cb:b0:736:31cf:2590 with SMTP id d2e1a72fcca58-74b5126405fmr437672b3a.16.1751404728498; Tue, 01 Jul 2025 14:18:48 -0700 (PDT) Received: from apollo.localdomain ([2601:646:8201:fd20::aa8c]) by smtp.gmail.com with ESMTPSA id d2e1a72fcca58-74af57ef669sm12116095b3a.158.2025.07.01.14.18.47 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 01 Jul 2025 14:18:47 -0700 (PDT) From: Khem Raj To: openembedded-core@lists.openembedded.org Cc: Khem Raj Subject: [PATCH v3 3/4] Remove oe.utils.host_gcc_version utility Date: Tue, 1 Jul 2025 14:18:42 -0700 Message-ID: <20250701211843.3960784-3-raj.khem@gmail.com> X-Mailer: git-send-email 2.50.0 In-Reply-To: <20250701211843.3960784-1-raj.khem@gmail.com> References: <20250701211843.3960784-1-raj.khem@gmail.com> MIME-Version: 1.0 List-Id: X-Webhook-Received: from li982-79.members.linode.com [45.33.32.79] by aws-us-west-2-korg-lkml-1.web.codeaurora.org with HTTPS for ; Tue, 01 Jul 2025 21:18:58 -0000 X-Groupsio-URL: https://lists.openembedded.org/g/openembedded-core/message/219727 This was used for the uninative gcc 4.8 and 4.9 incompatibility w.r.t runtime ABI, none of supported distros we use today use that old gcc Oldest we have is gcc 7.5.0 on opensuse 15.5 hosts Signed-off-by: Khem Raj --- meta/classes-global/uninative.bbclass | 2 +- meta/classes-recipe/populate_sdk_base.bbclass | 1 - meta/classes-recipe/populate_sdk_ext.bbclass | 3 +- meta/files/toolchain-shar-extract.sh | 8 - meta/lib/oe/utils.py | 147 +++++++++++------- 5 files changed, 90 insertions(+), 71 deletions(-) diff --git a/meta/classes-global/uninative.bbclass b/meta/classes-global/uninative.bbclass index 75e0c19704d..c246a1ecd6c 100644 --- a/meta/classes-global/uninative.bbclass +++ b/meta/classes-global/uninative.bbclass @@ -142,7 +142,7 @@ def enable_uninative(d): loader = d.getVar("UNINATIVE_LOADER") if os.path.exists(loader): bb.debug(2, "Enabling uninative") - d.setVar("NATIVELSBSTRING", "universal%s" % oe.utils.host_gcc_version(d)) + d.setVar("NATIVELSBSTRING", "universal") d.appendVar("SSTATEPOSTUNPACKFUNCS", " uninative_changeinterp") d.appendVarFlag("SSTATEPOSTUNPACKFUNCS", "vardepvalueexclude", "| uninative_changeinterp") d.appendVar("BUILD_LDFLAGS", " -Wl,--allow-shlib-undefined -Wl,--dynamic-linker=${UNINATIVE_LOADER} -pthread") diff --git a/meta/classes-recipe/populate_sdk_base.bbclass b/meta/classes-recipe/populate_sdk_base.bbclass index 8ef4b2be777..e6685cde97b 100644 --- a/meta/classes-recipe/populate_sdk_base.bbclass +++ b/meta/classes-recipe/populate_sdk_base.bbclass @@ -373,7 +373,6 @@ EOF -e 's#@SDK_VERSION@#${SDK_VERSION}#g' \ -e '/@SDK_PRE_INSTALL_COMMAND@/d' \ -e '/@SDK_POST_INSTALL_COMMAND@/d' \ - -e 's#@SDK_GCC_VER@#${@oe.utils.host_gcc_version(d, taskcontextonly=True)}#g' \ -e 's#@SDK_ARCHIVE_TYPE@#${SDK_ARCHIVE_TYPE}#g' \ ${SDKDEPLOYDIR}/${TOOLCHAIN_OUTPUTNAME}.sh diff --git a/meta/classes-recipe/populate_sdk_ext.bbclass b/meta/classes-recipe/populate_sdk_ext.bbclass index 2d7d661d259..eb4e16c57cb 100644 --- a/meta/classes-recipe/populate_sdk_ext.bbclass +++ b/meta/classes-recipe/populate_sdk_ext.bbclass @@ -491,8 +491,7 @@ def prepare_locked_cache(d, baseoutpath, derivative, conf_initpath): sstate_out = baseoutpath + '/sstate-cache' bb.utils.remove(sstate_out, True) - # uninative.bbclass sets NATIVELSBSTRING to 'universal%s' % oe.utils.host_gcc_version(d) - fixedlsbstring = "universal%s" % oe.utils.host_gcc_version(d) if bb.data.inherits_class('uninative', d) else "" + fixedlsbstring = "universal" if bb.data.inherits_class('uninative', d) else "" sdk_include_toolchain = (d.getVar('SDK_INCLUDE_TOOLCHAIN') == '1') sdk_ext_type = d.getVar('SDK_EXT_TYPE') diff --git a/meta/files/toolchain-shar-extract.sh b/meta/files/toolchain-shar-extract.sh index 06934e5a9ab..e2c617958a9 100644 --- a/meta/files/toolchain-shar-extract.sh +++ b/meta/files/toolchain-shar-extract.sh @@ -31,9 +31,6 @@ tweakpath /sbin INST_ARCH=$(uname -m | sed -e "s/i[3-6]86/ix86/" -e "s/x86[-_]64/x86_64/") SDK_ARCH=$(echo @SDK_ARCH@ | sed -e "s/i[3-6]86/ix86/" -e "s/x86[-_]64/x86_64/") -INST_GCC_VER=$(gcc --version 2>/dev/null | sed -ne 's/.* \([0-9]\+\.[0-9]\+\)\.[0-9]\+.*/\1/p') -SDK_GCC_VER='@SDK_GCC_VER@' - verlte () { [ "$1" = "`printf "$1\n$2" | sort -V | head -n1`" ] } @@ -145,11 +142,6 @@ fi # SDK_EXTENSIBLE is exposed from the SDK_PRE_INSTALL_COMMAND above if [ "$SDK_EXTENSIBLE" = "1" ]; then DEFAULT_INSTALL_DIR="@SDKEXTPATH@" - if [ "$INST_GCC_VER" = '4.8' -a "$SDK_GCC_VER" = '4.9' ] || [ "$INST_GCC_VER" = '4.8' -a "$SDK_GCC_VER" = '' ] || \ - [ "$INST_GCC_VER" = '4.9' -a "$SDK_GCC_VER" = '' ]; then - echo "Error: Incompatible SDK installer! Your host gcc version is $INST_GCC_VER and this SDK was built by gcc higher version." - exit 1 - fi fi if [ "$target_sdk_dir" = "" ]; then diff --git a/meta/lib/oe/utils.py b/meta/lib/oe/utils.py index a11db5f3cd9..ac127659632 100644 --- a/meta/lib/oe/utils.py +++ b/meta/lib/oe/utils.py @@ -11,42 +11,49 @@ import errno import bb.parse + def read_file(filename): try: - f = open( filename, "r" ) + f = open(filename, "r") except IOError as reason: - return "" # WARNING: can't raise an error now because of the new RDEPENDS handling. This is a bit ugly. :M: + return "" # WARNING: can't raise an error now because of the new RDEPENDS handling. This is a bit ugly. :M: else: data = f.read().strip() f.close() return data return None -def ifelse(condition, iftrue = True, iffalse = False): + +def ifelse(condition, iftrue=True, iffalse=False): if condition: return iftrue else: return iffalse + def conditional(variable, checkvalue, truevalue, falsevalue, d): if d.getVar(variable) == checkvalue: return truevalue else: return falsevalue + def vartrue(var, iftrue, iffalse, d): import oe.types + if oe.types.boolean(d.getVar(var)): return iftrue else: return iffalse + def less_or_equal(variable, checkvalue, truevalue, falsevalue, d): if float(d.getVar(variable)) <= float(checkvalue): return truevalue else: return falsevalue + def version_less_or_equal(variable, checkvalue, truevalue, falsevalue, d): result = bb.utils.vercmp_string(d.getVar(variable), checkvalue) if result <= 0: @@ -54,6 +61,7 @@ def version_less_or_equal(variable, checkvalue, truevalue, falsevalue, d): else: return falsevalue + def both_contain(variable1, variable2, checkvalue, d): val1 = d.getVar(variable1) val2 = d.getVar(variable2) @@ -68,6 +76,7 @@ def both_contain(variable1, variable2, checkvalue, d): else: return "" + def set_intersect(variable1, variable2, d): """ Expand both variables, interpret them as lists of strings, and return the @@ -83,36 +92,44 @@ def set_intersect(variable1, variable2, d): val2 = set(d.getVar(variable2).split()) return " ".join(val1 & val2) + def prune_suffix(var, suffixes, d): # See if var ends with any of the suffixes listed and # remove it if found for suffix in suffixes: if suffix and var.endswith(suffix): - var = var[:-len(suffix)] + var = var[: -len(suffix)] prefix = d.getVar("MLPREFIX") if prefix and var.startswith(prefix): - var = var[len(prefix):] + var = var[len(prefix) :] return var + def str_filter(f, str, d): from re import match + return " ".join([x for x in str.split() if match(f, x, 0)]) + def str_filter_out(f, str, d): from re import match + return " ".join([x for x in str.split() if not match(f, x, 0)]) + def build_depends_string(depends, task): """Append a taskname to a string of dependencies as used by the [depends] flag""" return " ".join(dep + ":" + task for dep in depends.split()) + def inherits(d, *classes): """Return True if the metadata inherits any of the specified classes""" return any(bb.data.inherits_class(cls, d) for cls in classes) -def features_backfill(var,d): + +def features_backfill(var, d): # This construct allows the addition of new features to variable specified # as var # Example for var = "DISTRO_FEATURES" @@ -122,8 +139,8 @@ def features_backfill(var,d): # Distributions wanting to elide a value in DISTRO_FEATURES_BACKFILL should # add the feature to DISTRO_FEATURES_BACKFILL_CONSIDERED features = (d.getVar(var) or "").split() - backfill = (d.getVar(var+"_BACKFILL") or "").split() - considered = (d.getVar(var+"_BACKFILL_CONSIDERED") or "").split() + backfill = (d.getVar(var + "_BACKFILL") or "").split() + considered = (d.getVar(var + "_BACKFILL_CONSIDERED") or "").split() addfeatures = [] for feature in backfill: @@ -133,6 +150,7 @@ def features_backfill(var,d): if addfeatures: d.appendVar(var, " " + " ".join(addfeatures)) + def all_distro_features(d, features, truevalue="1", falsevalue=""): """ Returns truevalue if *all* given features are set in DISTRO_FEATURES, @@ -153,6 +171,7 @@ def all_distro_features(d, features, truevalue="1", falsevalue=""): """ return bb.utils.contains("DISTRO_FEATURES", features, truevalue, falsevalue, d) + def any_distro_features(d, features, truevalue="1", falsevalue=""): """ Returns truevalue if at least *one* of the given features is set in DISTRO_FEATURES, @@ -174,6 +193,7 @@ def any_distro_features(d, features, truevalue="1", falsevalue=""): """ return bb.utils.contains_any("DISTRO_FEATURES", features, truevalue, falsevalue, d) + def parallel_make(d, makeinst=False): """ Return the integer value for the number of parallel threads to use when @@ -183,22 +203,23 @@ def parallel_make(d, makeinst=False): e.g. if PARALLEL_MAKE = "-j 10", this will return 10 as an integer. """ if makeinst: - pm = (d.getVar('PARALLEL_MAKEINST') or '').split() + pm = (d.getVar("PARALLEL_MAKEINST") or "").split() else: - pm = (d.getVar('PARALLEL_MAKE') or '').split() + pm = (d.getVar("PARALLEL_MAKE") or "").split() # look for '-j' and throw other options (e.g. '-l') away while pm: opt = pm.pop(0) - if opt == '-j': + if opt == "-j": v = pm.pop(0) - elif opt.startswith('-j'): + elif opt.startswith("-j"): v = opt[2:].strip() else: continue return int(v) - return '' + return "" + def parallel_make_argument(d, fmt, limit=None, makeinst=False): """ @@ -218,23 +239,28 @@ def parallel_make_argument(d, fmt, limit=None, makeinst=False): if limit: v = min(limit, v) return fmt % v - return '' + return "" + def packages_filter_out_system(d): """ Return a list of packages from PACKAGES with the "system" packages such as PN-dbg PN-doc PN-locale-eb-gb removed. """ - pn = d.getVar('PN') - pkgfilter = [pn + suffix for suffix in ('', '-dbg', '-dev', '-doc', '-locale', '-staticdev', '-src')] + pn = d.getVar("PN") + pkgfilter = [ + pn + suffix + for suffix in ("", "-dbg", "-dev", "-doc", "-locale", "-staticdev", "-src") + ] localepkg = pn + "-locale-" pkgs = [] - for pkg in d.getVar('PACKAGES').split(): + for pkg in d.getVar("PACKAGES").split(): if pkg not in pkgfilter and localepkg not in pkg: pkgs.append(pkg) return pkgs + def getstatusoutput(cmd): return subprocess.getstatusoutput(cmd) @@ -253,10 +279,12 @@ def trim_version(version, num_parts=2): trimmed = ".".join(parts[:num_parts]) return trimmed + def cpu_count(at_least=1, at_most=64): cpus = len(os.sched_getaffinity(0)) return max(min(cpus, at_most), at_least) + def execute_pre_post_process(d, cmds): if cmds is None: return @@ -267,14 +295,17 @@ def execute_pre_post_process(d, cmds): bb.note("Executing %s ..." % cmd) bb.build.exec_func(cmd, d) + @bb.parse.vardepsexclude("BB_NUMBER_THREADS") def get_bb_number_threads(d): return int(d.getVar("BB_NUMBER_THREADS") or os.cpu_count() or 1) + def multiprocess_launch(target, items, d, extraargs=None): max_process = get_bb_number_threads(d) return multiprocess_launch_mp(target, items, max_process, extraargs) + # For each item in items, call the function 'target' with item as the first # argument, extraargs as the other arguments and handle any exceptions in the # parent thread @@ -344,7 +375,7 @@ def multiprocess_launch_mp(target, items, max_process, extraargs=None): p.join() if errors: msg = "" - for (e, tb) in errors: + for e, tb in errors: if isinstance(e, subprocess.CalledProcessError) and e.output: msg = msg + str(e) + "\n" msg = msg + "Subprocess output:" @@ -354,24 +385,27 @@ def multiprocess_launch_mp(target, items, max_process, extraargs=None): bb.fatal("Fatal errors occurred in subprocesses:\n%s" % msg) return results + def squashspaces(string): import re + return re.sub(r"\s+", " ", string).strip() + def rprovides_map(pkgdata_dir, pkg_dict): # Map file -> pkg provider rprov_map = {} for pkg in pkg_dict: - path_to_pkgfile = os.path.join(pkgdata_dir, 'runtime-reverse', pkg) + path_to_pkgfile = os.path.join(pkgdata_dir, "runtime-reverse", pkg) if not os.path.isfile(path_to_pkgfile): continue with open(path_to_pkgfile) as f: for line in f: - if line.startswith('RPROVIDES') or line.startswith('FILERPROVIDES'): + if line.startswith("RPROVIDES") or line.startswith("FILERPROVIDES"): # List all components provided by pkg. # Exclude version strings, i.e. those starting with ( - provides = [x for x in line.split()[1:] if not x.startswith('(')] + provides = [x for x in line.split()[1:] if not x.startswith("(")] for prov in provides: if prov in rprov_map: rprov_map[prov].append(pkg) @@ -380,6 +414,7 @@ def rprovides_map(pkgdata_dir, pkg_dict): return rprov_map + def format_pkg_list(pkg_dict, ret_format=None, pkgdata_dir=None): output = [] @@ -388,10 +423,14 @@ def format_pkg_list(pkg_dict, ret_format=None, pkgdata_dir=None): output.append("%s %s" % (pkg, pkg_dict[pkg]["arch"])) elif ret_format == "file": for pkg in sorted(pkg_dict): - output.append("%s %s %s" % (pkg, pkg_dict[pkg]["filename"], pkg_dict[pkg]["arch"])) + output.append( + "%s %s %s" % (pkg, pkg_dict[pkg]["filename"], pkg_dict[pkg]["arch"]) + ) elif ret_format == "ver": for pkg in sorted(pkg_dict): - output.append("%s %s %s" % (pkg, pkg_dict[pkg]["arch"], pkg_dict[pkg]["ver"])) + output.append( + "%s %s %s" % (pkg, pkg_dict[pkg]["arch"], pkg_dict[pkg]["ver"]) + ) elif ret_format == "deps": rprov_map = rprovides_map(pkgdata_dir, pkg_dict) for pkg in sorted(pkg_dict): @@ -399,18 +438,20 @@ def format_pkg_list(pkg_dict, ret_format=None, pkgdata_dir=None): if dep in rprov_map: # There could be multiple providers within the image for pkg_provider in rprov_map[dep]: - output.append("%s|%s * %s [RPROVIDES]" % (pkg, pkg_provider, dep)) + output.append( + "%s|%s * %s [RPROVIDES]" % (pkg, pkg_provider, dep) + ) else: output.append("%s|%s" % (pkg, dep)) else: for pkg in sorted(pkg_dict): output.append(pkg) - output_str = '\n'.join(output) + output_str = "\n".join(output) if output_str: # make sure last line is newline terminated - output_str += '\n' + output_str += "\n" return output_str @@ -420,24 +461,27 @@ def format_pkg_list(pkg_dict, ret_format=None, pkgdata_dir=None): def get_host_compiler_version(d, taskcontextonly=False): import re, subprocess - if taskcontextonly and d.getVar('BB_WORKERCONTEXT') != '1': + if taskcontextonly and d.getVar("BB_WORKERCONTEXT") != "1": return compiler = d.getVar("BUILD_CC") # Get rid of ccache since it is not present when parsing. - if compiler.startswith('ccache '): + if compiler.startswith("ccache "): compiler = compiler[7:] try: env = os.environ.copy() # datastore PATH does not contain session PATH as set by environment-setup-... # this breaks the install-buildtools use-case # env["PATH"] = d.getVar("PATH") - output = subprocess.check_output("%s --version" % compiler, \ - shell=True, env=env, stderr=subprocess.STDOUT).decode("utf-8") + output = subprocess.check_output( + "%s --version" % compiler, shell=True, env=env, stderr=subprocess.STDOUT + ).decode("utf-8") except subprocess.CalledProcessError as e: - bb.fatal("Error running %s --version: %s" % (compiler, e.output.decode("utf-8"))) + bb.fatal( + "Error running %s --version: %s" % (compiler, e.output.decode("utf-8")) + ) - match = re.match(r".* (\d+\.\d+)\.\d+.*", output.split('\n')[0]) + match = re.match(r".* (\d+\.\d+)\.\d+.*", output.split("\n")[0]) if not match: bb.fatal("Can't get compiler version from %s --version output" % compiler) @@ -445,36 +489,13 @@ def get_host_compiler_version(d, taskcontextonly=False): return compiler, version -def host_gcc_version(d, taskcontextonly=False): - import re, subprocess - - if taskcontextonly and d.getVar('BB_WORKERCONTEXT') != '1': - return - - compiler = d.getVar("BUILD_CC") - # Get rid of ccache since it is not present when parsing. - if compiler.startswith('ccache '): - compiler = compiler[7:] - try: - env = os.environ.copy() - env["PATH"] = d.getVar("PATH") - output = subprocess.check_output("%s --version" % compiler, \ - shell=True, env=env, stderr=subprocess.STDOUT).decode("utf-8") - except subprocess.CalledProcessError as e: - bb.fatal("Error running %s --version: %s" % (compiler, e.output.decode("utf-8"))) - - match = re.match(r".* (\d+\.\d+)\.\d+.*", output.split('\n')[0]) - if not match: - bb.fatal("Can't get compiler version from %s --version output" % compiler) - - version = match.group(1) - return "-%s" % version if version in ("4.8", "4.9") else "" - @bb.parse.vardepsexclude("DEFAULTTUNE_MULTILIB_ORIGINAL", "OVERRIDES") def get_multilib_datastore(variant, d): localdata = bb.data.createCopy(d) if variant: - overrides = localdata.getVar("OVERRIDES", False) + ":virtclass-multilib-" + variant + overrides = ( + localdata.getVar("OVERRIDES", False) + ":virtclass-multilib-" + variant + ) localdata.setVar("OVERRIDES", overrides) localdata.setVar("MLPREFIX", variant + "-") else: @@ -482,25 +503,32 @@ def get_multilib_datastore(variant, d): if origdefault: localdata.setVar("DEFAULTTUNE", origdefault) overrides = localdata.getVar("OVERRIDES", False).split(":") - overrides = ":".join([x for x in overrides if not x.startswith("virtclass-multilib-")]) + overrides = ":".join( + [x for x in overrides if not x.startswith("virtclass-multilib-")] + ) localdata.setVar("OVERRIDES", overrides) localdata.setVar("MLPREFIX", "") return localdata + def sh_quote(string): import shlex + return shlex.quote(string) + def directory_size(root, blocksize=4096): """ Calculate the size of the directory, taking into account hard links, rounding up every size to multiples of the blocksize. """ + def roundup(size): """ Round the size up to the nearest multiple of the block size. """ import math + return math.ceil(size / blocksize) * blocksize def getsize(filename): @@ -522,6 +550,7 @@ def directory_size(root, blocksize=4096): total += roundup(getsize(root)) return total + # Update the mtime of a file, skip if permission/read-only issues def touch(filename): try: