build: refactor configure.py

- Explicitly specify the encoding when opening files.
- Use f-strings to format strings.
- Use `isinstance()` for type checks instead of `type()`.
- Use the `with` keyword for resource-allocating operations.
- Avoid using multiple statements in a single line.
- Remove unnecessary `else` clauses after `return`.
- Iterate with the `items()` method of dictionaries when both the key
and value are used.
- Remove unnecessary parentheses.
- Rename unused unpacked variables to `_`, `_1`, etc etc.
- Rename the `list` variable to avoid conflict with the global
`list()` function.

PR-URL: https://github.com/nodejs/node/pull/47667
Reviewed-By: Christian Clauss <cclauss@me.com>
This commit is contained in:
Mohammed Keyvanzadeh 2023-05-01 12:43:16 +03:30 committed by GitHub
parent b5fe45fb9a
commit a4cf6b204f
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -53,7 +53,7 @@ valid_mips_arch = ('loongson', 'r1', 'r2', 'r6', 'rx')
valid_mips_fpu = ('fp32', 'fp64', 'fpxx')
valid_mips_float_abi = ('soft', 'hard')
valid_intl_modes = ('none', 'small-icu', 'full-icu', 'system-icu')
with open ('tools/icu/icu_versions.json') as f:
with open('tools/icu/icu_versions.json', encoding='utf-8') as f:
icu_versions = json.load(f)
shareable_builtins = {'cjs_module_lexer/lexer': 'deps/cjs-module-lexer/lexer.js',
@ -108,7 +108,7 @@ parser.add_argument('--dest-cpu',
action='store',
dest='dest_cpu',
choices=valid_arch,
help='CPU architecture to build for ({0})'.format(', '.join(valid_arch)))
help=f"CPU architecture to build for ({', '.join(valid_arch)})")
parser.add_argument('--cross-compiling',
action='store_true',
@ -125,7 +125,7 @@ parser.add_argument('--dest-os',
action='store',
dest='dest_os',
choices=valid_os,
help='operating system to build for ({0})'.format(', '.join(valid_os)))
help=f"operating system to build for ({', '.join(valid_os)})")
parser.add_argument('--error-on-warn',
action='store_true',
@ -510,39 +510,34 @@ parser.add_argument('--with-arm-float-abi',
action='store',
dest='arm_float_abi',
choices=valid_arm_float_abi,
help='specifies which floating-point ABI to use ({0}).'.format(
', '.join(valid_arm_float_abi)))
help=f"specifies which floating-point ABI to use ({', '.join(valid_arm_float_abi)}).")
parser.add_argument('--with-arm-fpu',
action='store',
dest='arm_fpu',
choices=valid_arm_fpu,
help='ARM FPU mode ({0}) [default: %(default)s]'.format(
', '.join(valid_arm_fpu)))
help=f"ARM FPU mode ({', '.join(valid_arm_fpu)}) [default: %(default)s]")
parser.add_argument('--with-mips-arch-variant',
action='store',
dest='mips_arch_variant',
default='r2',
choices=valid_mips_arch,
help='MIPS arch variant ({0}) [default: %(default)s]'.format(
', '.join(valid_mips_arch)))
help=f"MIPS arch variant ({', '.join(valid_mips_arch)}) [default: %(default)s]")
parser.add_argument('--with-mips-fpu-mode',
action='store',
dest='mips_fpu_mode',
default='fp32',
choices=valid_mips_fpu,
help='MIPS FPU mode ({0}) [default: %(default)s]'.format(
', '.join(valid_mips_fpu)))
help=f"MIPS FPU mode ({', '.join(valid_mips_fpu)}) [default: %(default)s]")
parser.add_argument('--with-mips-float-abi',
action='store',
dest='mips_float_abi',
default='hard',
choices=valid_mips_float_abi,
help='MIPS floating-point ABI ({0}) [default: %(default)s]'.format(
', '.join(valid_mips_float_abi)))
help=f"MIPS floating-point ABI ({', '.join(valid_mips_float_abi)}) [default: %(default)s]")
parser.add_argument('--use-largepages',
action='store_true',
@ -569,8 +564,7 @@ intl_optgroup.add_argument('--with-intl',
dest='with_intl',
default='full-icu',
choices=valid_intl_modes,
help='Intl mode (valid choices: {0}) [default: %(default)s]'.format(
', '.join(valid_intl_modes)))
help=f"Intl mode (valid choices: {', '.join(valid_intl_modes)}) [default: %(default)s]")
intl_optgroup.add_argument('--without-intl',
action='store_const',
@ -597,7 +591,7 @@ intl_optgroup.add_argument('--with-icu-source',
dest='with_icu_source',
help='Intl mode: optional local path to icu/ dir, or path/URL of '
'the icu4c source archive. '
'v%d.x or later recommended.' % icu_versions['minimum_icu'])
f"v{icu_versions['minimum_icu']}.x or later recommended.")
intl_optgroup.add_argument('--with-icu-default-data-dir',
action='store',
@ -853,25 +847,25 @@ auto_downloads = nodedownload.parse(options.download_list)
def error(msg):
prefix = '\033[1m\033[31mERROR\033[0m' if os.isatty(1) else 'ERROR'
print('%s: %s' % (prefix, msg))
print(f'{prefix}: {msg}')
sys.exit(1)
def warn(msg):
warn.warned = True
prefix = '\033[1m\033[93mWARNING\033[0m' if os.isatty(1) else 'WARNING'
print('%s: %s' % (prefix, msg))
print(f'{prefix}: {msg}')
# track if warnings occurred
warn.warned = False
def info(msg):
prefix = '\033[1m\033[32mINFO\033[0m' if os.isatty(1) else 'INFO'
print('%s: %s' % (prefix, msg))
print(f'{prefix}: {msg}')
def print_verbose(x):
if not options.verbose:
return
if type(x) is str:
if isinstance(x, str):
print(x)
else:
pprint.pprint(x, indent=2)
@ -904,9 +898,11 @@ def pkg_config(pkg):
try:
proc = subprocess.Popen(shlex.split(pkg_config) + args,
stdout=subprocess.PIPE)
val = to_utf8(proc.communicate()[0]).strip()
with proc:
val = to_utf8(proc.communicate()[0]).strip()
except OSError as e:
if e.errno != errno.ENOENT: raise e # Unexpected error.
if e.errno != errno.ENOENT:
raise e # Unexpected error.
return (None, None, None, None) # No pkg-config/pkgconf installed.
retval.append(val)
args = ['--silence-errors']
@ -920,13 +916,14 @@ def try_check_compiler(cc, lang):
except OSError:
return (False, False, '', '')
proc.stdin.write(b'__clang__ __GNUC__ __GNUC_MINOR__ __GNUC_PATCHLEVEL__ '
b'__clang_major__ __clang_minor__ __clang_patchlevel__')
with proc:
proc.stdin.write(b'__clang__ __GNUC__ __GNUC_MINOR__ __GNUC_PATCHLEVEL__ '
b'__clang_major__ __clang_minor__ __clang_patchlevel__')
if sys.platform == 'zos':
values = (to_utf8(proc.communicate()[0]).split('\n')[-2].split() + ['0'] * 7)[0:7]
else:
values = (to_utf8(proc.communicate()[0]).split() + ['0'] * 7)[0:7]
if sys.platform == 'zos':
values = (to_utf8(proc.communicate()[0]).split('\n')[-2].split() + ['0'] * 7)[0:7]
else:
values = (to_utf8(proc.communicate()[0]).split() + ['0'] * 7)[0:7]
is_clang = values[0] == '1'
gcc_version = tuple(map(int, values[1:1+3]))
@ -952,12 +949,10 @@ def get_version_helper(cc, regexp):
consider adjusting the CC environment variable if you installed
it in a non-standard prefix.''')
match = re.search(regexp, to_utf8(proc.communicate()[1]))
with proc:
match = re.search(regexp, to_utf8(proc.communicate()[1]))
if match:
return match.group(2)
else:
return '0.0'
return match.group(2) if match else '0.0'
def get_nasm_version(asm):
try:
@ -970,13 +965,11 @@ def get_nasm_version(asm):
and refer BUILDING.md.''')
return '0.0'
match = re.match(r"NASM version ([2-9]\.[0-9][0-9]+)",
to_utf8(proc.communicate()[0]))
with proc:
match = re.match(r"NASM version ([2-9]\.[0-9][0-9]+)",
to_utf8(proc.communicate()[0]))
if match:
return match.group(1)
else:
return '0.0'
return match.group(1) if match else '0.0'
def get_llvm_version(cc):
return get_version_helper(
@ -1002,14 +995,16 @@ def get_gas_version(cc):
consider adjusting the CC environment variable if you installed
it in a non-standard prefix.''')
gas_ret = to_utf8(proc.communicate()[1])
with proc:
gas_ret = to_utf8(proc.communicate()[1])
match = re.match(r"GNU assembler version ([2-9]\.[0-9]+)", gas_ret)
if match:
return match.group(1)
else:
warn('Could not recognize `gas`: ' + gas_ret)
return '0.0'
warn(f'Could not recognize `gas`: {gas_ret}')
return '0.0'
# Note: Apple clang self-reports as clang 4.2.0 and gcc 4.2.1. It passes
# the version check more by accident than anything else but a more rigorous
@ -1027,26 +1022,22 @@ def check_compiler(o):
ok, is_clang, clang_version, gcc_version = try_check_compiler(CXX, 'c++')
version_str = ".".join(map(str, clang_version if is_clang else gcc_version))
print_verbose('Detected %sC++ compiler (CXX=%s) version: %s' %
('clang ' if is_clang else '', CXX, version_str))
print_verbose(f"Detected {'clang ' if is_clang else ''}C++ compiler (CXX={CXX}) version: {version_str}")
if not ok:
warn('failed to autodetect C++ compiler version (CXX=%s)' % CXX)
warn(f'failed to autodetect C++ compiler version (CXX={CXX})')
elif clang_version < (8, 0, 0) if is_clang else gcc_version < (10, 1, 0):
warn('C++ compiler (CXX=%s, %s) too old, need g++ 10.1.0 or clang++ 8.0.0' %
(CXX, version_str))
warn(f'C++ compiler (CXX={CXX}, {version_str}) too old, need g++ 10.1.0 or clang++ 8.0.0')
ok, is_clang, clang_version, gcc_version = try_check_compiler(CC, 'c')
version_str = ".".join(map(str, clang_version if is_clang else gcc_version))
print_verbose('Detected %sC compiler (CC=%s) version: %s' %
('clang ' if is_clang else '', CC, version_str))
print_verbose(f"Detected {'clang ' if is_clang else ''}C compiler (CC={CC}) version: {version_str}")
if not ok:
warn('failed to autodetect C compiler version (CC=%s)' % CC)
warn(f'failed to autodetect C compiler version (CC={CC})')
elif not is_clang and gcc_version < (4, 2, 0):
# clang 3.2 is a little white lie because any clang version will probably
# do for the C bits. However, we might as well encourage people to upgrade
# to a version that is not completely ancient.
warn('C compiler (CC=%s, %s) too old, need gcc 4.2 or clang 3.2' %
(CC, version_str))
warn(f'C compiler (CC={CC}, {version_str}) too old, need gcc 4.2 or clang 3.2')
o['variables']['llvm_version'] = get_llvm_version(CC) if is_clang else '0.0'
@ -1076,8 +1067,9 @@ def cc_macros(cc=None):
consider adjusting the CC environment variable if you installed
it in a non-standard prefix.''')
p.stdin.write(b'\n')
out = to_utf8(p.communicate()[0]).split('\n')
with p:
p.stdin.write(b'\n')
out = to_utf8(p.communicate()[0]).split('\n')
k = {}
for line in out:
@ -1134,9 +1126,9 @@ def host_arch_cc():
rtn = 'ia32' # default
for i in matchup:
if i in k and k[i] != '0':
rtn = matchup[i]
for key, value in matchup.items():
if k.get(key, 0) and k[key] != '0':
rtn = value
break
if rtn == 'mipsel' and '_LP64' in k:
@ -1195,7 +1187,7 @@ def configure_arm(o):
def configure_mips(o, target_arch):
can_use_fpu_instructions = (options.mips_float_abi != 'soft')
can_use_fpu_instructions = options.mips_float_abi != 'soft'
o['variables']['v8_can_use_fpu_instructions'] = b(can_use_fpu_instructions)
o['variables']['v8_use_mips_abi_hardfloat'] = b(can_use_fpu_instructions)
o['variables']['mips_arch_variant'] = options.mips_arch_variant
@ -1214,16 +1206,18 @@ def configure_zos(o):
def clang_version_ge(version_checked):
for compiler in [(CC, 'c'), (CXX, 'c++')]:
ok, is_clang, clang_version, gcc_version = \
_, is_clang, clang_version, _1 = (
try_check_compiler(compiler[0], compiler[1])
)
if is_clang and clang_version >= version_checked:
return True
return False
def gcc_version_ge(version_checked):
for compiler in [(CC, 'c'), (CXX, 'c++')]:
ok, is_clang, clang_version, gcc_version = \
_, is_clang, _1, gcc_version = (
try_check_compiler(compiler[0], compiler[1])
)
if is_clang or gcc_version < version_checked:
return False
return True
@ -1324,7 +1318,7 @@ def configure_node(o):
version_checked_str = ".".join(map(str, version_checked))
raise Exception(
'The options --enable-pgo-generate and --enable-pgo-use '
'are supported for gcc and gxx %s or newer only.' % (version_checked_str))
f'are supported for gcc and gxx {version_checked_str} or newer only.')
if options.enable_pgo_generate and options.enable_pgo_use:
raise Exception(
@ -1347,8 +1341,8 @@ def configure_node(o):
gcc_version_checked_str = ".".join(map(str, gcc_version_checked))
clang_version_checked_str = ".".join(map(str, clang_version_checked))
raise Exception(
'The option --enable-lto is supported for gcc %s+'
'or clang %s+ only.' % (gcc_version_checked_str, clang_version_checked_str))
f'The option --enable-lto is supported for gcc {gcc_version_checked_str}+'
f'or clang {clang_version_checked_str}+ only.')
o['variables']['enable_lto'] = b(options.enable_lto)
@ -1458,15 +1452,15 @@ def configure_library(lib, output, pkgname=None):
if 'msvs_settings' not in output:
output['msvs_settings'] = { 'VCLinkerTool': { 'AdditionalOptions': [] } }
output['msvs_settings']['VCLinkerTool']['AdditionalOptions'] += [
'/LIBPATH:%s' % options.__dict__[shared_lib + '_libpath']]
f"/LIBPATH:{options.__dict__[shared_lib + '_libpath']}"]
else:
output['libraries'] += [
'-L%s' % options.__dict__[shared_lib + '_libpath']]
f"-L{options.__dict__[shared_lib + '_libpath']}"]
elif pkg_libpath:
output['libraries'] += [pkg_libpath]
default_libs = getattr(options, shared_lib + '_libname')
default_libs = ['-l{0}'.format(l) for l in default_libs.split(',')]
default_libs = [f'-l{l}' for l in default_libs.split(',')]
if default_libs:
output['libraries'] += default_libs
@ -1528,7 +1522,7 @@ def configure_openssl(o):
if options.without_ssl:
def without_ssl_error(option):
error('--without-ssl is incompatible with %s' % option)
error(f'--without-ssl is incompatible with {option}')
if options.shared_openssl:
without_ssl_error('--shared-openssl')
if options.openssl_no_asm:
@ -1608,35 +1602,35 @@ def configure_static(o):
def write(filename, data):
print_verbose('creating %s' % filename)
with open(filename, 'w+') as f:
print_verbose(f'creating {filename}')
with open(filename, 'w+', encoding='utf-8') as f:
f.write(data)
do_not_edit = '# Do not edit. Generated by the configure script.\n'
def glob_to_var(dir_base, dir_sub, patch_dir):
list = []
dir_all = '%s/%s' % (dir_base, dir_sub)
file_list = []
dir_all = f'{dir_base}/{dir_sub}'
files = os.walk(dir_all)
for ent in files:
(path, dirs, files) = ent
(_, _1, files) = ent
for file in files:
if file.endswith(('.cpp', '.c', '.h')):
# srcfile uses "slash" as dir separator as its output is consumed by gyp
srcfile = '%s/%s' % (dir_sub, file)
srcfile = f'{dir_sub}/{file}'
if patch_dir:
patchfile = '%s/%s/%s' % (dir_base, patch_dir, file)
patchfile = f'{dir_base}{patch_dir}{file}'
if os.path.isfile(patchfile):
srcfile = '%s/%s' % (patch_dir, file)
info('Using floating patch "%s" from "%s"' % (patchfile, dir_base))
list.append(srcfile)
srcfile = f'{patch_dir}/{file}'
info(f'Using floating patch "{patchfile}" from "{dir_base}"')
file_list.append(srcfile)
break
return list
return file_list
def configure_intl(o):
def icu_download(path):
depFile = 'tools/icu/current_ver.dep'
with open(depFile) as f:
with open(depFile, encoding='utf-8') as f:
icus = json.load(f)
# download ICU, if needed
if not os.access(options.download_path, os.W_OK):
@ -1647,26 +1641,26 @@ def configure_intl(o):
url = icu['url']
(expectHash, hashAlgo, allAlgos) = nodedownload.findHash(icu)
if not expectHash:
error('''Could not find a hash to verify ICU download.
%s may be incorrect.
For the entry %s,
Expected one of these keys: %s''' % (depFile, url, ' '.join(allAlgos)))
error(f'''Could not find a hash to verify ICU download.
{depFile} may be incorrect.
For the entry {url},
Expected one of these keys: {' '.join(allAlgos)}''')
local = url.split('/')[-1]
targetfile = os.path.join(options.download_path, local)
if not os.path.isfile(targetfile):
if attemptdownload:
nodedownload.retrievefile(url, targetfile)
else:
print('Re-using existing %s' % targetfile)
print(f'Re-using existing {targetfile}')
if os.path.isfile(targetfile):
print('Checking file integrity with %s:\r' % hashAlgo)
print(f'Checking file integrity with {hashAlgo}:\r')
gotHash = nodedownload.checkHash(targetfile, hashAlgo)
print('%s: %s %s' % (hashAlgo, gotHash, targetfile))
if (expectHash == gotHash):
print(f'{hashAlgo}: {gotHash} {targetfile}')
if expectHash == gotHash:
return targetfile
else:
warn('Expected: %s *MISMATCH*' % expectHash)
warn('\n ** Corrupted ZIP? Delete %s to retry download.\n' % targetfile)
warn(f'Expected: {expectHash} *MISMATCH*')
warn(f'\n ** Corrupted ZIP? Delete {targetfile} to retry download.\n')
return None
icu_config = {
'variables': {}
@ -1694,12 +1688,14 @@ def configure_intl(o):
# use the .gyp given
o['variables']['icu_gyp_path'] = options.with_icu_path
return
# --with-intl=<with_intl>
# set the default
if with_intl in (None, 'none'):
o['variables']['v8_enable_i18n_support'] = 0
return # no Intl
elif with_intl == 'small-icu':
if with_intl == 'small-icu':
# small ICU (English only)
o['variables']['v8_enable_i18n_support'] = 1
o['variables']['icu_small'] = b(True)
@ -1722,8 +1718,7 @@ def configure_intl(o):
icu_ver_major = icuversion.split('.')[0]
o['variables']['icu_ver_major'] = icu_ver_major
if int(icu_ver_major) < icu_versions['minimum_icu']:
error('icu4c v%s is too old, v%d.x or later is required.' %
(icuversion, icu_versions['minimum_icu']))
error(f"icu4c v{icuversion} is too old, v{icu_versions['minimum_icu']}.x or later is required.")
# libpath provides linker path which may contain spaces
if libpath:
o['libraries'] += [libpath]
@ -1753,7 +1748,7 @@ def configure_intl(o):
canned_is_full = os.path.isfile(os.path.join(canned_icu_dir, 'README-FULL-ICU.txt'))
canned_is_small = os.path.isfile(os.path.join(canned_icu_dir, 'README-SMALL-ICU.txt'))
if canned_is_small:
warn('Ignoring %s - in-repo small icu is no longer supported.' % canned_icu_dir)
warn(f'Ignoring {canned_icu_dir} - in-repo small icu is no longer supported.')
# We can use 'deps/icu-small' - pre-canned ICU *iff*
# - canned_is_full AND
@ -1772,17 +1767,17 @@ def configure_intl(o):
# --with-icu-source processing
# now, check that they didn't pass --with-icu-source=deps/icu
elif with_icu_source and os.path.abspath(icu_full_path) == os.path.abspath(with_icu_source):
warn('Ignoring redundant --with-icu-source=%s' % with_icu_source)
warn(f'Ignoring redundant --with-icu-source={with_icu_source}')
with_icu_source = None
# if with_icu_source is still set, try to use it.
if with_icu_source:
if os.path.isdir(icu_full_path):
print('Deleting old ICU source: %s' % icu_full_path)
print(f'Deleting old ICU source: {icu_full_path}')
shutil.rmtree(icu_full_path)
# now, what path was given?
if os.path.isdir(with_icu_source):
# it's a path. Copy it.
print('%s -> %s' % (with_icu_source, icu_full_path))
print(f'{with_icu_source} -> {icu_full_path}')
shutil.copytree(with_icu_source, icu_full_path)
else:
# could be file or URL.
@ -1807,8 +1802,7 @@ def configure_intl(o):
shutil.rmtree(icu_tmp_path)
else:
shutil.rmtree(icu_tmp_path)
error('--with-icu-source=%s did not result in an "icu" dir.' % \
with_icu_source)
error(f'--with-icu-source={with_icu_source} did not result in an "icu" dir.')
# ICU mode. (icu-generic.gyp)
o['variables']['icu_gyp_path'] = 'tools/icu/icu-generic.gyp'
@ -1820,17 +1814,17 @@ def configure_intl(o):
if localzip:
nodedownload.unpack(localzip, icu_parent_path)
else:
warn('* ECMA-402 (Intl) support didn\'t find ICU in %s..' % icu_full_path)
warn("* ECMA-402 (Intl) support didn't find ICU in {icu_full_path}..")
if not os.path.isdir(icu_full_path):
error('''Cannot build Intl without ICU in %s.
Fix, or disable with "--with-intl=none"''' % icu_full_path)
error(f'''Cannot build Intl without ICU in {icu_full_path}.
Fix, or disable with "--with-intl=none"''')
else:
print_verbose('* Using ICU in %s' % icu_full_path)
print_verbose(f'* Using ICU in {icu_full_path}')
# Now, what version of ICU is it? We just need the "major", such as 54.
# uvernum.h contains it as a #define.
uvernum_h = os.path.join(icu_full_path, 'source/common/unicode/uvernum.h')
if not os.path.isfile(uvernum_h):
error('Could not load %s - is ICU installed?' % uvernum_h)
error(f'Could not load {uvernum_h} - is ICU installed?')
icu_ver_major = None
matchVerExp = r'^\s*#define\s+U_ICU_VERSION_SHORT\s+"([^"]*)".*'
match_version = re.compile(matchVerExp)
@ -1840,20 +1834,19 @@ def configure_intl(o):
if m:
icu_ver_major = str(m.group(1))
if not icu_ver_major:
error('Could not read U_ICU_VERSION_SHORT version from %s' % uvernum_h)
error(f'Could not read U_ICU_VERSION_SHORT version from {uvernum_h}')
elif int(icu_ver_major) < icu_versions['minimum_icu']:
error('icu4c v%s.x is too old, v%d.x or later is required.' %
(icu_ver_major, icu_versions['minimum_icu']))
error(f"icu4c v{icu_ver_major}.x is too old, v{icu_versions['minimum_icu']}.x or later is required.")
icu_endianness = sys.byteorder[0]
o['variables']['icu_ver_major'] = icu_ver_major
o['variables']['icu_endianness'] = icu_endianness
icu_data_file_l = 'icudt%s%s.dat' % (icu_ver_major, 'l') # LE filename
icu_data_file = 'icudt%s%s.dat' % (icu_ver_major, icu_endianness)
icu_data_file_l = f'icudt{icu_ver_major}l.dat' # LE filename
icu_data_file = f'icudt{icu_ver_major}{icu_endianness}.dat'
# relative to configure
icu_data_path = os.path.join(icu_full_path,
'source/data/in',
icu_data_file_l) # LE
compressed_data = '%s.bz2' % (icu_data_path)
compressed_data = f'{icu_data_path}.bz2'
if not os.path.isfile(icu_data_path) and os.path.isfile(compressed_data):
# unpack. deps/icu is a temporary path
if os.path.isdir(icu_tmp_path):
@ -1869,7 +1862,7 @@ def configure_intl(o):
# Now, proceed..
# relative to dep..
icu_data_in = os.path.join('..','..', icu_data_path)
icu_data_in = os.path.join('..', '..', icu_data_path)
if not os.path.isfile(icu_data_path) and icu_endianness != 'l':
# use host endianness
icu_data_path = os.path.join(icu_full_path,
@ -1877,8 +1870,8 @@ def configure_intl(o):
icu_data_file) # will be generated
if not os.path.isfile(icu_data_path):
# .. and we're not about to build it from .gyp!
error('''ICU prebuilt data file %s does not exist.
See the README.md.''' % icu_data_path)
error(f'''ICU prebuilt data file {icu_data_path} does not exist.
See the README.md.''')
# this is the input '.dat' file to use .. icudt*.dat
# may be little-endian if from a icu-project.org tarball
@ -1896,10 +1889,10 @@ def configure_intl(o):
}
# this creates a variable icu_src_XXX for each of the subdirs
# with a list of the src files to use
for i in icu_src:
var = 'icu_src_%s' % i
path = '../../%s/source/%s' % (icu_full_path, icu_src[i])
icu_config['variables'][var] = glob_to_var('tools/icu', path, 'patches/%s/source/%s' % (icu_ver_major, icu_src[i]) )
for key, value in icu_src.items():
var = f'icu_src_{key}'
path = f'../../{icu_full_path}/source/{value}'
icu_config['variables'][var] = glob_to_var('tools/icu', path, f'patches/{icu_ver_major}/source/{value}')
# calculate platform-specific genccode args
# print("platform %s, flavor %s" % (sys.platform, flavor))
# if sys.platform == 'darwin':
@ -1950,8 +1943,9 @@ def configure_section_file(o):
warn('''No acceptable ld.gold linker found!''')
return 0
match = re.match(r"^GNU gold.*([0-9]+)\.([0-9]+)$",
proc.communicate()[0].decode("utf-8"))
with proc:
match = re.match(r"^GNU gold.*([0-9]+)\.([0-9]+)$",
proc.communicate()[0].decode("utf-8"))
if match:
gold_major_version = match.group(1)
@ -1983,13 +1977,15 @@ def make_bin_override():
try:
os.makedirs(bin_override)
except OSError as e:
if e.errno != errno.EEXIST: raise e
if e.errno != errno.EEXIST:
raise e
python_link = os.path.join(bin_override, 'python')
try:
os.unlink(python_link)
except OSError as e:
if e.errno != errno.ENOENT: raise e
if e.errno != errno.ENOENT:
raise e
os.symlink(sys.executable, python_link)
# We need to set the environment right now so that when gyp (in run_gyp)
@ -2013,7 +2009,7 @@ check_compiler(output)
# determine the "flavor" (operating system) we're building for,
# leveraging gyp's GetFlavor function
flavor_params = {}
if (options.dest_os):
if options.dest_os:
flavor_params['flavor'] = options.dest_os
flavor = GetFlavor(flavor_params)
@ -2037,12 +2033,12 @@ configure_section_file(output)
# configure shareable builtins
output['variables']['node_builtin_shareable_builtins'] = []
for builtin in shareable_builtins:
for builtin, value in shareable_builtins.items():
builtin_id = 'node_shared_builtin_' + builtin.replace('/', '_') + '_path'
if getattr(options, builtin_id):
output['defines'] += [builtin_id.upper() + '=' + getattr(options, builtin_id)]
else:
output['variables']['node_builtin_shareable_builtins'] += [shareable_builtins[builtin]]
output['variables']['node_builtin_shareable_builtins'] += [value]
# Forward OSS-Fuzz settings
output['variables']['ossfuzz'] = b(options.ossfuzz)