| # Copyright (c) 2016, the R8 project authors. Please see the AUTHORS file |
| # for details. All rights reserved. Use of this source code is governed by a |
| # BSD-style license that can be found in the LICENSE file. |
| |
| # Different utility functions used accross scripts |
| |
| import hashlib |
| import jdk |
| import json |
| import os |
| import re |
| import shutil |
| import subprocess |
| import sys |
| import tarfile |
| import tempfile |
| import zipfile |
| |
| import defines |
| from thread_utils import print_thread |
| |
| ANDROID_JAR_DIR = 'third_party/android_jar/lib-v{api}' |
| ANDROID_JAR = os.path.join(ANDROID_JAR_DIR, 'android.jar') |
| TOOLS_DIR = defines.TOOLS_DIR |
| REPO_ROOT = defines.REPO_ROOT |
| THIRD_PARTY = defines.THIRD_PARTY |
| BUNDLETOOL_JAR_DIR = os.path.join(THIRD_PARTY, 'bundletool/bundletool-1.11.0') |
| BUNDLETOOL_JAR = os.path.join(BUNDLETOOL_JAR_DIR, 'bundletool-all-1.11.0.jar') |
| MEMORY_USE_TMP_FILE = 'memory_use.tmp' |
| DEX_SEGMENTS_RESULT_PATTERN = re.compile('- ([^:]+): ([0-9]+)') |
| |
| DEPENDENCIES_DIR = os.path.join(THIRD_PARTY, 'dependencies') |
| |
| BUILD = os.path.join(REPO_ROOT, 'build') |
| BUILD_JAVA_MAIN_DIR = os.path.join(BUILD, 'classes', 'java', 'main') |
| LIBS = os.path.join(BUILD, 'libs') |
| CUSTOM_CONVERSION_DIR = os.path.join(THIRD_PARTY, 'openjdk', |
| 'custom_conversion') |
| GENERATED_LICENSE_DIR = os.path.join(BUILD, 'generatedLicense') |
| SRC_ROOT = os.path.join(REPO_ROOT, 'src', 'main', 'java') |
| REPO_SOURCE = 'https://r8.googlesource.com/r8' |
| |
| GRADLE_TASK_CLEAN_TEST = ':test:cleanTest' |
| GRADLE_TASK_CONSOLIDATED_LICENSE = ':main:consolidatedLicense' |
| GRADLE_TASK_KEEP_ANNO_JAR = ':keepanno:keepAnnoAnnotationsJar' |
| GRADLE_TASK_KEEP_ANNO_DOC = ':keepanno:keepAnnoAnnotationsDoc' |
| GRADLE_TASK_R8 = ':main:r8WithRelocatedDeps' |
| GRADLE_TASK_R8LIB = ':test:assembleR8LibWithRelocatedDeps' |
| GRADLE_TASK_R8LIB_NO_DEPS = ':test:assembleR8LibNoDeps' |
| GRADLE_TASK_THREADING_MODULE_BLOCKING = ':main:threadingModuleBlockingJar' |
| GRADLE_TASK_THREADING_MODULE_SINGLE_THREADED = ':main:threadingModuleSingleThreadedJar' |
| GRADLE_TASK_SOURCE_JAR = ':test:packageSources' |
| GRADLE_TASK_SWISS_ARMY_KNIFE = ':main:swissArmyKnife' |
| GRADLE_TASK_TEST = ':test:test' |
| GRADLE_TASK_ALL_TESTS_WITH_APPLY_MAPPING_JAR = ':test:rewriteTestsForR8LibWithRelocatedDeps' |
| GRADLE_TASK_TESTBASE_WITH_APPLY_MAPPING_JAR = ':test:rewriteTestBaseForR8LibWithRelocatedDeps' |
| GRADLE_TASK_TEST_DEPS_JAR = ':test:packageTestDeps' |
| GRADLE_TASK_TEST_BASE_JAR = ':test:relocateTestBaseForR8LibWithRelocatedDeps' |
| GRADLE_TASK_TEST_JAR = ':test:relocateTestsForR8LibWithRelocatedDeps' |
| GRADLE_TASK_TEST_UNZIP_TESTBASE = ':test:unzipTestBase' |
| |
| R8 = 'r8' |
| R8LIB = 'r8lib' |
| |
| ALL_DEPS_JAR = os.path.join(LIBS, 'deps_all.jar') |
| R8_JAR = os.path.join(LIBS, 'r8.jar') |
| R8LIB_JAR = os.path.join(LIBS, 'r8lib.jar') |
| R8LIB_MAP = '%s.map' % R8LIB_JAR |
| R8_SRC_JAR = os.path.join(LIBS, 'r8-src.jar') |
| R8LIB_EXCLUDE_DEPS_JAR = os.path.join(LIBS, 'r8lib-exclude-deps.jar') |
| R8_FULL_EXCLUDE_DEPS_JAR = os.path.join(LIBS, 'r8-full-exclude-deps.jar') |
| THREADING_MODULE_BLOCKING_JAR = os.path.join(LIBS, |
| 'threading-module-blocking.jar') |
| THREADING_MODULE_SINGLE_THREADED_JAR = os.path.join( |
| LIBS, 'threading-module-single-threaded.jar') |
| R8_TESTS_JAR = os.path.join(LIBS, 'r8tests.jar') |
| R8_TESTBASE_JAR = os.path.join(LIBS, 'r8test_base.jar') |
| R8LIB_TESTBASE_JAR = os.path.join(LIBS, 'r8libtestbase-cf.jar') |
| R8LIB_TESTS_JAR = os.path.join(LIBS, 'r8libtestdeps-cf.jar') |
| R8_TESTS_DEPS_JAR = os.path.join(LIBS, 'test_deps_all.jar') |
| R8LIB_TESTS_DEPS_JAR = R8_TESTS_DEPS_JAR |
| MAVEN_ZIP_LIB = os.path.join(LIBS, 'r8lib.zip') |
| LIBRARY_DESUGAR_CONVERSIONS_LEGACY_ZIP = os.path.join( |
| CUSTOM_CONVERSION_DIR, 'library_desugar_conversions_legacy.jar') |
| LIBRARY_DESUGAR_CONVERSIONS_ZIP = os.path.join( |
| CUSTOM_CONVERSION_DIR, 'library_desugar_conversions.jar') |
| KEEPANNO_ANNOTATIONS_JAR = os.path.join(LIBS, 'keepanno-annotations.jar') |
| KEEPANNO_ANNOTATIONS_DOC = os.path.join('d8_r8', 'keepanno', 'build', 'docs', |
| 'javadoc') |
| |
| DESUGAR_CONFIGURATION = os.path.join('src', 'library_desugar', |
| 'desugar_jdk_libs.json') |
| DESUGAR_IMPLEMENTATION = os.path.join('third_party', 'openjdk', |
| 'desugar_jdk_libs', |
| 'desugar_jdk_libs.jar') |
| DESUGAR_CONFIGURATION_JDK11_LEGACY = os.path.join( |
| 'src', 'library_desugar', 'jdk11', 'desugar_jdk_libs_legacy.json') |
| DESUGAR_CONFIGURATION_JDK11_MINIMAL = os.path.join( |
| 'src', 'library_desugar', 'jdk11', 'desugar_jdk_libs_minimal.json') |
| DESUGAR_CONFIGURATION_JDK11 = os.path.join('src', 'library_desugar', 'jdk11', |
| 'desugar_jdk_libs.json') |
| DESUGAR_CONFIGURATION_JDK11_NIO = os.path.join('src', 'library_desugar', |
| 'jdk11', |
| 'desugar_jdk_libs_nio.json') |
| DESUGAR_IMPLEMENTATION_JDK11 = os.path.join('third_party', 'openjdk', |
| 'desugar_jdk_libs_11', |
| 'desugar_jdk_libs.jar') |
| DESUGAR_CONFIGURATION_MAVEN_ZIP = os.path.join( |
| LIBS, 'desugar_jdk_libs_configuration.zip') |
| DESUGAR_CONFIGURATION_JDK11_LEGACY_MAVEN_ZIP = os.path.join( |
| LIBS, 'desugar_jdk_libs_configuration_jdk11_legacy.zip') |
| DESUGAR_CONFIGURATION_JDK11_MINIMAL_MAVEN_ZIP = os.path.join( |
| LIBS, 'desugar_jdk_libs_configuration_jdk11_minimal.zip') |
| DESUGAR_CONFIGURATION_JDK11_MAVEN_ZIP = os.path.join( |
| LIBS, 'desugar_jdk_libs_configuration_jdk11.zip') |
| DESUGAR_CONFIGURATION_JDK11_NIO_MAVEN_ZIP = os.path.join( |
| LIBS, 'desugar_jdk_libs_configuration_jdk11_nio.zip') |
| GENERATED_LICENSE = os.path.join(GENERATED_LICENSE_DIR, 'LICENSE') |
| RT_JAR = os.path.join(REPO_ROOT, 'third_party/openjdk/openjdk-rt-1.8/rt.jar') |
| R8LIB_KEEP_RULES = os.path.join(REPO_ROOT, 'src/main/keep.txt') |
| CF_SEGMENTS_TOOL = os.path.join(THIRD_PARTY, 'cf_segments') |
| PINNED_R8_JAR = os.path.join(REPO_ROOT, 'third_party/r8/r8.jar') |
| PINNED_PGR8_JAR = os.path.join(REPO_ROOT, 'third_party/r8/r8-pg6.0.1.jar') |
| |
| OPENSOURCE_DUMPS_DIR = os.path.join(THIRD_PARTY, 'opensource-apps') |
| INTERNAL_DUMPS_DIR = os.path.join(THIRD_PARTY, 'internal-apps') |
| BAZEL_SHA_FILE = os.path.join(THIRD_PARTY, 'bazel.tar.gz.sha1') |
| BAZEL_TOOL = os.path.join(THIRD_PARTY, 'bazel') |
| JAVA8_SHA_FILE = os.path.join(THIRD_PARTY, 'openjdk', 'jdk8', |
| 'linux-x86.tar.gz.sha1') |
| JAVA11_SHA_FILE = os.path.join(THIRD_PARTY, 'openjdk', 'jdk-11', |
| 'linux.tar.gz.sha1') |
| DESUGAR_JDK_LIBS_11_SHA_FILE = os.path.join(THIRD_PARTY, 'openjdk', |
| 'desugar_jdk_libs_11.tar.gz.sha1') |
| IGNORE_WARNINGS_RULES = os.path.join(REPO_ROOT, 'src', 'test', |
| 'ignorewarnings.rules') |
| ANDROID_HOME_ENVIROMENT_NAME = "ANDROID_HOME" |
| ANDROID_TOOLS_VERSION_ENVIRONMENT_NAME = "ANDROID_TOOLS_VERSION" |
| USER_HOME = os.path.expanduser('~') |
| |
| R8_TEST_RESULTS_BUCKET = 'r8-test-results' |
| R8_INTERNAL_TEST_RESULTS_BUCKET = 'r8-internal-test-results' |
| |
| |
| def archive_file(name, gs_dir, src_file): |
| gs_file = '%s/%s' % (gs_dir, name) |
| upload_file_to_cloud_storage(src_file, gs_file) |
| |
| |
| def archive_value(name, gs_dir, value): |
| with TempDir() as temp: |
| temparchive = os.path.join(temp, name) |
| with open(temparchive, 'w') as f: |
| f.write(str(value)) |
| archive_file(name, gs_dir, temparchive) |
| |
| |
| def find_cloud_storage_file_from_options(name, options, orElse=None): |
| # Import archive on-demand since archive depends on utils. |
| from archive import GetUploadDestination |
| hash_or_version = find_hash_or_version_from_options(options) |
| if not hash_or_version: |
| return orElse |
| is_hash = options.commit_hash is not None |
| download_path = GetUploadDestination(hash_or_version, name, is_hash) |
| if file_exists_on_cloud_storage(download_path): |
| out = tempfile.NamedTemporaryFile().name |
| download_file_from_cloud_storage(download_path, out) |
| return out |
| else: |
| raise Exception('Could not find file {} from hash/version: {}.'.format( |
| name, hash_or_version)) |
| |
| |
| def find_r8_jar_from_options(options): |
| return find_cloud_storage_file_from_options('r8.jar', options) |
| |
| |
| def find_hash_or_version_from_options(options): |
| if options.tag: |
| return find_hash_or_version_from_tag(options.tag) |
| else: |
| return options.commit_hash or options.version |
| |
| |
| def find_hash_or_version_from_tag(tag_or_hash): |
| info = subprocess.check_output( |
| ['git', 'show', tag_or_hash, '-s', |
| '--format=oneline']).decode('utf-8').splitlines()[-1].split() |
| # The info should be on the following form [hash,"Version",version] |
| if len(info) == 3 and len(info[0]) == 40 and info[1] == "Version": |
| return info[2] |
| return None |
| |
| |
| def getAndroidHome(): |
| return os.environ.get(ANDROID_HOME_ENVIROMENT_NAME, |
| os.path.join(USER_HOME, 'Android', 'Sdk')) |
| |
| |
| def getAndroidBuildTools(): |
| if ANDROID_TOOLS_VERSION_ENVIRONMENT_NAME in os.environ: |
| version = os.environ.get(ANDROID_TOOLS_VERSION_ENVIRONMENT_NAME) |
| build_tools_dir = os.path.join(getAndroidHome(), 'build-tools', version) |
| assert os.path.exists(build_tools_dir) |
| return build_tools_dir |
| else: |
| versions = ['34.0.0', '33.0.1', '32.0.0'] |
| for version in versions: |
| build_tools_dir = os.path.join(getAndroidHome(), 'build-tools', |
| version) |
| if os.path.exists(build_tools_dir): |
| return build_tools_dir |
| raise Exception('Unable to find Android build-tools') |
| |
| |
| def is_python3(): |
| return sys.version_info.major == 3 |
| |
| |
| def Print(s, quiet=False): |
| if quiet: |
| return |
| print(s) |
| |
| |
| def Warn(message): |
| CRED = '\033[91m' |
| CEND = '\033[0m' |
| print(CRED + message + CEND) |
| |
| |
| def PrintCmd(cmd, env=None, quiet=False, worker_id=None): |
| if quiet: |
| return |
| if type(cmd) is list: |
| cmd = ' '.join(cmd) |
| if env: |
| env = ' '.join(['{}=\"{}\"'.format(x, y) for x, y in env.iteritems()]) |
| print_thread('Running: {} {}'.format(env, cmd), worker_id) |
| else: |
| print_thread('Running: {}'.format(cmd), worker_id) |
| # I know this will hit os on windows eventually if we don't do this. |
| sys.stdout.flush() |
| |
| |
| class ProgressLogger(object): |
| CLEAR_LINE = '\033[K' |
| UP = '\033[F' |
| |
| def __init__(self, quiet=False): |
| self._count = 0 |
| self._has_printed = False |
| self._quiet = quiet |
| |
| def log(self, text): |
| if len(text.strip()) == 0: |
| return |
| if self._quiet: |
| if self._has_printed: |
| sys.stdout.write(ProgressLogger.UP + ProgressLogger.CLEAR_LINE) |
| if len(text) > 140: |
| text = text[0:140] + '...' |
| print(text) |
| self._has_printed = True |
| |
| def done(self): |
| if self._quiet and self._has_printed: |
| sys.stdout.write(ProgressLogger.UP + ProgressLogger.CLEAR_LINE) |
| print('') |
| sys.stdout.write(ProgressLogger.UP) |
| |
| |
| def RunCmd(cmd, env_vars=None, quiet=False, fail=True, logging=True): |
| PrintCmd(cmd, env=env_vars, quiet=quiet) |
| env = os.environ.copy() |
| if env_vars: |
| env.update(env_vars) |
| process = subprocess.Popen(cmd, |
| env=env, |
| stdout=subprocess.PIPE, |
| stderr=subprocess.STDOUT) |
| stdout = [] |
| logger = ProgressLogger(quiet=quiet) if logging else None |
| failed = False |
| while True: |
| line = process.stdout.readline().decode('utf-8') |
| if line != '': |
| stripped = line.rstrip() |
| stdout.append(stripped) |
| if logger: |
| logger.log(stripped) |
| # TODO(christofferqa): r8 should fail with non-zero exit code. |
| if ('AssertionError:' in stripped or |
| 'CompilationError:' in stripped or |
| 'CompilationFailedException:' in stripped or |
| 'Compilation failed' in stripped or |
| 'FAILURE:' in stripped or |
| 'org.gradle.api.ProjectConfigurationException' in stripped |
| or 'BUILD FAILED' in stripped): |
| failed = True |
| else: |
| if logger: |
| logger.done() |
| exit_code = process.poll() |
| if exit_code or failed: |
| for line in stdout: |
| Warn(line) |
| if fail: |
| raise subprocess.CalledProcessError( |
| exit_code or -1, cmd, output='\n'.join(stdout)) |
| return stdout |
| |
| |
| def IsWindows(): |
| return defines.IsWindows() |
| |
| |
| def EnsureDepFromGoogleCloudStorage(dep, tgz, sha1, msg): |
| if (not os.path.exists(dep) or not os.path.exists(tgz) or |
| os.path.getmtime(tgz) < os.path.getmtime(sha1)): |
| DownloadFromGoogleCloudStorage(sha1) |
| # Update the mtime of the tar file to make sure we do not run again unless |
| # there is an update. |
| os.utime(tgz, None) |
| else: |
| print('Ensure cloud dependency:', msg, 'present') |
| |
| |
| def DownloadFromX20(sha1_file): |
| download_script = os.path.join(REPO_ROOT, 'tools', 'download_from_x20.py') |
| cmd = [download_script, sha1_file] |
| PrintCmd(cmd) |
| subprocess.check_call(cmd) |
| |
| |
| def DownloadFromGoogleCloudStorage(sha1_file, |
| bucket='r8-deps', |
| auth=False, |
| quiet=False): |
| suffix = '.bat' if IsWindows() else '' |
| download_script = 'download_from_google_storage%s' % suffix |
| cmd = [download_script] |
| if not auth: |
| cmd.append('-n') |
| cmd.extend(['-b', bucket, '-u', '-s', sha1_file]) |
| if not quiet: |
| PrintCmd(cmd) |
| subprocess.check_call(cmd) |
| else: |
| subprocess.check_output(cmd) |
| |
| |
| def get_nth_sha1_from_revision(n, revision): |
| result = subprocess.check_output([ |
| 'git', 'log', revision, f'--skip={n}', '--max-count=1', |
| '--pretty=format:%H' |
| ]).decode('utf-8') |
| return result.strip() |
| |
| |
| def get_sha1(filename): |
| sha1 = hashlib.sha1() |
| with open(filename, 'rb') as f: |
| while True: |
| chunk = f.read(1024 * 1024) |
| if not chunk: |
| break |
| sha1.update(chunk) |
| return sha1.hexdigest() |
| |
| |
| def get_sha1_from_revision(revision): |
| cmd = ['git', 'rev-parse', revision] |
| PrintCmd(cmd) |
| with ChangedWorkingDirectory(REPO_ROOT): |
| return subprocess.check_output(cmd).decode('utf-8').strip() |
| |
| |
| def get_HEAD_branch(): |
| result = subprocess.check_output( |
| ['git', 'rev-parse', '--abbrev-ref', 'HEAD']).decode('utf-8') |
| return result.strip() |
| |
| |
| def get_HEAD_sha1(): |
| return get_sha1_from_revision('HEAD') |
| |
| |
| def get_HEAD_diff_stat(): |
| return subprocess.check_output(['git', 'diff', '--stat']).decode('utf-8') |
| |
| |
| def makedirs_if_needed(path): |
| try: |
| os.makedirs(path) |
| except OSError: |
| if not os.path.isdir(path): |
| raise |
| |
| |
| def get_gsutil(): |
| return 'gsutil.py' if os.name != 'nt' else 'gsutil.py.bat' |
| |
| |
| def upload_file_to_cloud_storage(source, destination, header=None): |
| cmd = [get_gsutil()] |
| if header: |
| cmd.extend(['-h', header]) |
| cmd.extend(['cp', source, destination]) |
| PrintCmd(cmd) |
| subprocess.check_call(cmd) |
| |
| |
| def check_dir_args(source, destination): |
| # We require that the dirname of the paths coincide, e.g., src/dirname and dst/dirname |
| # The target is then stripped so the upload command will be: cp -R src/dirname dst/ |
| (destination_parent, destination_file) = os.path.split(destination) |
| if os.path.basename(source) != destination_file: |
| raise Exception( |
| 'Attempt to upload directory with non-matching directory name: ' + |
| f'{source} and {destination}') |
| if len(destination_parent.strip()) == 0: |
| raise Exception( |
| 'Attempt to upload directory to empty destination directory: ' + |
| destination) |
| return destination_parent |
| |
| |
| def upload_directory_to_cloud_storage(source, destination, parallel=True): |
| destination_parent = check_dir_args(source, destination) |
| cmd = [get_gsutil()] |
| if parallel: |
| cmd += ['-m'] |
| cmd += ['cp', '-R'] |
| cmd += [source, destination_parent + '/'] |
| PrintCmd(cmd) |
| subprocess.check_call(cmd) |
| |
| |
| def rsync_directory_to_cloud_storage(source, destination, parallel=True): |
| check_dir_args(source, destination) |
| cmd = [get_gsutil()] |
| if parallel: |
| cmd += ['-m'] |
| cmd += ['rsync', '-R'] |
| cmd += [source, destination] |
| PrintCmd(cmd) |
| subprocess.check_call(cmd) |
| |
| |
| def delete_file_from_cloud_storage(destination): |
| cmd = [get_gsutil(), 'rm', destination] |
| PrintCmd(cmd) |
| subprocess.check_call(cmd) |
| |
| |
| def ls_files_on_cloud_storage(destination): |
| cmd = [get_gsutil(), 'ls', destination] |
| PrintCmd(cmd) |
| return subprocess.check_output(cmd).decode('utf-8') |
| |
| |
| def cat_file_on_cloud_storage(destination, ignore_errors=False): |
| cmd = [get_gsutil(), 'cat', destination] |
| PrintCmd(cmd) |
| try: |
| return subprocess.check_output(cmd).decode('utf-8').strip() |
| except subprocess.CalledProcessError as e: |
| if ignore_errors: |
| return '' |
| else: |
| raise e |
| |
| |
| def file_exists_on_cloud_storage(destination): |
| cmd = [get_gsutil(), 'ls', destination] |
| PrintCmd(cmd) |
| return subprocess.call(cmd) == 0 |
| |
| |
| def download_file_from_cloud_storage(source, |
| destination, |
| concurrent=False, |
| flags=None, |
| quiet=False): |
| cmd = [get_gsutil()] |
| if concurrent: |
| cmd.append('-m') |
| cmd.append('cp') |
| if flags: |
| cmd.extend(flags) |
| cmd.extend([source, destination]) |
| PrintCmd(cmd, quiet=quiet) |
| subprocess.check_call(cmd) |
| |
| |
| def create_archive(name, sources=None): |
| if not sources: |
| sources = [name] |
| tarname = '%s.tar.gz' % name |
| with tarfile.open(tarname, 'w:gz') as tar: |
| for source in sources: |
| tar.add(source) |
| return tarname |
| |
| |
| def extract_dir(filename): |
| return filename[0:len(filename) - len('.tar.gz')] |
| |
| |
| def unpack_archive(filename): |
| dest_dir = extract_dir(filename) |
| if os.path.exists(dest_dir): |
| print('Deleting existing dir %s' % dest_dir) |
| shutil.rmtree(dest_dir) |
| dirname = os.path.dirname(os.path.abspath(filename)) |
| with tarfile.open(filename, 'r:gz') as tar: |
| tar.extractall(path=dirname) |
| |
| |
| def check_gcert(): |
| status = subprocess.call(['gcertstatus']) |
| if status != 0: |
| subprocess.check_call(['gcert']) |
| |
| |
| # Note that gcs is eventually consistent with regards to list operations. |
| # This is not a problem in our case, but don't ever use this method |
| # for synchronization. |
| def cloud_storage_exists(destination): |
| cmd = [get_gsutil(), 'ls', destination] |
| PrintCmd(cmd) |
| exit_code = subprocess.call(cmd) |
| return exit_code == 0 |
| |
| |
| class TempDir(object): |
| |
| def __init__(self, prefix='', delete=True): |
| self._temp_dir = None |
| self._prefix = prefix |
| self._delete = delete |
| |
| def __enter__(self): |
| self._temp_dir = tempfile.mkdtemp(self._prefix) |
| return self._temp_dir |
| |
| def __exit__(self, *_): |
| if self._delete: |
| shutil.rmtree(self._temp_dir, ignore_errors=True) |
| |
| |
| class ChangedWorkingDirectory(object): |
| |
| def __init__(self, working_directory, quiet=False): |
| self._quiet = quiet |
| self._working_directory = working_directory |
| |
| def __enter__(self): |
| self._old_cwd = os.getcwd() |
| if not self._quiet: |
| print('Enter directory:', self._working_directory) |
| os.chdir(self._working_directory) |
| |
| def __exit__(self, *_): |
| if not self._quiet: |
| print('Enter directory:', self._old_cwd) |
| os.chdir(self._old_cwd) |
| |
| |
| # Reading Android CTS test_result.xml |
| |
| |
| class CtsModule(object): |
| |
| def __init__(self, module_name): |
| self.name = module_name |
| |
| |
| class CtsTestCase(object): |
| |
| def __init__(self, test_case_name): |
| self.name = test_case_name |
| |
| |
| class CtsTest(object): |
| |
| def __init__(self, test_name, outcome): |
| self.name = test_name |
| self.outcome = outcome |
| |
| |
| # Generator yielding CtsModule, CtsTestCase or CtsTest from |
| # reading through a CTS test_result.xml file. |
| def read_cts_test_result(file_xml): |
| re_module = re.compile('<Module name="([^"]*)"') |
| re_test_case = re.compile('<TestCase name="([^"]*)"') |
| re_test = re.compile('<Test result="(pass|fail)" name="([^"]*)"') |
| with open(file_xml) as f: |
| for line in f: |
| m = re_module.search(line) |
| if m: |
| yield CtsModule(m.groups()[0]) |
| continue |
| m = re_test_case.search(line) |
| if m: |
| yield CtsTestCase(m.groups()[0]) |
| continue |
| m = re_test.search(line) |
| if m: |
| outcome = m.groups()[0] |
| assert outcome in ['fail', 'pass'] |
| yield CtsTest(m.groups()[1], outcome == 'pass') |
| |
| |
| def grep_memoryuse(logfile): |
| re_vmhwm = re.compile('^VmHWM:[ \t]*([0-9]+)[ \t]*([a-zA-Z]*)') |
| result = None |
| with open(logfile) as f: |
| for line in f: |
| m = re_vmhwm.search(line) |
| if m: |
| groups = m.groups() |
| s = len(groups) |
| if s >= 1: |
| result = int(groups[0]) |
| if s >= 2: |
| unit = groups[1] |
| if unit == 'kB': |
| result *= 1024 |
| elif unit != '': |
| raise Exception( |
| 'Unrecognized unit in memory usage log: {}'. |
| format(unit)) |
| if result is None: |
| raise Exception('No memory usage found in log: {}'.format(logfile)) |
| return result |
| |
| |
| # Return a dictionary: {segment_name -> segments_size} |
| def getDexSegmentSizes(dex_files): |
| assert len(dex_files) > 0 |
| cmd = [jdk.GetJavaExecutable(), '-jar', R8_JAR, 'dexsegments'] |
| cmd.extend(dex_files) |
| PrintCmd(cmd) |
| output = subprocess.check_output(cmd).decode('utf-8') |
| |
| matches = DEX_SEGMENTS_RESULT_PATTERN.findall(output) |
| |
| if matches is None or len(matches) == 0: |
| raise Exception('DexSegments failed to return any output for' \ |
| ' these files: {}'.format(dex_files)) |
| |
| result = {} |
| |
| for match in matches: |
| result[match[0]] = int(match[1]) |
| |
| return result |
| |
| |
| # Return a dictionary: {segment_name -> segments_size} |
| def getCfSegmentSizes(cfFile): |
| cmd = [ |
| jdk.GetJavaExecutable(), '-cp', CF_SEGMENTS_TOOL, |
| 'com.android.tools.r8.cf_segments.MeasureLib', cfFile |
| ] |
| PrintCmd(cmd) |
| output = subprocess.check_output(cmd).decode('utf-8') |
| |
| matches = DEX_SEGMENTS_RESULT_PATTERN.findall(output) |
| |
| if matches is None or len(matches) == 0: |
| raise Exception('CfSegments failed to return any output for' \ |
| ' the file: ' + cfFile) |
| |
| result = {} |
| |
| for match in matches: |
| result[match[0]] = int(match[1]) |
| |
| return result |
| |
| |
| def get_maven_path(artifact, version): |
| return os.path.join('com', 'android', 'tools', artifact, version) |
| |
| |
| def print_cfsegments(prefix, cf_files): |
| for cf_file in cf_files: |
| for segment_name, size in getCfSegmentSizes(cf_file).items(): |
| print('{}-{}(CodeSize): {}'.format(prefix, segment_name, size)) |
| |
| |
| def print_dexsegments(prefix, dex_files, worker_id=None): |
| for segment_name, size in getDexSegmentSizes(dex_files).items(): |
| print_thread('{}-{}(CodeSize): {}'.format(prefix, segment_name, size), |
| worker_id) |
| |
| |
| # Ensure that we are not benchmarking with a google jvm. |
| def check_java_version(): |
| cmd = [jdk.GetJavaExecutable(), '-version'] |
| output = subprocess.check_output(cmd, |
| stderr=subprocess.STDOUT).decode('utf-8') |
| m = re.search('openjdk version "([^"]*)"', output) |
| if m is None: |
| raise Exception("Can't check java version: no version string in output" |
| " of 'java -version': '{}'".format(output)) |
| version = m.groups(0)[0] |
| m = re.search('google', version) |
| if m is not None: |
| raise Exception("Do not use google JVM for benchmarking: " + version) |
| |
| |
| def get_android_jar_dir(api): |
| return os.path.join(REPO_ROOT, ANDROID_JAR_DIR.format(api=api)) |
| |
| |
| def get_android_jar(api): |
| return os.path.join(REPO_ROOT, ANDROID_JAR.format(api=api)) |
| |
| |
| def is_bot(): |
| return 'SWARMING_BOT_ID' in os.environ |
| |
| |
| def uncompressed_size(path): |
| return sum(z.file_size for z in zipfile.ZipFile(path).infolist()) |
| |
| |
| def desugar_configuration_name_and_version(configuration, is_for_maven): |
| name = 'desugar_jdk_libs_configuration' |
| with open(configuration, 'r') as f: |
| configuration_json = json.loads(f.read()) |
| configuration_format_version = \ |
| configuration_json.get('configuration_format_version') |
| if (not configuration_format_version): |
| raise Exception('No "configuration_format_version" found in ' + |
| configuration) |
| if (configuration_format_version != 3 and |
| configuration_format_version != 5 and |
| configuration_format_version != (200 if is_for_maven else 101)): |
| raise Exception( |
| 'Unsupported "configuration_format_version" "%s" found in %s' % |
| (configuration_format_version, configuration)) |
| version = configuration_json.get('version') |
| if not version: |
| if configuration_format_version == (200 if is_for_maven else 101): |
| identifier = configuration_json.get('identifier') |
| if not identifier: |
| raise Exception('No "identifier" found in ' + configuration) |
| identifier_split = identifier.split(':') |
| if (len(identifier_split) != 3): |
| raise Exception('Invalid "identifier" found in ' + |
| configuration) |
| if (identifier_split[0] != 'com.tools.android'): |
| raise Exception('Invalid "identifier" found in ' + |
| configuration) |
| if not identifier_split[1].startswith( |
| 'desugar_jdk_libs_configuration'): |
| raise Exception('Invalid "identifier" found in ' + |
| configuration) |
| name = identifier_split[1] |
| version = identifier_split[2] |
| else: |
| raise Exception('No "version" found in ' + configuration) |
| else: |
| if configuration_format_version == (200 if is_for_maven else 101): |
| raise Exception('No "version" expected in ' + configuration) |
| # Disallow prerelease, as older R8 versions cannot parse it causing hard to |
| # understand errors. |
| check_basic_semver_version(version, |
| 'in ' + configuration, |
| allowPrerelease=False) |
| return (name, version) |
| |
| |
| class SemanticVersion: |
| |
| def __init__(self, major, minor, patch, prerelease): |
| self.major = major |
| self.minor = minor |
| self.patch = patch |
| self.prerelease = prerelease |
| # Build metadata currently not suppported |
| |
| def larger_than(self, other): |
| if self.major > other.major: |
| return True |
| if self.major == other.major and self.minor > other.minor: |
| return True |
| if self.patch: |
| return (self.major == other.major and self.minor == other.minor and |
| self.patch > other.patch) |
| if self.prerelease: |
| if other.prerelease: |
| return self.prerelease > other.prerelease |
| else: |
| return False |
| |
| else: |
| return False |
| |
| |
| # Check that the passed string is formatted as a basic semver version (x.y.z or x.y.z-prerelease |
| # depending on the value of allowPrerelease). |
| # See https://semver.org/. The regexp parts used are not all complient with what is suggested |
| # on https://semver.org/#is-there-a-suggested-regular-expression-regex-to-check-a-semver-string. |
| def check_basic_semver_version(version, |
| error_context='', |
| components=3, |
| allowPrerelease=False): |
| regexp = '^' |
| for x in range(components): |
| regexp += '([0-9]+)' |
| if x < components - 1: |
| regexp += '\\.' |
| if allowPrerelease: |
| # This part is from |
| # https://semver.org/#is-there-a-suggested-regular-expression-regex-to-check-a-semver-string |
| regexp += r'(?:-(?P<prerelease>(?:0|[1-9]\d*|\d*[a-zA-Z-][0-9a-zA-Z-]*)(?:\.(?:0|[1-9]\d*|\d*[a-zA-Z-][0-9a-zA-Z-]*))*))?' |
| regexp += '$' |
| reg = re.compile(regexp) |
| match = reg.match(version) |
| if not match: |
| raise Exception("Invalid version '" + version + "'" + |
| (' ' + error_context) if len(error_context) > 0 else '') |
| if components == 2: |
| return SemanticVersion(int(match.group(1)), int(match.group(2)), None, |
| None) |
| elif components == 3 and not allowPrerelease: |
| return SemanticVersion(int(match.group(1)), int(match.group(2)), |
| int(match.group(3)), None) |
| elif components == 3 and allowPrerelease: |
| return SemanticVersion(int(match.group(1)), int(match.group(2)), |
| int(match.group(3)), match.group('prerelease')) |
| else: |
| raise Exception('Argument "components" must be 2 or 3') |