[48/49] cd /home/petkantchin/ws/nodlabs/compiler_gym/build/ninja/RelWithDebInfo/tests && bash -c "cd \"/home/petkantchin/ws/nodlabs/compiler_gym/repo/tests/..\" && make install-test" mkdir -p "/tmp/compiler_gym_petkantchin/install_tests" rm -f "/tmp/compiler_gym_petkantchin/install_tests/tests" "/tmp/compiler_gym_petkantchin/install_tests/tox.ini" ln -s "/home/petkantchin/ws/nodlabs/compiler_gym/repo//tests" "/tmp/compiler_gym_petkantchin/install_tests" ln -s "/home/petkantchin/ws/nodlabs/compiler_gym/repo//tox.ini" "/tmp/compiler_gym_petkantchin/install_tests" cd "/tmp/compiler_gym_petkantchin/install_tests" && pytest tests --no-success-flaky-report --benchmark-disable -n auto -k "not fuzz" --durations=5 ============================= test session starts ============================== platform linux -- Python 3.8.10, pytest-6.2.5, py-1.11.0, pluggy-0.13.1 benchmark: 3.4.1 (defaults: timer=time.perf_counter disable_gc=False min_rounds=5 min_time=0.000005 max_time=1.0 calibration_precision=10 warmup=False warmup_iterations=100000) rootdir: /tmp/compiler_gym_petkantchin/install_tests, configfile: tox.ini plugins: flaky-3.7.0, xdist-2.4.0, shard-0.1.2, sugar-0.9.4, forked-1.3.0, hydra-core-1.1.0, cov-2.12.1, benchmark-3.4.1, timeout-1.4.2, mock-3.6.1, stress-1.0.1 gw0 I / gw1 I / gw2 I / gw3 I / gw4 I / gw5 I / gw6 I / gw7 I gw0 [2668] / gw1 [2668] / gw2 [2668] / gw3 [2668] / gw4 [2668] / gw5 [2668] / gw6 [2668] / gw7 [2668] ................................................F......s................ [ 2%] .s.....X.ss.......................................s..ss.............s..s [ 5%] s...................................s...s...........s...ss.............. [ 8%] .....s..ss....XF.......................sX.ss............................ [ 10%] .....................................s...........F.s.................... [ 13%] .....................s.................................................. [ 16%] ........................................................................ [ 18%] ........................................................................ [ 21%] ........................................................................ [ 24%] ............................................................s......s.... [ 26%] ........................................................................ [ 29%] ........................................................................ [ 32%] ........................................................................ [ 35%] ........................................................................ [ 37%] ........................................................................ [ 40%] .........................................s.............................. [ 43%] ........................................................................ [ 45%] ........................................................................ [ 48%] ........................................................................ [ 51%] ........................................................................ [ 53%] ......................................................................... [ 56%] ........................................................................ [ 59%] ........................................................................ [ 62%] ........................................................................ [ 64%] ........................................................................ [ 67%] ......................................................................... [ 70%] ........................................................................ [ 72%] ........................................................................ [ 75%] ........................................................................ [ 78%] ........................................................................ [ 81%] ........................................................................ [ 83%] ........................................................................ [ 86%] ........................................................................ [ 89%] ..................................................................F..... [ 91%] ........................................................................ [ 94%] ........................................................................ [ 97%] ........................................................................ [ 99%] .. [100%] =================================== FAILURES =================================== ___________ test_print_service_capabilities_smoke_test[loop_tool-v0] ___________ [gw7] linux -- Python 3.8.10 /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/bin/python cls = func = . at 0x7f46e21779d0> when = 'call' reraise = (, ) @classmethod def from_call( cls, func: "Callable[[], TResult]", when: "Literal['collect', 'setup', 'call', 'teardown']", reraise: Optional[ Union[Type[BaseException], Tuple[Type[BaseException], ...]] ] = None, ) -> "CallInfo[TResult]": excinfo = None start = timing.time() precise_start = timing.perf_counter() try: > result: Optional[TResult] = func() /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/runner.py:311: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/runner.py:255: in lambda: ihook(item=item, **kwds), when=when, reraise=reraise /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pluggy/hooks.py:286: in __call__ return self._hookexec(self, self.get_hookimpls(), kwargs) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pluggy/manager.py:93: in _hookexec return self._inner_hookexec(hook, methods, kwargs) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pluggy/manager.py:84: in self._inner_hookexec = lambda hook, methods, kwargs: hook.multicall( /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/unraisableexception.py:88: in pytest_runtest_call yield from unraisable_exception_runtest_hook() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ def unraisable_exception_runtest_hook() -> Generator[None, None, None]: with catch_unraisable_exception() as cm: yield if cm.unraisable: if cm.unraisable.err_msg is not None: err_msg = cm.unraisable.err_msg else: err_msg = "Exception ignored in" msg = f"{err_msg}: {cm.unraisable.object!r}\n\n" msg += "".join( traceback.format_exception( cm.unraisable.exc_type, cm.unraisable.exc_value, cm.unraisable.exc_traceback, ) ) > warnings.warn(pytest.PytestUnraisableExceptionWarning(msg)) E pytest.PytestUnraisableExceptionWarning: Exception ignored in: E E Traceback (most recent call last): E File "/home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/compiler_gym/service/connection.py", line 616, in __init__ E self.endpoint = endpoint E ResourceWarning: unclosed /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/unraisableexception.py:78: PytestUnraisableExceptionWarning ----------------------------- Captured stdout call ----------------------------- Datasets -------- +-------------------------------+--------------------------+-------------------+----------------------+ | Dataset | Num. Benchmarks [#f1]_ | Description | Validatable [#f2]_ | +===============================+==========================+===================+======================+ | benchmark://loop_tool-cpu-v0 | ∞ | loop_tool dataset | No | +-------------------------------+--------------------------+-------------------+----------------------+ | benchmark://loop_tool-cuda-v0 | ∞ | loop_tool dataset | No | +-------------------------------+--------------------------+-------------------+----------------------+ | Total | ∞ | | | +-------------------------------+--------------------------+-------------------+----------------------+ .. [#f1] Values obtained on linux. Datasets are platform-specific. .. [#f2] A **validatable** dataset is one where the behavior of the benchmarks can be checked by compiling the programs to binaries and executing them. If the benchmarks crash, or are found to have different behavior, then validation fails. This type of validation is used to check that the compiler has not broken the semantics of the program. See :mod:`compiler_gym.bin.validate`. Observation Spaces ------------------ +---------------------+----------------------------------------+ | Observation space | Shape | +=====================+========================================+ | action_state | `Box([0], [68719476736], (1,), int64)` | +---------------------+----------------------------------------+ | flops | `float<-inf,inf>` | +---------------------+----------------------------------------+ | loop_tree | `str_list<>[0,inf])` | +---------------------+----------------------------------------+ Reward Spaces ------------- +----------------+-------------+---------------------+------------------+-----------------------+ | Reward space | Range | Success threshold | Deterministic? | Platform dependent? | +================+=============+=====================+==================+=======================+ | flops | (-inf, inf) | | No | Yes | +----------------+-------------+---------------------+------------------+-----------------------+ simple Action Space ------------------- +---------------+ | Action | +===============+ | down | +---------------+ | toggle_mode | +---------------+ | toggle_thread | +---------------+ | up | +---------------+ split Action Space ------------------ +---------------+ | Action | +===============+ | down | +---------------+ | split | +---------------+ | toggle_mode | +---------------+ | toggle_thread | +---------------+ | up | +---------------+ ___________________________ test_setuptools_version ____________________________ [gw5] linux -- Python 3.8.10 /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/bin/python self = gym 0.20.0 (/home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages) @property def _dep_map(self): try: > return self.__dep_map /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:3031: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = gym 0.20.0 (/home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages) attr = '_DistInfoDistribution__dep_map' def __getattr__(self, attr): """Delegate all unrecognized public attributes to .metadata provider""" if attr.startswith('_'): > raise AttributeError(attr) E AttributeError: _DistInfoDistribution__dep_map /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:2828: AttributeError During handling of the above exception, another exception occurred: self = <[AttributeError("'SpecifierSet' object has no attribute '_prereleases'") raised in repr()] SpecifierSet object at 0x7f5b1c22a070> specifiers = '>=3.', prereleases = None def __init__( self, specifiers: str = "", prereleases: Optional[bool] = None ) -> None: # Split on , to break each individual specifier into it's own item, and # strip each item to remove leading/trailing whitespace. split_specifiers = [s.strip() for s in specifiers.split(",") if s.strip()] # Parsed each individual specifier, attempting first to make it a # Specifier and falling back to a LegacySpecifier. parsed: Set[_IndividualSpecifier] = set() for specifier in split_specifiers: try: > parsed.add(Specifier(specifier)) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/_vendor/packaging/specifiers.py:652: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <[AttributeError("'Specifier' object has no attribute '_prereleases'") raised in repr()] Specifier object at 0x7f5b1c245c10> spec = '>=3.', prereleases = None def __init__(self, spec: str = "", prereleases: Optional[bool] = None) -> None: match = self._regex.search(spec) if not match: > raise InvalidSpecifier(f"Invalid specifier: '{spec}'") E pkg_resources.extern.packaging.specifiers.InvalidSpecifier: Invalid specifier: '>=3.' /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/_vendor/packaging/specifiers.py:105: InvalidSpecifier During handling of the above exception, another exception occurred: @install_test_only def test_setuptools_version(): > version = pkg_resources.require("compiler_gym")[0].version /home/petkantchin/ws/nodlabs/compiler_gym/build/ninja/RelWithDebInfo/tests/py_pkg/tests/version_test.py:24: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:891: in require needed = self.resolve(parse_requirements(requirements)) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:785: in resolve new_requirements = dist.requires(req.extras)[::-1] /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:2749: in requires dm = self._dep_map /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:3033: in _dep_map self.__dep_map = self._compute_dependencies() /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:3043: in _compute_dependencies reqs.extend(parse_requirements(req)) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:3099: in parse_requirements yield Requirement(line) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/__init__.py:3109: in __init__ super(Requirement, self).__init__(requirement_string) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/_vendor/packaging/requirements.py:122: in __init__ self.specifier: SpecifierSet = SpecifierSet(req.specifier) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/_vendor/packaging/specifiers.py:654: in __init__ parsed.add(LegacySpecifier(specifier)) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = =3.')>, spec = '>=3.', prereleases = None def __init__(self, spec: str = "", prereleases: Optional[bool] = None) -> None: super().__init__(spec, prereleases) > warnings.warn( "Creating a LegacyVersion has been deprecated and will be " "removed in the next major release", DeprecationWarning, ) E DeprecationWarning: Creating a LegacyVersion has been deprecated and will be removed in the next major release /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pkg_resources/_vendor/packaging/specifiers.py:273: DeprecationWarning __________________ test_reset_invalid_benchmark[/usr/bin/gcc] __________________ [gw5] linux -- Python 3.8.10 /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/bin/python cls = func = . at 0x7f5b19c893a0> when = 'call' reraise = (, ) @classmethod def from_call( cls, func: "Callable[[], TResult]", when: "Literal['collect', 'setup', 'call', 'teardown']", reraise: Optional[ Union[Type[BaseException], Tuple[Type[BaseException], ...]] ] = None, ) -> "CallInfo[TResult]": excinfo = None start = timing.time() precise_start = timing.perf_counter() try: > result: Optional[TResult] = func() /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/runner.py:311: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/runner.py:255: in lambda: ihook(item=item, **kwds), when=when, reraise=reraise /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pluggy/hooks.py:286: in __call__ return self._hookexec(self, self.get_hookimpls(), kwargs) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pluggy/manager.py:93: in _hookexec return self._inner_hookexec(hook, methods, kwargs) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/pluggy/manager.py:84: in self._inner_hookexec = lambda hook, methods, kwargs: hook.multicall( /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/unraisableexception.py:88: in pytest_runtest_call yield from unraisable_exception_runtest_hook() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ def unraisable_exception_runtest_hook() -> Generator[None, None, None]: with catch_unraisable_exception() as cm: yield if cm.unraisable: if cm.unraisable.err_msg is not None: err_msg = cm.unraisable.err_msg else: err_msg = "Exception ignored in" msg = f"{err_msg}: {cm.unraisable.object!r}\n\n" msg += "".join( traceback.format_exception( cm.unraisable.exc_type, cm.unraisable.exc_value, cm.unraisable.exc_traceback, ) ) > warnings.warn(pytest.PytestUnraisableExceptionWarning(msg)) E pytest.PytestUnraisableExceptionWarning: Exception ignored in: E E Traceback (most recent call last): E File "/home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/compiler_gym/envs/gcc/gcc.py", line 727, in _get_spec E spec = pickle.load(f) E ResourceWarning: unclosed /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/unraisableexception.py:78: PytestUnraisableExceptionWarning ___________________________ test_no_args_call[slurm] ___________________________ [gw2] linux -- Python 3.8.10 /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/bin/python tmpdir = local('/tmp/pytest-of-petkantchin/pytest-13/popen-gw2/test_no_args_call_slurm_0') executor_type = 'slurm' @pytest.mark.xfail( sys.platform == "darwin", reason="'ResourceWarning: unclosed ' when type == local", ) def test_no_args_call(tmpdir, executor_type: str): with Executor(type=executor_type, cpus=1).get_executor(logs_dir=tmpdir) as executor: > job = executor.submit(_hello_fn) /home/petkantchin/ws/nodlabs/compiler_gym/build/ninja/RelWithDebInfo/tests/py_pkg/tests/util/executor_test.py:48: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/compiler_gym/util/executor.py:188: in submit logger.info("Submitting job %s to %s ...", job.job_id, self) /usr/lib/python3.8/logging/__init__.py:1446: in info self._log(INFO, msg, args, **kwargs) /usr/lib/python3.8/logging/__init__.py:1589: in _log self.handle(record) /usr/lib/python3.8/logging/__init__.py:1599: in handle self.callHandlers(record) /usr/lib/python3.8/logging/__init__.py:1661: in callHandlers hdlr.handle(record) /usr/lib/python3.8/logging/__init__.py:954: in handle self.emit(record) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/logging.py:331: in emit super().emit(record) /usr/lib/python3.8/logging/__init__.py:1093: in emit self.handleError(record) /usr/lib/python3.8/logging/__init__.py:1085: in emit msg = self.format(record) /usr/lib/python3.8/logging/__init__.py:929: in format return fmt.format(record) /home/petkantchin/ws/nodlabs/compiler_gym/compiler_gym_venv/lib/python3.8/site-packages/_pytest/logging.py:92: in format return super().format(record) /usr/lib/python3.8/logging/__init__.py:668: in format record.message = record.getMessage() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def getMessage(self): """ Return the message for this LogRecord. Return the message for this LogRecord after merging any user-supplied arguments with the message. """ msg = str(self.msg) if self.args: > msg = msg % self.args E TypeError: __str__ returned non-string (type NoneType) /usr/lib/python3.8/logging/__init__.py:373: TypeError ============================= slowest 5 durations ============================== 216.61s call tests/llvm/validation_regression_test.py::test_validate_known_bad_trajectory[benchmark://cbench-v1/susan,opt -mem2reg -reg2mem -gvn -reg2mem -gvn -newgvn input.bc -o output.bc] 183.90s call tests/random_search_test.py::test_random_search_smoke_test 180.25s call tests/llvm/custom_benchmarks_test.py::test_make_benchmark_single_clang_job 160.90s call tests/loop_tool/actions_test.py::test_basic[cpu] 152.51s call tests/bin/validate_bin_test.py::test_validate_cbench_null_options[benchmarks5] =========================== short test summary info ============================ FAILED tests/bin/service_bin_test.py::test_print_service_capabilities_smoke_test[loop_tool-v0] FAILED tests/version_test.py::test_setuptools_version - DeprecationWarning: C... FAILED tests/gcc/gcc_env_test.py::test_reset_invalid_benchmark[/usr/bin/gcc] FAILED tests/util/executor_test.py::test_no_args_call[slurm] - TypeError: __s... ====== 4 failed, 2634 passed, 27 skipped, 3 xpassed in 1348.91s (0:22:28) ====== make: *** [Makefile:334: install-test] Error 1 FAILED: tests/CMakeFiles/tests__python_package_tests cd /home/petkantchin/ws/nodlabs/compiler_gym/build/ninja/RelWithDebInfo/tests && bash -c "cd \"/home/petkantchin/ws/nodlabs/compiler_gym/repo/tests/..\" && make install-test" ninja: build stopped: subcommand failed.