We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
test_dashboard
________________________________ test_dashboard ________________________________ ConnectionRefusedError: [Errno 61] Connection refused The above exception was the direct cause of the following exception: addr = 'tcp://127.0.0.1:49538', timeout = 5, deserialize = True handshake_overrides = None connection_args = {'extra_conn_args': {}, 'require_encryption': False, 'ssl_context': None} scheme = 'tcp', loc = '127.0.0.1:49538' backend = <distributed.comm.tcp.TCPBackend object at 0x116bb3dc0> connector = <distributed.comm.tcp.TCPConnector object at 0x143253b80> comm = None, time_left = <function connect.<locals>.time_left at 0x14363f9a0> backoff_base = 0.01 asyncdefconnect( addr, timeout=None, deserialize=True, handshake_overrides=None, **connection_args ): """ Connect to the given address (a URI such as ``tcp://127.0.0.1:1234``) and yield a ``Comm`` object. If the connection attempt fails, it is retried until the *timeout* is expired. """ if timeout isNone: timeout = dask.config.get("distributed.comm.timeouts.connect") timeout = parse_timedelta(timeout, default="seconds") scheme, loc = parse_address(addr) backend = registry.get_backend(scheme) connector = backend.get_connector() comm = None start = time() deftime_left(): deadline = start + timeout returnmax(0, deadline - time()) backoff_base = 0.01 attempt = 0 # Prefer multiple small attempts than one long attempt. This should protect # primarily from DNS race conditions # gh3104, gh4176, gh4167 intermediate_cap = timeout / 5 active_exception = None while time_left() > 0: try: > comm = await asyncio.wait_for( connector.connect(loc, deserialize=deserialize, **connection_args), timeout=min(intermediate_cap, time_left()), ) distributed/comm/core.py:291: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ fut = <Task finished name='Task-150' coro=<BaseTCPConnector.connect() done, defined at /Users/runner/work/distributed/distri...'in <distributed.comm.tcp.TCPConnector object at 0x143253b80>: ConnectionRefusedError: [Errno 61] Connection refused')> timeout = 2.8829848766326904 asyncdefwait_for(fut, timeout): """Wait for the single Future or coroutine to complete, with timeout. Coroutine will be wrapped in Task. Returns result of the Future or coroutine. When a timeout occurs, it cancels the task and raises TimeoutError. To avoid the task cancellation, wrap it in shield(). If the wait is cancelled, the task is also cancelled. This function is a coroutine. """ loop = events.get_running_loop() if timeout isNone: returnawait fut if timeout <= 0: fut = ensure_future(fut, loop=loop) if fut.done(): return fut.result() await _cancel_and_wait(fut, loop=loop) try: return fut.result() except exceptions.CancelledError as exc: raise exceptions.TimeoutError() fromexc waiter = loop.create_future() timeout_handle = loop.call_later(timeout, _release_waiter, waiter) cb = functools.partial(_release_waiter, waiter) fut = ensure_future(fut, loop=loop) fut.add_done_callback(cb) try: # wait until the future completes or the timeout try: await waiter except exceptions.CancelledError: if fut.done(): return fut.result() else: fut.remove_done_callback(cb) # We must ensure that the task is not running # after wait_for() returns. # See https://bugs.python.org/issue32751 await _cancel_and_wait(fut, loop=loop) raise if fut.done(): > return fut.result() ../../../miniconda3/envs/dask-distributed/lib/python3.10/asyncio/tasks.py:445: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <distributed.comm.tcp.TCPConnector object at 0x143253b80> address = '127.0.0.1:49538', deserialize = True connection_args = {'extra_conn_args': {}, 'require_encryption': False, 'ssl_context': None} ip = '127.0.0.1', port = 49538, kwargs = {} asyncdefconnect(self, address, deserialize=True, **connection_args): self._check_encryption(address, connection_args) ip, port = parse_host_port(address) kwargs = self._get_connect_args(**connection_args) try: stream = awaitself.client.connect( ip, port, max_buffer_size=MAX_BUFFER_SIZE, **kwargs ) # Under certain circumstances tornado will have a closed connnection with an # error and not raise a StreamClosedError. # # This occurs with tornado 5.x and openssl 1.1+ if stream.closed() and stream.error: raise StreamClosedError(stream.error) except StreamClosedError as e: # The socket connect() call failed > convert_stream_closed_error(self, e) distributed/comm/tcp.py:461: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ obj = <distributed.comm.tcp.TCPConnector object at 0x143253b80> exc = ConnectionRefusedError(61, 'Connection refused') defconvert_stream_closed_error(obj, exc): """ Re-raise StreamClosedError as CommClosedError. """ if exc.real_error isnotNone: # The stream was closed because of an underlying OS error exc = exc.real_error ifisinstance(exc, ssl.SSLError): if"UNKNOWN_CA"in exc.reason: raise FatalCommClosedError(f"in {obj}: {exc.__class__.__name__}: {exc}") > raise CommClosedError(f"in {obj}: {exc.__class__.__name__}: {exc}") fromexc E distributed.comm.core.CommClosedError: in <distributed.comm.tcp.TCPConnector object at 0x143253b80>: ConnectionRefusedError: [Errno 61] Connection refused distributed/comm/tcp.py:142: CommClosedError The above exception was the direct cause of the following exception: loop = <tornado.platform.asyncio.AsyncIOLoop object at 0x142a95ba0> deftest_dashboard(loop): pytest.importorskip("bokeh") port = open_port() with popen( ["dask-scheduler", "--host", f"127.0.0.1:{port}"], ): > with Client(f"127.0.0.1:{port}", loop=loop) as c: distributed/cli/tests/test_dask_scheduler.py:82: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ distributed/client.py:940: in __init__ self.start(timeout=timeout) distributed/client.py:1098: in start sync(self.loop, self._start, **kwargs) distributed/utils.py:405: in sync raise exc.with_traceback(tb) distributed/utils.py:378: in f result = yield future ../../../miniconda3/envs/dask-distributed/lib/python3.10/site-packages/tornado/gen.py:762: in run value = future.result() distributed/client.py:1178: in _start awaitself._ensure_connected(timeout=timeout) distributed/client.py:1241: in _ensure_connected comm = await connect( _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ addr = 'tcp://127.0.0.1:49538', timeout = 5, deserialize = True handshake_overrides = None connection_args = {'extra_conn_args': {}, 'require_encryption': False, 'ssl_context': None} scheme = 'tcp', loc = '127.0.0.1:49538' backend = <distributed.comm.tcp.TCPBackend object at 0x116bb3dc0> connector = <distributed.comm.tcp.TCPConnector object at 0x143253b80> comm = None, time_left = <function connect.<locals>.time_left at 0x14363f9a0> backoff_base = 0.01 asyncdefconnect( addr, timeout=None, deserialize=True, handshake_overrides=None, **connection_args ): """ Connect to the given address (a URI such as ``tcp://127.0.0.1:1234``) and yield a ``Comm`` object. If the connection attempt fails, it is retried until the *timeout* is expired. """ if timeout isNone: timeout = dask.config.get("distributed.comm.timeouts.connect") timeout = parse_timedelta(timeout, default="seconds") scheme, loc = parse_address(addr) backend = registry.get_backend(scheme) connector = backend.get_connector() comm = None start = time() deftime_left(): deadline = start + timeout returnmax(0, deadline - time()) backoff_base = 0.01 attempt = 0 # Prefer multiple small attempts than one long attempt. This should protect # primarily from DNS race conditions # gh3104, gh4176, gh4167 intermediate_cap = timeout / 5 active_exception = None while time_left() > 0: try: comm = await asyncio.wait_for( connector.connect(loc, deserialize=deserialize, **connection_args), timeout=min(intermediate_cap, time_left()), ) break except FatalCommClosedError: raise # Note: CommClosed inherits from OSError except (asyncio.TimeoutError, OSError) as exc: active_exception = exc # As descibed above, the intermediate timeout is used to distributed # initial, bulk connect attempts homogeneously. In particular with # the jitter upon retries we should not be worred about overloading # any more DNS servers intermediate_cap = timeout # FullJitter see https://aws.amazon.com/blogs/architecture/exponential-backoff-and-jitter/ upper_cap = min(time_left(), backoff_base * (2**attempt)) backoff = random.uniform(0, upper_cap) attempt += 1 logger.debug( "Could not connect to %s, waiting for %s before retrying", loc, backoff ) await asyncio.sleep(backoff) else: > raiseOSError( f"Timed out trying to connect to {addr} after {timeout} s" ) fromactive_exception E OSError: Timed out trying to connect to tcp://127.0.0.1:49538 after 5 s distributed/comm/core.py:317: OSError ----------------------------- Captured stderr call ----------------------------- [2022](https://github.com/dask/distributed/runs/7416275344?check_suite_focus=true#step:11:2023)-07-19 19:25:37,688 - distributed.scheduler - INFO - ----------------------------------------------- 2022-07-19 19:25:39,042 - distributed.http.proxy - INFO - To route to workers diagnostics web server please install jupyter-server-proxy: python -m pip install jupyter-server-proxy 2022-07-19 19:25:39,163 - distributed.scheduler - INFO - State start 2022-07-19 19:25:39,167 - distributed.scheduler - INFO - ----------------------------------------------- 2022-07-19 19:25:39,168 - distributed.scheduler - INFO - Clear task state 2022-07-19 19:25:39,169 - distributed.scheduler - INFO - Scheduler at: tcp://127.0.0.1:49538 2022-07-19 19:25:39,169 - distributed.scheduler - INFO - dashboard at: 127.0.0.1:8787 2022-07-19 19:25:41,316 - distributed._signals - INFO - Received signal SIGINT (2) 2022-07-19 19:25:41,317 - distributed.scheduler - INFO - Scheduler closing... 2022-07-19 19:25:41,318 - distributed.scheduler - INFO - Scheduler closing all comms 2022-07-19 19:25:41,319 - distributed.scheduler - INFO - Stopped scheduler at 'tcp://127.0.0.1:49538' 2022-07-19 19:25:41,320 - distributed.scheduler - INFO - End scheduler
https://github.com/dask/distributed/runs/7416275344?check_suite_focus=true#step:11:1897
The text was updated successfully, but these errors were encountered:
This one looks newly flaky as of ~Jul 8:
Sorry, something went wrong.
OSError: Timed out trying to connect to tcp://127.0.0.1:8786 after 5 s
gen_cluster
Successfully merging a pull request may close this issue.
https://github.com/dask/distributed/runs/7416275344?check_suite_focus=true#step:11:1897
The text was updated successfully, but these errors were encountered: