Source code for distributed.worker

from __future__ import print_function, division, absolute_import

import bisect
from collections import defaultdict, deque
from datetime import timedelta
import heapq
import logging
import os
from pickle import PicklingError
import random
import threading
import sys
import warnings
import weakref

import dask
from dask.core import istask
from dask.compatibility import apply
try:
    from cytoolz import pluck, partial, merge
except ImportError:
    from toolz import pluck, partial, merge
from tornado.gen import Return
from tornado import gen
from tornado.ioloop import IOLoop
from tornado.locks import Event

from . import profile, comm
from .batched import BatchedSend
from .comm import get_address_host, get_local_address_for, connect
from .comm.utils import offload
from .compatibility import unicode, get_thread_identity, finalize
from .core import (error_message, CommClosedError, send_recv,
                   pingpong, coerce_to_address)
from .diskutils import WorkSpace
from .metrics import time
from .node import ServerNode
from .preloading import preload_modules
from .proctitle import setproctitle
from .protocol import (pickle, to_serialize, deserialize_bytes,
                       serialize_bytelist)
from .pubsub import PubSubWorkerExtension
from .security import Security
from .sizeof import safe_sizeof as sizeof
from .threadpoolexecutor import ThreadPoolExecutor, secede as tpe_secede
from .utils import (funcname, get_ip, has_arg, _maybe_complex, log_errors,
                    ignoring, mp_context, import_file,
                    silence_logging, thread_state, json_load_robust, key_split,
                    format_bytes, DequeHandler, PeriodicCallback,
                    parse_bytes, parse_timedelta, iscoroutinefunction)
from .utils_comm import pack_data, gather_from_workers
from .utils_perf import ThrottledGC, enable_gc_diagnosis, disable_gc_diagnosis

_ncores = mp_context.cpu_count()

logger = logging.getLogger(__name__)

LOG_PDB = dask.config.get('distributed.admin.pdb-on-err')

no_value = '--no-value-sentinel--'

try:
    import psutil
    TOTAL_MEMORY = psutil.virtual_memory().total
except ImportError:
    logger.warning("Please install psutil to estimate worker memory use")
    TOTAL_MEMORY = 8e9
    psutil = None


IN_PLAY = ('waiting', 'ready', 'executing', 'long-running')
PENDING = ('waiting', 'ready', 'constrained')
PROCESSING = ('waiting', 'ready', 'constrained', 'executing', 'long-running')
READY = ('ready', 'constrained')


DEFAULT_EXTENSIONS = [
    PubSubWorkerExtension,
]

_global_workers = []


[docs]class Worker(ServerNode): """ Worker node in a Dask distributed cluster Workers perform two functions: 1. **Serve data** from a local dictionary 2. **Perform computation** on that data and on data from peers Workers keep the scheduler informed of their data and use that scheduler to gather data from other workers when necessary to perform a computation. You can start a worker with the ``dask-worker`` command line application:: $ dask-worker scheduler-ip:port Use the ``--help`` flag to see more options $ dask-worker --help The rest of this docstring is about the internal state the the worker uses to manage and track internal computations. **State** **Informational State** These attributes don't change significantly during execution. * **ncores:** ``int``: Number of cores used by this worker process * **executor:** ``concurrent.futures.ThreadPoolExecutor``: Executor used to perform computation * **local_dir:** ``path``: Path on local machine to store temporary files * **scheduler:** ``rpc``: Location of scheduler. See ``.ip/.port`` attributes. * **name:** ``string``: Alias * **services:** ``{str: Server}``: Auxiliary web servers running on this worker * **service_ports:** ``{str: port}``: * **total_out_connections**: ``int`` The maximum number of concurrent outgoing requests for data * **total_in_connections**: ``int`` The maximum number of concurrent incoming requests for data * **total_comm_nbytes**: ``int`` * **batched_stream**: ``BatchedSend`` A batched stream along which we communicate to the scheduler * **log**: ``[(message)]`` A structured and queryable log. See ``Worker.story`` **Volatile State** This attributes track the progress of tasks that this worker is trying to complete. In the descriptions below a ``key`` is the name of a task that we want to compute and ``dep`` is the name of a piece of dependent data that we want to collect from others. * **data:** ``{key: object}``: Dictionary mapping keys to actual values * **task_state**: ``{key: string}``: The state of all tasks that the scheduler has asked us to compute. Valid states include waiting, constrained, executing, memory, erred * **tasks**: ``{key: dict}`` The function, args, kwargs of a task. We run this when appropriate * **dependencies**: ``{key: {deps}}`` The data needed by this key to run * **dependents**: ``{dep: {keys}}`` The keys that use this dependency * **data_needed**: deque(keys) The keys whose data we still lack, arranged in a deque * **waiting_for_data**: ``{kep: {deps}}`` A dynamic verion of dependencies. All dependencies that we still don't have for a particular key. * **ready**: [keys] Keys that are ready to run. Stored in a LIFO stack * **constrained**: [keys] Keys for which we have the data to run, but are waiting on abstract resources like GPUs. Stored in a FIFO deque * **executing**: {keys} Keys that are currently executing * **executed_count**: int A number of tasks that this worker has run in its lifetime * **long_running**: {keys} A set of keys of tasks that are running and have started their own long-running clients. * **dep_state**: ``{dep: string}``: The state of all dependencies required by our tasks Valid states include waiting, flight, and memory * **who_has**: ``{dep: {worker}}`` Workers that we believe have this data * **has_what**: ``{worker: {deps}}`` The data that we care about that we think a worker has * **pending_data_per_worker**: ``{worker: [dep]}`` The data on each worker that we still want, prioritized as a deque * **in_flight_tasks**: ``{task: worker}`` All dependencies that are coming to us in current peer-to-peer connections and the workers from which they are coming. * **in_flight_workers**: ``{worker: {task}}`` The workers from which we are currently gathering data and the dependencies we expect from those connections * **comm_bytes**: ``int`` The total number of bytes in flight * **suspicious_deps**: ``{dep: int}`` The number of times a dependency has not been where we expected it * **nbytes**: ``{key: int}`` The size of a particular piece of data * **types**: ``{key: type}`` The type of a particular piece of data * **threads**: ``{key: int}`` The ID of the thread on which the task ran * **active_threads**: ``{int: key}`` The keys currently running on active threads * **exceptions**: ``{key: exception}`` The exception caused by running a task if it erred * **tracebacks**: ``{key: traceback}`` The exception caused by running a task if it erred * **startstops**: ``{key: [(str, float, float)]}`` Log of transfer, load, and compute times for a task * **priorities**: ``{key: tuple}`` The priority of a key given by the scheduler. Determines run order. * **durations**: ``{key: float}`` Expected duration of a task * **resource_restrictions**: ``{key: {str: number}}`` Abstract resources required to run a task Parameters ---------- scheduler_ip: str scheduler_port: int ip: str, optional ncores: int, optional loop: tornado.ioloop.IOLoop local_dir: str, optional Directory where we place local resources name: str, optional memory_limit: int, float, string Number of bytes of memory that this worker should use. Set to zero for no limit. Set to 'auto' for 60% of memory use. Use strings or numbers like 5GB or 5e9 memory_target_fraction: float Fraction of memory to try to stay beneath memory_spill_fraction: float Fraction of memory at which we start spilling to disk memory_pause_fraction: float Fraction of memory at which we stop running new tasks executor: concurrent.futures.Executor resources: dict Resources that thiw worker has like ``{'GPU': 2}`` Examples -------- Use the command line to start a worker:: $ dask-scheduler Start scheduler at 127.0.0.1:8786 $ dask-worker 127.0.0.1:8786 Start worker at: 127.0.0.1:1234 Registered with scheduler at: 127.0.0.1:8786 See Also -------- distributed.scheduler.Scheduler distributed.nanny.Nanny """ def __init__(self, scheduler_ip=None, scheduler_port=None, scheduler_file=None, ncores=None, loop=None, local_dir='dask-worker-space', services=None, service_ports=None, name=None, reconnect=True, memory_limit='auto', executor=None, resources=None, silence_logs=None, death_timeout=None, preload=None, preload_argv=None, security=None, contact_address=None, memory_monitor_interval='200ms', extensions=None, metrics=None, **kwargs): self.tasks = dict() self.task_state = dict() self.dep_state = dict() self.dependencies = dict() self.dependents = dict() self.waiting_for_data = dict() self.who_has = dict() self.has_what = defaultdict(set) self.pending_data_per_worker = defaultdict(deque) self._lock = threading.Lock() self.data_needed = deque() # TODO: replace with heap? self.in_flight_tasks = dict() self.in_flight_workers = dict() self.total_out_connections = dask.config.get('distributed.worker.connections.outgoing') self.total_in_connections = dask.config.get('distributed.worker.connections.incoming') self.total_comm_nbytes = 10e6 self.comm_nbytes = 0 self.suspicious_deps = defaultdict(lambda: 0) self._missing_dep_flight = set() self.nbytes = dict() self.types = dict() self.threads = dict() self.exceptions = dict() self.tracebacks = dict() self.active_threads_lock = threading.Lock() self.active_threads = dict() self.profile_keys = defaultdict(profile.create) self.profile_keys_history = deque(maxlen=3600) self.profile_recent = profile.create() self.profile_history = deque(maxlen=3600) self.priorities = dict() self.generation = 0 self.durations = dict() self.startstops = defaultdict(list) self.resource_restrictions = dict() self.ready = list() self.constrained = deque() self.executing = set() self.executed_count = 0 self.long_running = set() self.batched_stream = None self.recent_messages_log = deque(maxlen=dask.config.get('distributed.comm.recent-messages-log-length')) self.target_message_size = 50e6 # 50 MB self.log = deque(maxlen=100000) self.validate = kwargs.pop('validate', False) self._transitions = { ('waiting', 'ready'): self.transition_waiting_ready, ('waiting', 'memory'): self.transition_waiting_done, ('waiting', 'error'): self.transition_waiting_done, ('ready', 'executing'): self.transition_ready_executing, ('ready', 'memory'): self.transition_ready_memory, ('constrained', 'executing'): self.transition_constrained_executing, ('executing', 'memory'): self.transition_executing_done, ('executing', 'error'): self.transition_executing_done, ('executing', 'rescheduled'): self.transition_executing_done, ('executing', 'long-running'): self.transition_executing_long_running, ('long-running', 'error'): self.transition_executing_done, ('long-running', 'memory'): self.transition_executing_done, ('long-running', 'rescheduled'): self.transition_executing_done, } self._dep_transitions = { ('waiting', 'flight'): self.transition_dep_waiting_flight, ('waiting', 'memory'): self.transition_dep_waiting_memory, ('flight', 'waiting'): self.transition_dep_flight_waiting, ('flight', 'memory'): self.transition_dep_flight_memory, } self.incoming_transfer_log = deque(maxlen=(100000)) self.incoming_count = 0 self.outgoing_transfer_log = deque(maxlen=(100000)) self.outgoing_count = 0 self.outgoing_current_count = 0 self.repetitively_busy = 0 self._client = None profile_cycle_interval = kwargs.pop('profile_cycle_interval', dask.config.get('distributed.worker.profile.cycle')) profile_cycle_interval = parse_timedelta(profile_cycle_interval, default='ms') self._setup_logging() if scheduler_file: cfg = json_load_robust(scheduler_file) scheduler_addr = cfg['address'] elif scheduler_ip is None and dask.config.get('scheduler-address', None): scheduler_addr = dask.config.get('scheduler-address') elif scheduler_port is None: scheduler_addr = coerce_to_address(scheduler_ip) else: scheduler_addr = coerce_to_address((scheduler_ip, scheduler_port)) self._port = 0 self.ncores = ncores or _ncores self.total_resources = resources or {} self.available_resources = (resources or {}).copy() self.death_timeout = death_timeout self.preload = preload if self.preload is None: self.preload = dask.config.get('distributed.worker.preload') self.preload_argv = preload_argv if self.preload_argv is None: self.preload_argv = dask.config.get('distributed.worker.preload-argv') self.contact_address = contact_address self.memory_monitor_interval = parse_timedelta(memory_monitor_interval, default='ms') self.extensions = dict() if silence_logs: silence_logging(level=silence_logs) self._workspace = WorkSpace(os.path.abspath(local_dir)) self._workdir = self._workspace.new_work_dir(prefix='worker-') self.local_dir = self._workdir.dir_path self.security = security or Security() assert isinstance(self.security, Security) self.connection_args = self.security.get_connection_args('worker') self.listen_args = self.security.get_listen_args('worker') self.memory_limit = parse_memory_limit(memory_limit, self.ncores) self.paused = False if 'memory_target_fraction' in kwargs: self.memory_target_fraction = kwargs.pop('memory_target_fraction') else: self.memory_target_fraction = dask.config.get('distributed.worker.memory.target') if 'memory_spill_fraction' in kwargs: self.memory_spill_fraction = kwargs.pop('memory_spill_fraction') else: self.memory_spill_fraction = dask.config.get('distributed.worker.memory.spill') if 'memory_pause_fraction' in kwargs: self.memory_pause_fraction = kwargs.pop('memory_pause_fraction') else: self.memory_pause_fraction = dask.config.get('distributed.worker.memory.pause') if (self.memory_limit and (self.memory_target_fraction or self.memory_spill_fraction)): try: from zict import Buffer, File, Func except ImportError: raise ImportError("Please `pip install zict` for spill-to-disk workers") path = os.path.join(self.local_dir, 'storage') storage = Func(partial(serialize_bytelist, on_error='raise'), deserialize_bytes, File(path)) target = int(float(self.memory_limit) * self.memory_target_fraction) self.data = Buffer({}, storage, target, weight) else: self.data = dict() self.actors = {} self.loop = loop or IOLoop.current() self.status = None self._closed = Event() self.reconnect = reconnect self.executor = executor or ThreadPoolExecutor(self.ncores, thread_name_prefix="Dask-Worker-Threads'") self.actor_executor = ThreadPoolExecutor(1, thread_name_prefix="Dask-Actor-Threads") self.name = name self.scheduler_delay = 0 self.stream_comms = dict() self.heartbeat_active = False self._ipython_kernel = None if self.local_dir not in sys.path: sys.path.insert(0, self.local_dir) self.services = {} self.service_ports = service_ports or {} self.service_specs = services or {} self.metrics = dict(metrics) if metrics else {} handlers = { 'gather': self.gather, 'run': self.run, 'run_coroutine': self.run_coroutine, 'get_data': self.get_data, 'update_data': self.update_data, 'delete_data': self.delete_data, 'terminate': self.terminate, 'ping': pingpong, 'upload_file': self.upload_file, 'start_ipython': self.start_ipython, 'call_stack': self.get_call_stack, 'profile': self.get_profile, 'profile_metadata': self.get_profile_metadata, 'get_logs': self.get_logs, 'keys': self.keys, 'versions': self.versions, 'actor_execute': self.actor_execute, 'actor_attribute': self.actor_attribute, } stream_handlers = { 'close': self._close, 'compute-task': self.add_task, 'release-task': partial(self.release_key, report=False), 'delete-data': self.delete_data, 'steal-request': self.steal_request, } super(Worker, self).__init__( handlers=handlers, stream_handlers=stream_handlers, io_loop=self.loop, connection_args=self.connection_args, **kwargs) self.scheduler = self.rpc(scheduler_addr) self.execution_state = {'scheduler': self.scheduler.address, 'ioloop': self.loop, 'worker': self} pc = PeriodicCallback(self.heartbeat, 1000, io_loop=self.io_loop) self.periodic_callbacks['heartbeat'] = pc self._address = contact_address if self.memory_limit: self._memory_monitoring = False pc = PeriodicCallback(self.memory_monitor, self.memory_monitor_interval * 1000, io_loop=self.io_loop) self.periodic_callbacks['memory'] = pc if extensions is None: extensions = DEFAULT_EXTENSIONS for ext in extensions: ext(self) self._throttled_gc = ThrottledGC(logger=logger) setproctitle("dask-worker [not started]") pc = PeriodicCallback( self.trigger_profile, parse_timedelta(dask.config.get('distributed.worker.profile.interval'), default='ms') * 1000, io_loop=self.io_loop ) self.periodic_callbacks['profile'] = pc pc = PeriodicCallback(self.cycle_profile, profile_cycle_interval * 1000, io_loop=self.io_loop) self.periodic_callbacks['profile-cycle'] = pc _global_workers.append(weakref.ref(self)) ################## # Administrative # ################## def __repr__(self): return "<%s: %s, %s, stored: %d, running: %d/%d, ready: %d, comm: %d, waiting: %d>" % ( self.__class__.__name__, self.address, self.status, len(self.data), len(self.executing), self.ncores, len(self.ready), len(self.in_flight_tasks), len(self.waiting_for_data)) def _setup_logging(self): self._deque_handler = DequeHandler(n=dask.config.get('distributed.admin.log-length')) self._deque_handler.setFormatter(logging.Formatter(dask.config.get('distributed.admin.log-format'))) logger.addHandler(self._deque_handler) finalize(self, logger.removeHandler, self._deque_handler) @property def worker_address(self): """ For API compatibility with Nanny """ return self.address def get_metrics(self): core = dict(executing=len(self.executing), in_memory=len(self.data), ready=len(self.ready), in_flight=len(self.in_flight_tasks)) custom = {k: metric(self) for k, metric in self.metrics.items()} return merge(custom, self.monitor.recent(), core) def identity(self, comm=None): return {'type': type(self).__name__, 'id': self.id, 'scheduler': self.scheduler.address, 'ncores': self.ncores, 'memory_limit': self.memory_limit} ##################### # External Services # ##################### @gen.coroutine def _register_with_scheduler(self): self.periodic_callbacks['heartbeat'].stop() start = time() if self.contact_address is None: self.contact_address = self.address logger.info('-' * 49) while True: if self.death_timeout and time() > start + self.death_timeout: yield self._close(timeout=1) return if self.status in ('closed', 'closing'): raise gen.Return try: _start = time() comm = yield connect(self.scheduler.address, connection_args=self.connection_args) yield comm.write(dict(op='register-worker', reply=False, address=self.contact_address, keys=list(self.data), ncores=self.ncores, name=self.name, nbytes=self.nbytes, now=time(), resources=self.total_resources, memory_limit=self.memory_limit, local_directory=self.local_dir, services=self.service_ports, pid=os.getpid(), metrics=self.get_metrics()), serializers=['msgpack']) future = comm.read(deserializers=['msgpack']) if self.death_timeout: diff = self.death_timeout - (time() - start) if diff < 0: continue future = gen.with_timeout(timedelta(seconds=diff), future) response = yield future _end = time() middle = (_start + _end) / 2 self.scheduler_delay = response['time'] - middle self.status = 'running' break except EnvironmentError: logger.info('Waiting to connect to: %26s', self.scheduler.address) yield gen.sleep(0.1) except gen.TimeoutError: logger.info("Timed out when connecting to scheduler") if response['status'] != 'OK': raise ValueError("Unexpected response from register: %r" % (response,)) else: # Retrieve eventual init functions and run them for function_bytes in response['worker-setups']: setup_function = pickle.loads(function_bytes) if has_arg(setup_function, 'dask_worker'): result = setup_function(dask_worker=self) else: result = setup_function() logger.info('Init function %s ran: output=%s' % (setup_function, result)) logger.info(' Registered to: %26s', self.scheduler.address) logger.info('-' * 49) self.batched_stream = BatchedSend(interval='2ms', loop=self.loop) self.batched_stream.start(comm) self.periodic_callbacks['heartbeat'].start() self.loop.add_callback(self.handle_scheduler, comm) @gen.coroutine def heartbeat(self): if not self.heartbeat_active: self.heartbeat_active = True logger.debug("Heartbeat: %s" % self.address) try: start = time() response = yield self.scheduler.heartbeat_worker( address=self.contact_address, now=time(), metrics=self.get_metrics() ) end = time() middle = (start + end) / 2 if response['status'] == 'missing': yield self._register_with_scheduler() return self.scheduler_delay = response['time'] - middle self.periodic_callbacks['heartbeat'].callback_time = response['heartbeat-interval'] * 1000 except CommClosedError: logger.warning("Heartbeat to scheduler failed") finally: self.heartbeat_active = False else: logger.debug("Heartbeat skipped: channel busy") @gen.coroutine def handle_scheduler(self, comm): try: yield self.handle_stream(comm, every_cycle=[self.ensure_communicating, self.ensure_computing]) except Exception as e: logger.exception(e) raise finally: if self.reconnect: logger.info("Connection to scheduler broken. Reconnecting...") self.loop.add_callback(self._register_with_scheduler) else: yield self._close(report=False) def start_ipython(self, comm): """Start an IPython kernel Returns Jupyter connection info dictionary. """ from ._ipython_utils import start_ipython if self._ipython_kernel is None: self._ipython_kernel = start_ipython( ip=self.ip, ns={'worker': self}, log=logger, ) return self._ipython_kernel.get_connection_info() @gen.coroutine def upload_file(self, comm, filename=None, data=None, load=True): out_filename = os.path.join(self.local_dir, filename) def func(data): if isinstance(data, unicode): data = data.encode() with open(out_filename, 'wb') as f: f.write(data) f.flush() return data if len(data) < 10000: data = func(data) else: data = yield offload(func, data) if load: try: import_file(out_filename) except Exception as e: logger.exception(e) raise gen.Return({'status': 'error', 'exception': to_serialize(e)}) raise gen.Return({'status': 'OK', 'nbytes': len(data)}) def keys(self, comm=None): return list(self.data) @gen.coroutine def gather(self, comm=None, who_has=None): who_has = {k: [coerce_to_address(addr) for addr in v] for k, v in who_has.items() if k not in self.data} result, missing_keys, missing_workers = yield gather_from_workers( who_has, rpc=self.rpc, who=self.address) if missing_keys: logger.warning("Could not find data: %s on workers: %s (who_has: %s)", missing_keys, missing_workers, who_has) raise Return({'status': 'missing-data', 'keys': missing_keys}) else: self.update_data(data=result, report=False) raise Return({'status': 'OK'}) def get_logs(self, comm=None, n=None): deque_handler = self._deque_handler if n is None: L = list(deque_handler.deque) else: L = deque_handler.deque L = [L[-i] for i in range(min(n, len(L)))] return [(msg.levelname, deque_handler.format(msg)) for msg in L] ############# # Lifecycle # ############# def start_services(self, listen_ip=''): for k, v in self.service_specs.items(): if isinstance(k, tuple): k, port = k else: port = 0 if isinstance(v, tuple): v, kwargs = v else: v, kwargs = v, {} self.services[k] = v(self, io_loop=self.loop, **kwargs) self.services[k].listen((listen_ip, port)) self.service_ports[k] = self.services[k].port @gen.coroutine def _start(self, addr_or_port=0): assert self.status is None enable_gc_diagnosis() thread_state.on_event_loop_thread = True # XXX Factor this out if not addr_or_port: # Default address is the required one to reach the scheduler listen_host = get_address_host(self.scheduler.address) self.listen(get_local_address_for(self.scheduler.address), listen_args=self.listen_args) self.ip = get_address_host(self.address) elif isinstance(addr_or_port, int): # addr_or_port is an integer => assume TCP listen_host = self.ip = get_ip( get_address_host(self.scheduler.address) ) self.listen((listen_host, addr_or_port), listen_args=self.listen_args) else: self.listen(addr_or_port, listen_args=self.listen_args) self.ip = get_address_host(self.address) try: listen_host = get_address_host(addr_or_port) except ValueError: listen_host = addr_or_port if '://' in listen_host: protocol, listen_host = listen_host.split('://') self.name = self.name or self.address preload_modules(self.preload, parameter=self, file_dir=self.local_dir, argv=self.preload_argv) # Services listen on all addresses # Note Nanny is not a "real" service, just some metadata # passed in service_ports... self.start_services(listen_host) try: listening_address = '%s%s:%d' % (self.listener.prefix, listen_host, self.port) except Exception: listening_address = '%s%s' % (self.listener.prefix, listen_host) logger.info(' Start worker at: %26s', self.address) logger.info(' Listening to: %26s', listening_address) for k, v in self.service_ports.items(): logger.info(' %16s at: %26s' % (k, listen_host + ':' + str(v))) logger.info('Waiting to connect to: %26s', self.scheduler.address) logger.info('-' * 49) logger.info(' Threads: %26d', self.ncores) if self.memory_limit: logger.info(' Memory: %26s', format_bytes(self.memory_limit)) logger.info(' Local Directory: %26s', self.local_dir) setproctitle("dask-worker [%s]" % self.address) yield self._register_with_scheduler() self.start_periodic_callbacks() def start(self, port=0): self.loop.add_callback(self._start, port) @gen.coroutine def _close(self, report=True, timeout=10, nanny=True, executor_wait=True): with log_errors(): if self.status in ('closed', 'closing'): return disable_gc_diagnosis() logger.info("Stopping worker at %s", self.address) self.status = 'closing' setproctitle("dask-worker [closing]") self.stop() for pc in self.periodic_callbacks.values(): pc.stop() with ignoring(EnvironmentError, gen.TimeoutError): if report: yield gen.with_timeout(timedelta(seconds=timeout), self.scheduler.unregister(address=self.contact_address)) self.scheduler.close_rpc() self.actor_executor._work_queue.queue.clear() if isinstance(self.executor, ThreadPoolExecutor): self.executor._work_queue.queue.clear() self.executor.shutdown(wait=executor_wait, timeout=timeout) else: self.executor.shutdown(wait=False) self.actor_executor.shutdown(wait=executor_wait, timeout=timeout) self._workdir.release() for k, v in self.services.items(): v.stop() self.status = 'closed' if nanny and 'nanny' in self.service_ports: with self.rpc((self.ip, self.service_ports['nanny'])) as r: yield r.terminate() if self.batched_stream and not self.batched_stream.comm.closed(): self.batched_stream.send({'op': 'close-stream'}) if self.batched_stream: self.batched_stream.close() self.rpc.close() self._closed.set() self._remove_from_global_workers() yield self.close() setproctitle("dask-worker [closed]") def __del__(self): self._remove_from_global_workers() def _remove_from_global_workers(self): for ref in list(_global_workers): if ref() is self: _global_workers.remove(ref) if ref() is None: _global_workers.remove(ref) @gen.coroutine def terminate(self, comm, report=True): yield self._close(report=report) raise Return('OK') @gen.coroutine def wait_until_closed(self): yield self._closed.wait() assert self.status == 'closed' ################ # Worker Peers # ################ def send_to_worker(self, address, msg): if address not in self.stream_comms: bcomm = BatchedSend(interval='1ms', loop=self.loop) self.stream_comms[address] = bcomm @gen.coroutine def batched_send_connect(): comm = yield connect(address, # TODO, serialization connection_args=self.connection_args) yield comm.write({'op': 'connection_stream'}) bcomm.start(comm) self.loop.add_callback(batched_send_connect) self.stream_comms[address].send(msg) @gen.coroutine def get_data(self, comm, keys=None, who=None, serializers=None, max_connections=None): start = time() if max_connections is None: max_connections = self.total_in_connections # Allow same-host connections more liberally if max_connections and comm and get_address_host(comm.peer_address) == get_address_host(self.address): max_connections = max_connections * 2 if max_connections is not False and self.outgoing_current_count > max_connections: raise gen.Return({'status': 'busy'}) self.outgoing_current_count += 1 data = {k: self.data[k] for k in keys if k in self.data} if len(data) < len(keys): for k in set(keys) - set(data): if k in self.actors: from .actor import Actor data[k] = Actor(type(self.actors[k]), self.address, k) msg = {'status': 'OK', 'data': {k: to_serialize(v) for k, v in data.items()}} nbytes = {k: self.nbytes.get(k) for k in data} stop = time() if self.digests is not None: self.digests['get-data-load-duration'].add(stop - start) start = time() try: compressed = yield comm.write(msg, serializers=serializers) response = yield comm.read(deserializers=serializers) assert response == 'OK', response except EnvironmentError: logger.exception('failed during get data with %s -> %s', self.address, who, exc_info=True) comm.abort() raise finally: self.outgoing_current_count -= 1 stop = time() if self.digests is not None: self.digests['get-data-send-duration'].add(stop - start) total_bytes = sum(filter(None, nbytes.values())) self.outgoing_count += 1 duration = (stop - start) or 0.5 # windows self.outgoing_transfer_log.append({ 'start': start + self.scheduler_delay, 'stop': stop + self.scheduler_delay, 'middle': (start + stop) / 2, 'duration': duration, 'who': who, 'keys': nbytes, 'total': total_bytes, 'compressed': compressed, 'bandwidth': total_bytes / duration }) raise gen.Return('dont-reply') ################### # Local Execution # ################### def update_data(self, comm=None, data=None, report=True, serializers=None): for key, value in data.items(): if key in self.task_state: self.transition(key, 'memory', value=value) else: self.put_key_in_memory(key, value) self.task_state[key] = 'memory' self.tasks[key] = None self.priorities[key] = None self.durations[key] = None self.dependencies[key] = set() if key in self.dep_state: self.transition_dep(key, 'memory', value=value) self.log.append((key, 'receive-from-scatter')) if report: self.batched_stream.send({'op': 'add-keys', 'keys': list(data)}) info = {'nbytes': {k: sizeof(v) for k, v in data.items()}, 'status': 'OK'} return info @gen.coroutine def delete_data(self, comm=None, keys=None, report=True): if keys: for key in list(keys): self.log.append((key, 'delete')) if key in self.task_state: self.release_key(key) if key in self.dep_state: self.release_dep(key) logger.debug("Deleted %d keys", len(keys)) if report: logger.debug("Reporting loss of keys to scheduler") # TODO: this route seems to not exist? yield self.scheduler.remove_keys(address=self.contact_address, keys=list(keys)) raise Return('OK') @gen.coroutine def set_resources(self, **resources): for r, quantity in resources.items(): if r in self.total_resources: self.available_resources[r] += quantity - self.total_resources[r] else: self.available_resources[r] = quantity self.total_resources[r] = quantity yield self.scheduler.set_resources(resources=self.total_resources, worker=self.contact_address) ################### # Task Management # ################### def add_task(self, key, function=None, args=None, kwargs=None, task=None, who_has=None, nbytes=None, priority=None, duration=None, resource_restrictions=None, actor=False, **kwargs2): try: if key in self.tasks: state = self.task_state[key] if state == 'memory': assert key in self.data or key in self.actors logger.debug("Asked to compute pre-existing result: %s: %s", key, state) self.send_task_state_to_scheduler(key) return if state in IN_PLAY: return if state == 'erred': del self.exceptions[key] del self.tracebacks[key] if priority is not None: priority = tuple(priority) + (self.generation,) self.generation -= 1 if self.dep_state.get(key) == 'memory': self.task_state[key] = 'memory' self.send_task_state_to_scheduler(key) self.tasks[key] = None self.log.append((key, 'new-task-already-in-memory')) self.priorities[key] = priority self.durations[key] = duration return self.log.append((key, 'new')) try: start = time() self.tasks[key] = _deserialize(function, args, kwargs, task) if actor: self.actors[key] = None stop = time() if stop - start > 0.010: self.startstops[key].append(('deserialize', start, stop)) except Exception as e: logger.warning("Could not deserialize task", exc_info=True) emsg = error_message(e) emsg['key'] = key emsg['op'] = 'task-erred' self.batched_stream.send(emsg) self.log.append((key, 'deserialize-error')) return self.priorities[key] = priority self.durations[key] = duration if resource_restrictions: self.resource_restrictions[key] = resource_restrictions self.task_state[key] = 'waiting' if nbytes is not None: self.nbytes.update(nbytes) who_has = who_has or {} self.dependencies[key] = set(who_has) self.waiting_for_data[key] = set() for dep in who_has: if dep not in self.dependents: self.dependents[dep] = set() self.dependents[dep].add(key) if dep not in self.dep_state: if self.task_state.get(dep) == 'memory': state = 'memory' else: state = 'waiting' self.dep_state[dep] = state self.log.append((dep, 'new-dep', state)) if self.dep_state[dep] != 'memory': self.waiting_for_data[key].add(dep) for dep, workers in who_has.items(): assert workers if dep not in self.who_has: self.who_has[dep] = set(workers) self.who_has[dep].update(workers) for worker in workers: self.has_what[worker].add(dep) if self.dep_state[dep] != 'memory': self.pending_data_per_worker[worker].append(dep) if self.waiting_for_data[key]: self.data_needed.append(key) else: self.transition(key, 'ready') if self.validate: if who_has: assert all(dep in self.dep_state for dep in who_has) assert all(dep in self.nbytes for dep in who_has) for dep in who_has: self.validate_dep(dep) self.validate_key(key) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_dep(self, dep, finish, **kwargs): try: start = self.dep_state[dep] except KeyError: return if start == finish: return func = self._dep_transitions[start, finish] state = func(dep, **kwargs) self.log.append(('dep', dep, start, state or finish)) if dep in self.dep_state: self.dep_state[dep] = state or finish if self.validate: self.validate_dep(dep) def transition_dep_waiting_flight(self, dep, worker=None): try: if self.validate: assert dep not in self.in_flight_tasks assert self.dependents[dep] self.in_flight_tasks[dep] = worker except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_dep_flight_waiting(self, dep, worker=None, remove=True): try: if self.validate: assert dep in self.in_flight_tasks del self.in_flight_tasks[dep] if remove: try: self.who_has[dep].remove(worker) except KeyError: pass try: self.has_what[worker].remove(dep) except KeyError: pass if not self.who_has.get(dep): if dep not in self._missing_dep_flight: self._missing_dep_flight.add(dep) self.loop.add_callback(self.handle_missing_dep, dep) for key in self.dependents.get(dep, ()): if self.task_state[key] == 'waiting': if remove: # try a new worker immediately self.data_needed.appendleft(key) else: # worker was probably busy, wait a while self.data_needed.append(key) if not self.dependents[dep]: self.release_dep(dep) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_dep_flight_memory(self, dep, value=None): try: if self.validate: assert dep in self.in_flight_tasks del self.in_flight_tasks[dep] if self.dependents[dep]: self.dep_state[dep] = 'memory' self.put_key_in_memory(dep, value) self.batched_stream.send({'op': 'add-keys', 'keys': [dep]}) else: self.release_dep(dep) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_dep_waiting_memory(self, dep, value=None): try: if self.validate: assert dep in self.data assert dep in self.nbytes assert dep in self.types assert self.task_state[dep] == 'memory' except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise if value is not no_value and dep not in self.data: self.put_key_in_memory(dep, value, transition=False) def transition(self, key, finish, **kwargs): start = self.task_state[key] if start == finish: return func = self._transitions[start, finish] state = func(key, **kwargs) self.log.append((key, start, state or finish)) self.task_state[key] = state or finish if self.validate: self.validate_key(key) def transition_waiting_ready(self, key): try: if self.validate: assert self.task_state[key] == 'waiting' assert key in self.waiting_for_data assert not self.waiting_for_data[key] assert all(dep in self.data or dep in self.actors for dep in self.dependencies[key]) assert key not in self.executing assert key not in self.ready self.waiting_for_data.pop(key, None) if key in self.resource_restrictions: self.constrained.append(key) return 'constrained' else: heapq.heappush(self.ready, (self.priorities[key], key)) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_waiting_done(self, key, value=None): try: if self.validate: assert self.task_state[key] == 'waiting' assert key in self.waiting_for_data assert key not in self.executing assert key not in self.ready del self.waiting_for_data[key] self.send_task_state_to_scheduler(key) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_ready_executing(self, key): try: if self.validate: assert key not in self.waiting_for_data # assert key not in self.data assert self.task_state[key] in READY assert key not in self.ready assert all(dep in self.data or dep in self.actors for dep in self.dependencies[key]) self.executing.add(key) self.loop.add_callback(self.execute, key) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_ready_memory(self, key, value=None): self.send_task_state_to_scheduler(key) def transition_constrained_executing(self, key): self.transition_ready_executing(key) for resource, quantity in self.resource_restrictions[key].items(): self.available_resources[resource] -= quantity if self.validate: assert all(v >= 0 for v in self.available_resources.values()) def transition_executing_done(self, key, value=no_value, report=True): try: if self.validate: assert key in self.executing or key in self.long_running assert key not in self.waiting_for_data assert key not in self.ready out = None if key in self.resource_restrictions: for resource, quantity in self.resource_restrictions[key].items(): self.available_resources[resource] += quantity if self.task_state[key] == 'executing': self.executing.remove(key) self.executed_count += 1 elif self.task_state[key] == 'long-running': self.long_running.remove(key) if value is not no_value: try: self.task_state[key] = 'memory' self.put_key_in_memory(key, value, transition=False) except Exception as e: logger.info("Failed to put key in memory", exc_info=True) msg = error_message(e) self.exceptions[key] = msg['exception'] self.tracebacks[key] = msg['traceback'] self.task_state[key] = 'error' out = 'error' if key in self.dep_state: self.transition_dep(key, 'memory') if report and self.batched_stream: self.send_task_state_to_scheduler(key) else: raise CommClosedError return out except EnvironmentError: logger.info("Comm closed") except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def transition_executing_long_running(self, key, compute_duration=None): try: if self.validate: assert key in self.executing self.executing.remove(key) self.long_running.add(key) self.batched_stream.send({'op': 'long-running', 'key': key, 'compute_duration': compute_duration}) self.ensure_computing() except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def maybe_transition_long_running(self, key, compute_duration=None): if self.task_state.get(key) == 'executing': self.transition(key, 'long-running', compute_duration=compute_duration) def stateof(self, key): return {'executing': key in self.executing, 'waiting_for_data': key in self.waiting_for_data, 'heap': key in pluck(1, self.ready), 'data': key in self.data} def story(self, *keys): return [msg for msg in self.log if any(key in msg for key in keys) or any(key in c for key in keys for c in msg if isinstance(c, (tuple, list, set)))] def ensure_communicating(self): changed = True try: while changed and self.data_needed and len(self.in_flight_workers) < self.total_out_connections: changed = False logger.debug("Ensure communicating. Pending: %d. Connections: %d/%d", len(self.data_needed), len(self.in_flight_workers), self.total_out_connections) key = self.data_needed[0] if key not in self.tasks: self.data_needed.popleft() changed = True continue if self.task_state.get(key) != 'waiting': self.log.append((key, 'communication pass')) self.data_needed.popleft() changed = True continue deps = self.dependencies[key] if self.validate: assert all(dep in self.dep_state for dep in deps) deps = [dep for dep in deps if self.dep_state[dep] == 'waiting'] missing_deps = {dep for dep in deps if not self.who_has.get(dep)} if missing_deps: logger.info("Can't find dependencies for key %s", key) missing_deps2 = {dep for dep in missing_deps if dep not in self._missing_dep_flight} for dep in missing_deps2: self._missing_dep_flight.add(dep) self.loop.add_callback(self.handle_missing_dep, *missing_deps2) deps = [dep for dep in deps if dep not in missing_deps] self.log.append(('gather-dependencies', key, deps)) in_flight = False while deps and (len(self.in_flight_workers) < self.total_out_connections or self.comm_nbytes < self.total_comm_nbytes): dep = deps.pop() if self.dep_state[dep] != 'waiting': continue if dep not in self.who_has: continue workers = [w for w in self.who_has[dep] if w not in self.in_flight_workers] if not workers: in_flight = True continue host = get_address_host(self.address) local = [w for w in workers if get_address_host(w) == host] if local: worker = random.choice(local) else: worker = random.choice(list(workers)) to_gather, total_nbytes = self.select_keys_for_gather(worker, dep) self.comm_nbytes += total_nbytes self.in_flight_workers[worker] = to_gather for d in to_gather: self.transition_dep(d, 'flight', worker=worker) self.loop.add_callback(self.gather_dep, worker, dep, to_gather, total_nbytes, cause=key) changed = True if not deps and not in_flight: self.data_needed.popleft() except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def send_task_state_to_scheduler(self, key): if key in self.data or self.actors.get(key): try: value = self.data[key] except KeyError: value = self.actors[key] nbytes = self.nbytes[key] or sizeof(value) typ = self.types.get(key) or type(value) del value try: typ = dumps_function(typ) except PicklingError: # Some types fail pickling (example: _thread.lock objects), # send their name as a best effort. typ = pickle.dumps(typ.__name__) d = {'op': 'task-finished', 'status': 'OK', 'key': key, 'nbytes': nbytes, 'thread': self.threads.get(key), 'type': typ} elif key in self.exceptions: d = {'op': 'task-erred', 'status': 'error', 'key': key, 'thread': self.threads.get(key), 'exception': self.exceptions[key], 'traceback': self.tracebacks[key]} else: logger.error("Key not ready to send to worker, %s: %s", key, self.task_state[key]) return if key in self.startstops: d['startstops'] = self.startstops[key] self.batched_stream.send(d) def put_key_in_memory(self, key, value, transition=True): if key in self.data: return if key in self.actors: self.actors[key] = value else: start = time() self.data[key] = value stop = time() if stop - start > 0.020: self.startstops[key].append(('disk-write', start, stop)) if key not in self.nbytes: self.nbytes[key] = sizeof(value) self.types[key] = type(value) for dep in self.dependents.get(key, ()): if dep in self.waiting_for_data: if key in self.waiting_for_data[dep]: self.waiting_for_data[dep].remove(key) if not self.waiting_for_data[dep]: self.transition(dep, 'ready') if transition and key in self.task_state: self.transition(key, 'memory') self.log.append((key, 'put-in-memory')) def select_keys_for_gather(self, worker, dep): deps = {dep} total_bytes = self.nbytes[dep] L = self.pending_data_per_worker[worker] while L: d = L.popleft() if self.dep_state.get(d) != 'waiting': continue if total_bytes + self.nbytes[d] > self.target_message_size: break deps.add(d) total_bytes += self.nbytes[d] return deps, total_bytes @gen.coroutine def gather_dep(self, worker, dep, deps, total_nbytes, cause=None): if self.status != 'running': return with log_errors(): response = {} try: if self.validate: self.validate_state() # dep states may have changed before gather_dep runs # if a dep is no longer in-flight then don't fetch it deps = tuple(dep for dep in deps if self.dep_state.get(dep) == 'flight') self.log.append(('request-dep', dep, worker, deps)) logger.debug("Request %d keys", len(deps)) start = time() response = yield get_data_from_worker(self.rpc, deps, worker, who=self.address) stop = time() if response['status'] == 'busy': self.log.append(('busy-gather', worker, deps)) for dep in deps: if self.dep_state.get(dep, None) == 'flight': self.transition_dep(dep, 'waiting') return if cause: self.startstops[cause].append(( 'transfer', start + self.scheduler_delay, stop + self.scheduler_delay )) total_bytes = sum(self.nbytes.get(dep, 0) for dep in response['data']) duration = (stop - start) or 0.5 self.incoming_transfer_log.append({ 'start': start + self.scheduler_delay, 'stop': stop + self.scheduler_delay, 'middle': (start + stop) / 2.0 + self.scheduler_delay, 'duration': duration, 'keys': {dep: self.nbytes.get(dep, None) for dep in response['data']}, 'total': total_bytes, 'bandwidth': total_bytes / duration, 'who': worker }) if self.digests is not None: self.digests['transfer-bandwidth'].add(total_bytes / duration) self.digests['transfer-duration'].add(duration) self.counters['transfer-count'].add(len(response['data'])) self.incoming_count += 1 self.log.append(('receive-dep', worker, list(response['data']))) if response['data']: self.batched_stream.send({'op': 'add-keys', 'keys': list(response['data'])}) except EnvironmentError as e: logger.exception("Worker stream died during communication: %s", worker) self.log.append(('receive-dep-failed', worker)) for d in self.has_what.pop(worker): self.who_has[d].remove(worker) if not self.who_has[d]: del self.who_has[d] except Exception as e: logger.exception(e) if self.batched_stream and LOG_PDB: import pdb pdb.set_trace() raise finally: self.comm_nbytes -= total_nbytes busy = response.get('status', '') == 'busy' data = response.get('data', {}) for d in self.in_flight_workers.pop(worker): if not busy and d in data: self.transition_dep(d, 'memory', value=data[d]) elif self.dep_state.get(d) != 'memory': self.transition_dep(d, 'waiting', worker=worker, remove=not busy) if not busy and d not in data and d in self.dependents: self.log.append(('missing-dep', d)) self.batched_stream.send({'op': 'missing-data', 'errant_worker': worker, 'key': d}) if self.validate: self.validate_state() self.ensure_computing() if not busy: self.repetitively_busy = 0 self.ensure_communicating() else: # Exponential backoff to avoid hammering scheduler/worker self.repetitively_busy += 1 yield gen.sleep(0.100 * 1.5 ** self.repetitively_busy) # See if anyone new has the data yield self.query_who_has(dep) self.ensure_communicating() def bad_dep(self, dep): exc = ValueError("Could not find dependent %s. Check worker logs" % str(dep)) for key in self.dependents[dep]: msg = error_message(exc) self.exceptions[key] = msg['exception'] self.tracebacks[key] = msg['traceback'] self.transition(key, 'error') self.release_dep(dep) @gen.coroutine def handle_missing_dep(self, *deps, **kwargs): original_deps = list(deps) self.log.append(('handle-missing', deps)) try: deps = {dep for dep in deps if dep in self.dependents} if not deps: return for dep in list(deps): suspicious = self.suspicious_deps[dep] if suspicious > 5: deps.remove(dep) self.bad_dep(dep) if not deps: return for dep in deps: logger.info("Dependent not found: %s %s . Asking scheduler", dep, self.suspicious_deps[dep]) who_has = yield self.scheduler.who_has(keys=list(deps)) who_has = {k: v for k, v in who_has.items() if v} self.update_who_has(who_has) for dep in deps: self.suspicious_deps[dep] += 1 if not who_has.get(dep): self.log.append((dep, 'no workers found', self.dependents.get(dep))) self.release_dep(dep) else: self.log.append((dep, 'new workers found')) for key in self.dependents.get(dep, ()): if key in self.waiting_for_data: self.data_needed.append(key) except Exception: logger.error("Handle missing dep failed, retrying", exc_info=True) retries = kwargs.get('retries', 5) self.log.append(('handle-missing-failed', retries, deps)) if retries > 0: yield self.handle_missing_dep(self, *deps, retries=retries - 1) else: raise finally: try: for dep in original_deps: self._missing_dep_flight.remove(dep) except KeyError: pass self.ensure_communicating() @gen.coroutine def query_who_has(self, *deps): with log_errors(): response = yield self.scheduler.who_has(keys=deps) self.update_who_has(response) raise gen.Return(response) def update_who_has(self, who_has): try: for dep, workers in who_has.items(): if not workers: continue if dep in self.who_has: self.who_has[dep].update(workers) else: self.who_has[dep] = set(workers) for worker in workers: self.has_what[worker].add(dep) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def steal_request(self, key): state = self.task_state.get(key, None) response = {'op': 'steal-response', 'key': key, 'state': state} self.batched_stream.send(response) if state in ('ready', 'waiting'): self.release_key(key) def release_key(self, key, cause=None, reason=None, report=True): try: if key not in self.task_state: return state = self.task_state.pop(key) if cause: self.log.append((key, 'release-key', {'cause': cause})) else: self.log.append((key, 'release-key')) del self.tasks[key] if key in self.data and key not in self.dep_state: try: del self.data[key] except FileNotFoundError: logger.error("Tried to delete %s but no file found", exc_info=True) del self.nbytes[key] del self.types[key] if key in self.actors and key not in self.dep_state: del self.actors[key] del self.nbytes[key] del self.types[key] if key in self.waiting_for_data: del self.waiting_for_data[key] for dep in self.dependencies.pop(key, ()): if dep in self.dependents: self.dependents[dep].discard(key) if not self.dependents[dep] and self.dep_state[dep] in ('waiting', 'flight'): self.release_dep(dep) if key in self.threads: del self.threads[key] del self.priorities[key] del self.durations[key] if key in self.exceptions: del self.exceptions[key] if key in self.tracebacks: del self.tracebacks[key] if key in self.startstops: del self.startstops[key] if key in self.executing: self.executing.remove(key) if key in self.resource_restrictions: if state == 'executing': for resource, quantity in self.resource_restrictions[key].items(): self.available_resources[resource] += quantity del self.resource_restrictions[key] if report and state in PROCESSING: # not finished self.batched_stream.send({'op': 'release', 'key': key, 'cause': cause}) except CommClosedError: pass except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def release_dep(self, dep, report=False): try: if dep not in self.dep_state: return self.log.append((dep, 'release-dep')) state = self.dep_state.pop(dep) if dep in self.suspicious_deps: del self.suspicious_deps[dep] if dep in self.who_has: for worker in self.who_has.pop(dep): self.has_what[worker].remove(dep) if dep not in self.task_state: if dep in self.data: del self.data[dep] del self.types[dep] if dep in self.actors: del self.actors[dep] del self.types[dep] del self.nbytes[dep] if dep in self.in_flight_tasks: worker = self.in_flight_tasks.pop(dep) self.in_flight_workers[worker].remove(dep) for key in self.dependents.pop(dep, ()): if self.task_state[key] != 'memory': self.release_key(key, cause=dep) if report and state == 'memory': self.batched_stream.send({'op': 'release-worker-data', 'keys': [dep]}) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def rescind_key(self, key): try: if self.task_state.get(key) not in PENDING: return del self.task_state[key] del self.tasks[key] if key in self.waiting_for_data: del self.waiting_for_data[key] for dep in self.dependencies.pop(key, ()): self.dependents[dep].remove(key) if not self.dependents[dep]: del self.dependents[dep] if key not in self.dependents: # if key in self.nbytes: # del self.nbytes[key] if key in self.priorities: del self.priorities[key] if key in self.durations: del self.durations[key] except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise ################ # Execute Task # ################ @gen.coroutine def executor_submit(self, key, function, args=(), kwargs=None, executor=None): """ Safely run function in thread pool executor We've run into issues running concurrent.future futures within tornado. Apparently it's advantageous to use timeouts and periodic callbacks to ensure things run smoothly. This can get tricky, so we pull it off into an separate method. """ executor = executor or self.executor job_counter[0] += 1 # logger.info("%s:%d Starts job %d, %s", self.ip, self.port, i, key) kwargs = kwargs or {} future = executor.submit(function, *args, **kwargs) pc = PeriodicCallback(lambda: logger.debug("future state: %s - %s", key, future._state), 1000) pc.start() try: yield future finally: pc.stop() result = future.result() # logger.info("Finish job %d, %s", i, key) raise gen.Return(result) def run(self, comm, function, args=(), wait=True, kwargs=None): kwargs = kwargs or {} return run(self, comm, function=function, args=args, kwargs=kwargs, wait=wait) def run_coroutine(self, comm, function, args=(), kwargs=None, wait=True): return run(self, comm, function=function, args=args, kwargs=kwargs, wait=wait) @gen.coroutine def actor_execute(self, comm=None, actor=None, function=None, args=(), kwargs={}): separate_thread = kwargs.pop('separate_thread', True) key = actor actor = self.actors[key] func = getattr(actor, function) name = key_split(key) + '.' + function if iscoroutinefunction(func): result = yield func(*args, **kwargs) elif separate_thread: result = yield self.executor_submit(name, apply_function_actor, args=(func, args, kwargs, self.execution_state, name, self.active_threads, self.active_threads_lock), executor=self.actor_executor) else: result = func(*args, **kwargs) raise gen.Return({'status': 'OK', 'result': to_serialize(result)}) def actor_attribute(self, comm=None, actor=None, attribute=None): value = getattr(self.actors[actor], attribute) return {'status': 'OK', 'result': to_serialize(value)} def meets_resource_constraints(self, key): if key not in self.resource_restrictions: return True for resource, needed in self.resource_restrictions[key].items(): if self.available_resources[resource] < needed: return False return True def ensure_computing(self): if self.paused: return try: while self.constrained and len(self.executing) < self.ncores: key = self.constrained[0] if self.task_state.get(key) != 'constrained': self.constrained.popleft() continue if self.meets_resource_constraints(key): self.constrained.popleft() self.transition(key, 'executing') else: break while self.ready and len(self.executing) < self.ncores: _, key = heapq.heappop(self.ready) if self.task_state.get(key) in READY: self.transition(key, 'executing') except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise @gen.coroutine def execute(self, key, report=False): executor_error = None if self.status in ('closing', 'closed'): return try: if key not in self.executing or key not in self.task_state: return if self.validate: assert key not in self.waiting_for_data assert self.task_state[key] == 'executing' function, args, kwargs = self.tasks[key] start = time() data = {} for k in self.dependencies[key]: try: data[k] = self.data[k] except KeyError: from .actor import Actor # TODO: create local actor data[k] = Actor(type(self.actors[k]), self.address, k, self) args2 = pack_data(args, data, key_types=(bytes, unicode)) kwargs2 = pack_data(kwargs, data, key_types=(bytes, unicode)) stop = time() if stop - start > 0.005: self.startstops[key].append(('disk-read', start, stop)) if self.digests is not None: self.digests['disk-load-duration'].add(stop - start) logger.debug("Execute key: %s worker: %s", key, self.address) # TODO: comment out? try: result = yield self.executor_submit(key, apply_function, args=(function, args2, kwargs2, self.execution_state, key, self.active_threads, self.active_threads_lock, self.scheduler_delay)) except RuntimeError as e: executor_error = e raise if self.task_state.get(key) not in ('executing', 'long-running'): return result['key'] = key value = result.pop('result', None) self.startstops[key].append(('compute', result['start'], result['stop'])) self.threads[key] = result['thread'] if result['op'] == 'task-finished': self.nbytes[key] = result['nbytes'] self.types[key] = result['type'] self.transition(key, 'memory', value=value) if self.digests is not None: self.digests['task-duration'].add(result['stop'] - result['start']) else: if isinstance(result.pop('actual-exception'), Reschedule): self.batched_stream.send({'op': 'reschedule', 'key': key}) self.transition(key, 'rescheduled', report=False) self.release_key(key, report=False) else: self.exceptions[key] = result['exception'] self.tracebacks[key] = result['traceback'] logger.warning(" Compute Failed\n" "Function: %s\n" "args: %s\n" "kwargs: %s\n" "Exception: %s\n", str(funcname(function))[:1000], convert_args_to_str(args2, max_len=1000), convert_kwargs_to_str(kwargs2, max_len=1000), repr(result['exception'].data)) self.transition(key, 'error') logger.debug("Send compute response to scheduler: %s, %s", key, result) if self.validate: assert key not in self.executing assert key not in self.waiting_for_data self.ensure_computing() self.ensure_communicating() except Exception as e: if executor_error is e: logger.error("Thread Pool Executor error: %s", e) else: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise finally: if key in self.executing: self.executing.remove(key) ################## # Administrative # ################## @gen.coroutine def memory_monitor(self): """ Track this process's memory usage and act accordingly If we rise above 70% memory use, start dumping data to disk. If we rise above 80% memory use, stop execution of new tasks """ if self._memory_monitoring: return self._memory_monitoring = True total = 0 proc = self.monitor.proc memory = proc.memory_info().rss frac = memory / self.memory_limit # Pause worker threads if above 80% memory use if self.memory_pause_fraction and frac > self.memory_pause_fraction: # Try to free some memory while in paused state self._throttled_gc.collect() if not self.paused: logger.warning("Worker is at %d%% memory usage. Pausing worker. " "Process memory: %s -- Worker memory limit: %s", int(frac * 100), format_bytes(proc.memory_info().rss), format_bytes(self.memory_limit)) self.paused = True elif self.paused: logger.warning("Worker is at %d%% memory usage. Resuming worker. " "Process memory: %s -- Worker memory limit: %s", int(frac * 100), format_bytes(proc.memory_info().rss), format_bytes(self.memory_limit)) self.paused = False self.ensure_computing() # Dump data to disk if above 70% if self.memory_spill_fraction and frac > self.memory_spill_fraction: target = self.memory_limit * self.memory_target_fraction count = 0 need = memory - target while memory > target: if not self.data.fast: logger.warning("Memory use is high but worker has no data " "to store to disk. Perhaps some other process " "is leaking memory? Process memory: %s -- " "Worker memory limit: %s", format_bytes(proc.memory_info().rss), format_bytes(self.memory_limit)) break k, v, weight = self.data.fast.evict() del k, v total += weight count += 1 yield gen.moment memory = proc.memory_info().rss if total > need and memory > target: # Issue a GC to ensure that the evicted data is actually # freed from memory and taken into account by the monitor # before trying to evict even more data. self._throttled_gc.collect() memory = proc.memory_info().rss if count: logger.debug("Moved %d pieces of data data and %s to disk", count, format_bytes(total)) self._memory_monitoring = False raise gen.Return(total) def cycle_profile(self): now = time() + self.scheduler_delay prof, self.profile_recent = self.profile_recent, profile.create() self.profile_history.append((now, prof)) self.profile_keys_history.append((now, dict(self.profile_keys))) self.profile_keys.clear() def trigger_profile(self): """ Get a frame from all actively computing threads Merge these frames into existing profile counts """ if not self.active_threads: # hope that this is thread-atomic? return start = time() with self.active_threads_lock: active_threads = self.active_threads.copy() frames = sys._current_frames() frames = {ident: frames[ident] for ident in active_threads} for ident, frame in frames.items(): if frame is not None: key = key_split(active_threads[ident]) profile.process(frame, None, self.profile_recent, stop='distributed/worker.py') profile.process(frame, None, self.profile_keys[key], stop='distributed/worker.py') stop = time() if self.digests is not None: self.digests['profile-duration'].add(stop - start) def get_profile(self, comm=None, start=None, stop=None, key=None): now = time() + self.scheduler_delay if key is None: history = self.profile_history else: history = [(t, d[key]) for t, d in self.profile_keys_history if key in d] if start is None: istart = 0 else: istart = bisect.bisect_left(history, (start,)) if stop is None: istop = None else: istop = bisect.bisect_right(history, (stop,)) + 1 if istop >= len(history): istop = None # include end if istart == 0 and istop is None: history = list(history) else: iistop = len(history) if istop is None else istop history = [history[i] for i in range(istart, iistop)] prof = profile.merge(*pluck(1, history)) if not history: return profile.create() if istop is None and (start is None or start < now): if key is None: recent = self.profile_recent else: recent = self.profile_keys[key] prof = profile.merge(prof, recent) return prof def get_profile_metadata(self, comm=None, start=0, stop=None): if stop is None: add_recent = True now = time() + self.scheduler_delay stop = stop or now start = start or 0 result = {'counts': [(t, d['count']) for t, d in self.profile_history if start < t < stop], 'keys': [(t, {k: d['count'] for k, d in v.items()}) for t, v in self.profile_keys_history if start < t < stop]} if add_recent: result['counts'].append((now, self.profile_recent['count'])) result['keys'].append((now, {k: v['count'] for k, v in self.profile_keys.items()})) return result def get_call_stack(self, comm=None, keys=None): with self.active_threads_lock: frames = sys._current_frames() active_threads = self.active_threads.copy() frames = {k: frames[ident] for ident, k in active_threads.items()} if keys is not None: frames = {k: frame for k, frame in frames.items() if k in keys} result = {k: profile.call_stack(frame) for k, frame in frames.items()} return result ############## # Validation # ############## def validate_key_memory(self, key): assert key in self.data or key in self.actors assert key in self.nbytes assert key not in self.waiting_for_data assert key not in self.executing assert key not in self.ready if key in self.dep_state: assert self.dep_state[key] == 'memory' def validate_key_executing(self, key): assert key in self.executing assert key not in self.data assert key not in self.waiting_for_data assert all(dep in self.data or dep in self.actors for dep in self.dependencies[key]) def validate_key_ready(self, key): assert key in pluck(1, self.ready) assert key not in self.data assert key not in self.executing assert key not in self.waiting_for_data assert all(dep in self.data or dep in self.actors for dep in self.dependencies[key]) def validate_key_waiting(self, key): assert key not in self.data assert not all(dep in self.data for dep in self.dependencies[key]) def validate_key(self, key): try: state = self.task_state[key] if state == 'memory': self.validate_key_memory(key) elif state == 'waiting': self.validate_key_waiting(key) elif state == 'ready': self.validate_key_ready(key) elif state == 'executing': self.validate_key_executing(key) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def validate_dep_waiting(self, dep): assert dep not in self.data assert dep in self.nbytes assert self.dependents[dep] assert not any(key in self.ready for key in self.dependents[dep]) def validate_dep_flight(self, dep): assert dep not in self.data assert dep in self.nbytes assert not any(key in self.ready for key in self.dependents[dep]) peer = self.in_flight_tasks[dep] assert dep in self.in_flight_workers[peer] def validate_dep_memory(self, dep): assert dep in self.data or dep in self.actors assert dep in self.nbytes assert dep in self.types if dep in self.task_state: assert self.task_state[dep] == 'memory' def validate_dep(self, dep): try: state = self.dep_state[dep] if state == 'waiting': self.validate_dep_waiting(dep) elif state == 'flight': self.validate_dep_flight(dep) elif state == 'memory': self.validate_dep_memory(dep) else: raise ValueError("Unknown dependent state", state) except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise def validate_state(self): if self.status != 'running': return try: for key, workers in self.who_has.items(): for w in workers: assert key in self.has_what[w] for worker, keys in self.has_what.items(): for k in keys: assert worker in self.who_has[k] for key in self.task_state: self.validate_key(key) for dep in self.dep_state: self.validate_dep(dep) for key, deps in self.waiting_for_data.items(): if key not in self.data_needed: for dep in deps: assert (dep in self.in_flight_tasks or dep in self._missing_dep_flight or self.who_has[dep].issubset(self.in_flight_workers)) for key in self.tasks: if self.task_state[key] == 'memory': assert isinstance(self.nbytes[key], int) assert key not in self.waiting_for_data assert key in self.data or key in self.actors except Exception as e: logger.exception(e) if LOG_PDB: import pdb pdb.set_trace() raise ####################################### # Worker Clients (advanced workloads) # ####################################### @property def client(self): with self._lock: if self._client: return self._client else: return self._get_client() def _get_client(self, timeout=3): """ Get local client attached to this worker If no such client exists, create one See Also -------- get_client """ try: from .client import default_client client = default_client() except ValueError: # no clients found, need to make a new one pass else: if (client.scheduler and client.scheduler.address == self.scheduler.address or client._start_arg == self.scheduler.address): self._client = client if not self._client: from .client import Client asynchronous = self.loop is IOLoop.current() self._client = Client(self.scheduler, loop=self.loop, security=self.security, set_as_default=True, asynchronous=asynchronous, direct_to_workers=True, name='worker', timeout=timeout) if not asynchronous: assert self._client.status == 'running' return self._client def get_current_task(self): """ Get the key of the task we are currently running This only makes sense to run within a task Examples -------- >>> from dask.distributed import get_worker >>> def f(): ... return get_worker().get_current_task() >>> future = client.submit(f) # doctest: +SKIP >>> future.result() # doctest: +SKIP 'f-1234' See Also -------- get_worker """ return self.active_threads[get_thread_identity()]
[docs]def get_worker(): """ Get the worker currently running this task Examples -------- >>> def f(): ... worker = get_worker() # The worker on which this task is running ... return worker.address >>> future = client.submit(f) # doctest: +SKIP >>> future.result() # doctest: +SKIP 'tcp://127.0.0.1:47373' See Also -------- get_client worker_client """ try: return thread_state.execution_state['worker'] except AttributeError: for ref in _global_workers[::-1]: worker = ref() if worker: return worker raise ValueError("No workers found")
[docs]def get_client(address=None, timeout=3, resolve_address=True): """Get a client while within a task. This client connects to the same scheduler to which the worker is connected Parameters ---------- address : str, optional The address of the scheduler to connect to. Defaults to the scheduler the worker is connected to. timeout : int, default 3 Timeout (in seconds) for getting the Client resolve_address : bool, default True Whether to resolve `address` to its canonical form. Returns ------- Client Examples -------- >>> def f(): ... client = get_client() ... futures = client.map(lambda x: x + 1, range(10)) # spawn many tasks ... results = client.gather(futures) ... return sum(results) >>> future = client.submit(f) # doctest: +SKIP >>> future.result() # doctest: +SKIP 55 See Also -------- get_worker worker_client secede """ if address and resolve_address: address = comm.resolve_address(address) try: worker = get_worker() except ValueError: # could not find worker pass else: if not address or worker.scheduler.address == address: return worker._get_client(timeout=timeout) from .client import _get_global_client client = _get_global_client() # TODO: assumes the same scheduler if client and (not address or client.scheduler.address == address): return client elif address: from .client import Client return Client(address, timeout=timeout) else: raise ValueError("No global client found and no address provided")
[docs]def secede(): """ Have this task secede from the worker's thread pool This opens up a new scheduling slot and a new thread for a new task. This enables the client to schedule tasks on this node, which is especially useful while waiting for other jobs to finish (e.g., with ``client.gather``). Examples -------- >>> def mytask(x): ... # do some work ... client = get_client() ... futures = client.map(...) # do some remote work ... secede() # while that work happens, remove ourself from the pool ... return client.gather(futures) # return gathered results See Also -------- get_client get_worker """ worker = get_worker() tpe_secede() # have this thread secede from the thread pool duration = time() - thread_state.start_time worker.loop.add_callback(worker.maybe_transition_long_running, thread_state.key, compute_duration=duration)
class Reschedule(Exception): """ Reschedule this task Raising this exception will stop the current execution of the task and ask the scheduler to reschedule this task, possibly on a different machine. This does not guarantee that the task will move onto a different machine. The scheduler will proceed through its normal heuristics to determine the optimal machine to accept this task. The machine will likely change if the load across the cluster has significantly changed since first scheduling the task. """ pass def parse_memory_limit(memory_limit, ncores): if memory_limit is None: return None if memory_limit == 'auto': memory_limit = int(TOTAL_MEMORY * min(1, ncores / _ncores)) with ignoring(ValueError, TypeError): x = float(memory_limit) if isinstance(x, float) and x <= 1: return int(x * TOTAL_MEMORY) if isinstance(memory_limit, (unicode, str)): return parse_bytes(memory_limit) else: return int(memory_limit) @gen.coroutine def get_data_from_worker(rpc, keys, worker, who=None, max_connections=None, serializers=None, deserializers=None): """ Get keys from worker The worker has a two step handshake to acknowledge when data has been fully delivered. This function implements that handshake. See Also -------- Worker.get_data Worker.gather_deps utils_comm.gather_data_from_workers """ if serializers is None: serializers = rpc.serializers if deserializers is None: deserializers = rpc.deserializers comm = yield rpc.connect(worker) try: response = yield send_recv(comm, serializers=serializers, deserializers=deserializers, op='get_data', keys=keys, who=who, max_connections=max_connections) try: status = response['status'] except KeyError: raise ValueError("Unexpected response", response) else: if status == 'OK': yield comm.write('OK') finally: rpc.reuse(worker, comm) raise gen.Return(response) job_counter = [0] def _deserialize(function=None, args=None, kwargs=None, task=None): """ Deserialize task inputs and regularize to func, args, kwargs """ if function is not None: function = pickle.loads(function) if args: args = pickle.loads(args) if kwargs: kwargs = pickle.loads(kwargs) if task is not None: assert not function and not args and not kwargs function = execute_task args = (task,) return function, args or (), kwargs or {} def execute_task(task): """ Evaluate a nested task >>> inc = lambda x: x + 1 >>> execute_task((inc, 1)) 2 >>> execute_task((sum, [1, 2, (inc, 3)])) 7 """ if istask(task): func, args = task[0], task[1:] return func(*map(execute_task, args)) elif isinstance(task, list): return list(map(execute_task, task)) else: return task cache = dict() def dumps_function(func): """ Dump a function to bytes, cache functions """ try: result = cache[func] except KeyError: result = pickle.dumps(func) if len(result) < 100000: cache[func] = result except TypeError: result = pickle.dumps(func) return result def dumps_task(task): """ Serialize a dask task Returns a dict of bytestrings that can each be loaded with ``loads`` Examples -------- Either returns a task as a function, args, kwargs dict >>> from operator import add >>> dumps_task((add, 1)) # doctest: +SKIP {'function': b'\x80\x04\x95\x00\x8c\t_operator\x94\x8c\x03add\x94\x93\x94.' 'args': b'\x80\x04\x95\x07\x00\x00\x00K\x01K\x02\x86\x94.'} Or as a single task blob if it can't easily decompose the result. This happens either if the task is highly nested, or if it isn't a task at all >>> dumps_task(1) # doctest: +SKIP {'task': b'\x80\x04\x95\x03\x00\x00\x00\x00\x00\x00\x00K\x01.'} """ if istask(task): if task[0] is apply and not any(map(_maybe_complex, task[2:])): d = {'function': dumps_function(task[1]), 'args': warn_dumps(task[2])} if len(task) == 4: d['kwargs'] = warn_dumps(task[3]) return d elif not any(map(_maybe_complex, task[1:])): return {'function': dumps_function(task[0]), 'args': warn_dumps(task[1:])} return to_serialize(task) _warn_dumps_warned = [False] def warn_dumps(obj, dumps=pickle.dumps, limit=1e6): """ Dump an object to bytes, warn if those bytes are large """ b = dumps(obj) if not _warn_dumps_warned[0] and len(b) > limit: _warn_dumps_warned[0] = True s = str(obj) if len(s) > 70: s = s[:50] + ' ... ' + s[-15:] warnings.warn("Large object of size %s detected in task graph: \n" " %s\n" "Consider scattering large objects ahead of time\n" "with client.scatter to reduce scheduler burden and \n" "keep data on workers\n\n" " future = client.submit(func, big_data) # bad\n\n" " big_future = client.scatter(big_data) # good\n" " future = client.submit(func, big_future) # good" % (format_bytes(len(b)), s)) return b def apply_function(function, args, kwargs, execution_state, key, active_threads, active_threads_lock, time_delay): """ Run a function, collect information Returns ------- msg: dictionary with status, result/error, timings, etc.. """ ident = get_thread_identity() with active_threads_lock: active_threads[ident] = key thread_state.start_time = time() thread_state.execution_state = execution_state thread_state.key = key start = time() try: result = function(*args, **kwargs) except Exception as e: msg = error_message(e) msg['op'] = 'task-erred' msg['actual-exception'] = e else: msg = {'op': 'task-finished', 'status': 'OK', 'result': result, 'nbytes': sizeof(result), 'type': type(result) if result is not None else None} finally: end = time() msg['start'] = start + time_delay msg['stop'] = end + time_delay msg['thread'] = ident with active_threads_lock: del active_threads[ident] return msg def apply_function_actor(function, args, kwargs, execution_state, key, active_threads, active_threads_lock): """ Run a function, collect information Returns ------- msg: dictionary with status, result/error, timings, etc.. """ ident = get_thread_identity() with active_threads_lock: active_threads[ident] = key thread_state.execution_state = execution_state thread_state.key = key result = function(*args, **kwargs) with active_threads_lock: del active_threads[ident] return result def get_msg_safe_str(msg): """ Make a worker msg, which contains args and kwargs, safe to cast to str: allowing for some arguments to raise exceptions during conversion and ignoring them. """ class Repr(object): def __init__(self, f, val): self._f = f self._val = val def __repr__(self): return self._f(self._val) msg = msg.copy() if "args" in msg: msg["args"] = Repr(convert_args_to_str, msg["args"]) if "kwargs" in msg: msg["kwargs"] = Repr(convert_kwargs_to_str, msg["kwargs"]) return msg def convert_args_to_str(args, max_len=None): """ Convert args to a string, allowing for some arguments to raise exceptions during conversion and ignoring them. """ length = 0 strs = ["" for i in range(len(args))] for i, arg in enumerate(args): try: sarg = repr(arg) except Exception: sarg = "< could not convert arg to str >" strs[i] = sarg length += len(sarg) + 2 if max_len is not None and length > max_len: return "({}".format(", ".join(strs[:i + 1]))[:max_len] else: return "({})".format(", ".join(strs)) def convert_kwargs_to_str(kwargs, max_len=None): """ Convert kwargs to a string, allowing for some arguments to raise exceptions during conversion and ignoring them. """ length = 0 strs = ["" for i in range(len(kwargs))] for i, (argname, arg) in enumerate(kwargs.items()): try: sarg = repr(arg) except Exception: sarg = "< could not convert arg to str >" skwarg = repr(argname) + ": " + sarg strs[i] = skwarg length += len(skwarg) + 2 if max_len is not None and length > max_len: return "{{{}".format(", ".join(strs[:i + 1]))[:max_len] else: return "{{{}}}".format(", ".join(strs)) def weight(k, v): return sizeof(v) @gen.coroutine def run(server, comm, function, args=(), kwargs={}, is_coro=None, wait=True): function = pickle.loads(function) if is_coro is None: is_coro = iscoroutinefunction(function) else: warnings.warn("The is_coro= parameter is deprecated. " "We now automatically detect coroutines/async functions") assert wait or is_coro, "Combination not supported" if args: args = pickle.loads(args) if kwargs: kwargs = pickle.loads(kwargs) if has_arg(function, 'dask_worker'): kwargs['dask_worker'] = server if has_arg(function, 'dask_scheduler'): kwargs['dask_scheduler'] = server logger.info("Run out-of-band function %r", funcname(function)) try: if not is_coro: result = function(*args, **kwargs) else: if wait: result = yield function(*args, **kwargs) else: server.loop.add_callback(function, *args, **kwargs) result = None except Exception as e: logger.warning(" Run Failed\n" "Function: %s\n" "args: %s\n" "kwargs: %s\n", str(funcname(function))[:1000], convert_args_to_str(args, max_len=1000), convert_kwargs_to_str(kwargs, max_len=1000), exc_info=True) response = error_message(e) else: response = { 'status': 'OK', 'result': to_serialize(result), } raise Return(response)