Coursera Learner working on a presentation with Coursera logo and
Coursera Learner working on a presentation with Coursera logo and

Dask may be a flexible library for parallel computing in Python.

Dask consists of two parts:

Dynamic task scheduling optimized for computation. this is often almost like Airflow, Luigi, Celery, or Make, but optimized for interactive computational workloads.

Big Data” collections like parallel arrays, dataframes, and lists that reach common interfaces like NumPy, Pandas, or Python iterators to larger-than-memory or distributed environments. These parallel collections run on top of dynamic task schedulers.

Dask emphasizes the subsequent virtues:

Familiar: Provides parallelized NumPy array and Pandas DataFrame objects

Flexible: Provides a task scheduling interface for more custom workloads and integration with other projects.

Native: Enables distributed computing in pure Python with access to the PyData stack.

Fast: Operates with low overhead, low latency, and minimal serialization necessary for fast numerical algorithms

Scales up: Runs resiliently on clusters with 1000s of cores

Scales down: Trivial to line up and run on a laptop during a single process

Responsive: Designed with interactive computing in mind, it provides rapid feedback and diagnostics to assist humans

Dask collections and schedulers

import pandas as pd                     import dask.dataframe as dd

df = pd.read_csv(‘2015-01-01.csv’)      df = dd.read_csv(‘2015-*-*.csv’)

df.groupby(df.user_id).value.mean()     df.groupby(df.user_id).value.mean().compute()

Dask Array mimics NumPy – documentation

import numpy as np                       import dask.array as da

f = h5py.File(‘myfile.hdf5’)             f = h5py.File(‘myfile.hdf5’)

x = np.array(f[‘/small-data’])           x = da.from_array(f[‘/big-data’],

                                                           chunks=(1000, 1000))

x – x.mean(axis=1)                       x – x.mean(axis=1).compute()

Dask Bag mimics iterators, Toolz, and PySpark – documentation

import dask.bag as db

b = db.read_text(‘2015-*-*.json.gz’).map(json.loads)

b.pluck(‘name’).frequencies().topk(10, lambda pair: pair[1]).compute()

Dask Delayed mimics for loops and wraps custom code – documentation

from dask import delayed

L = []

for fn in filenames:                  # Use for loops to build up computation

    data = delayed(load)(fn)          # Delay execution of function

    L.append(delayed(process)(data))  # Build connections between variables

result = delayed(summarize)(L)


The concurrent.futures interface provides general submission of custom tasks: – documentation

from dask.distributed import Client

client = Client(‘scheduler:port’)

futures = []

for fn in filenames:

    future = client.submit(load, fn)


summary = client.submit(summarize, futures)


Scales from laptops to clusters

Dask is convenient on a laptop. It installs trivially with conda or pip and extends the dimensions of convenient datasets from “fits in memory” to “fits on disk”.

Dask can scale to a cluster of 100s of machines. it’s resilient, elastic, data local, and low latency. For more information, see the documentation about the distributed scheduler.

This simple transition between single-machine to moderate cluster enables users to both start simple and grow when necessary.

Complex Algorithms

Dask represents parallel computations with task graphs. These directed acyclic graphs may have arbitrary structure, which enables both developers and users the liberty to create sophisticated algorithms and to handle messy situations not easily managed by the map/filter/groupby paradigm common in most data engineering frameworks.

We originally needed this complexity to create complex algorithms for n-dimensional arrays but have found it to be equally valuable when handling messy situations in everyday problems.


Weekly newsletter

No spam. Just the latest releases and tips, interesting articles, and exclusive interviews in your inbox every week.