# Best Practices¶

It is easy to get started with Dask delayed, but using it well does require some experience. This page contains suggestions for best practices, and includes solutions to common problems:

## Call delayed on the function, not the result¶

Dask.delayed operates on functions, like dask.delayed(f)(x, y), not on their results like dask.delayed(f(x, y)). When you do the latter Python first calculates f(x, y) before Dask has a chance to step in

Don’t

dask.delayed(f(x, y))


Do

dask.delayed(f)(x, y)


## Compute on lots of computation at once¶

To improve parallelism you want to include lots of computation in each compute call. Ideally you want to make many dask.delayed calls to define your computation and then only call dask.compute at the end. It’s ok to call dask.compute in the middle of your computation as well, but everything will stop there as Dask computes those results before moving forward with your code.

Don’t

for x in L:
y.compute()  # calling compute after every delayed call stops parallelism


Do

results = []
for x in L:
results.append(y)

results = dask.compute(*results)  # call compute after you have collected many delayed calls


## Don’t mutate inputs¶

Your functions should not change the inputs directly

Don’t

@dask.delayed
def f(x):
x += 1
return x


Do

@dask.delayed
def f(x):
return x + 1


If you need to use a mutable operation then make a copy within your function first

@dask.delayed
def f(x):
x = copy(x)
x += 1
return x


## Avoid global state¶

Ideally your operations shouldn’t rely on global state. Using global state might work if you only use threads, but when you move to multiprocessing or distributed computing then you will likely encounter confusing errors

Don’t

L = []

def f(x):
L.append(x)


## Don’t rely on side effects¶

Delayed functions only do something if they are computed. You will always need to pass the output to something that eventually calls compute.

Don’t

dask.delayed(f)(1, 2, 3)  # this has no effect


Do

x = dask.delayed(f)(1, 2, 3)
...
dask.compute(x, ...)  # need to call compute for something to happen


## Break up computations into many pieces¶

Every dask.delayed function call is a single operation from Dask’s perspective. You achieve parallelism by having many dask.delayed calls, not by using only a single one. Dask will not look inside a function decorated with dask.delayed and parallelize that code internally. It needs your help to find good places to break up a computation.

Don’t

def load(filename):
...

def process(data):
...

def save(data):
...

def f(filenames):
results = []
for filename in filenames:
data = process(data)
results.append(save(data))

return results



Do

@dask.delayed
...

def process(data):
...

def save(data):
...

def f(filenames):
results = []
for filename in filenames:
data = process(data)
results.append(save(data))

return results



Every delayed task has an overhead of a few hundred microseconds. Usually this is ok, but it can become a problem if you apply dask.delayed too finely. In this case it’s often best to break up your many tasks into batches, or use one of the dask collections to help you.

Don’t

results = []
for x in range(1000000000):  # Too many dask.delayed calls
results.append(y)


Do

# Use collections

b = db.from_sequence(1000000000, npartitions=1000)
b = b.map(f)

# Or batch manually

def batch(seq):
sub_results = []
for x in seq:
sub_results.append(f(x))
return sub_results

batches = []
for i in range(0, 1000000000, 1000000):  # in steps of 1000000
result_batch = dask.delayed(batch, range(i, i + 1000000))
batches.append(result_batch)


## Avoid calling delayed within delayed functions¶

Often if you are new to using Dask.delayed you place dask.delayed calls everywhere and hope for the best. While this may actually work it’s usually slow and results in hard-to-understand solutions.

Don’t

@dask.delayed
def process_all(L):
result = []
for x in L:
result.append(y)
return result


Do

Instead, because this function only does delayed work it is very fast and so there is no reason to delay it.

def process_all(L):
result = []
for x in L:
result.append(y)
return result


When you place a dask array or dask dataframe into a delayed call that function will receive the Numpy or Pandas equivalent. Beware that if your array is large then this might crash your workers.

Instead, it’s more common to use methods like da.map_blocks or df.map_partitions, or to turn your arrays or dataframes into many delayed objects

Don’t

import dask.dataframe as dd



Do

import dask.dataframe as dd

df.map_partitions(train)
# or
partitions = df.to_delayed()

delayed_values = [dask.delayed(train)(part) for part in partitions]


However, if you don’t mind turning your dask array/dataframe into a single chunk then this is ok.

dask.delayed(train)(..., y=df.sum())


## Avoid repeatedly putting large inputs into delayed calls¶

Every time you pass a concrete result (anything that isn’t delayed) Dask will hash it by default to give it a name. This is fairly fast (around 500 MB/s) but can be slow if you do it over and over again. Instead, it is better to delay your data as well.

This is especially important when using a distributed cluster to avoid sending your data separately for each function call.

Don’t

x = np.array(...)  # some large array

results = [dask.delayed(train)(x, i) for i in range(1000)]


Every call to dask.delayed(train)(x, ...) has to hash the numpy array x, which slows things down.

Do

x = np.array(...)  # some large array
x = dask.delayed(x)  # delay the data, hashing once

results = [dask.delayed(train)(x, i) for i in range(1000)]