Threading and multiprocessing

There are two modules in the Python standard library that support threading and multiprocessing:

  • threading module

  • multiprocessing module

Lean on these modules as much as possible because parallel programming can be hard but is also a pretty solved problem for most use cases.

If your problem can be solved sequentially, consider the costs and benefits before going parallel.

Parallelization Strategy for Performance

  1. Break problem down into chunks

  2. Execute chunks in parallel

  3. Reassemble output of chunks into result

multitasking flow diagram

Remember:

  • Not every problem can be parallelized.

  • There is an optimal number of threads for each problem in each environment, so make it tunable.

  • Working concurrently opens up synchronization issues.

  • There are many methods for synchronizing threads:
    • locks

    • queues

    • signaling/messaging mechanisms

The Mechanics: How Do You Use Threads and/or Processes

Python provides the threading and multiprocessing modules to facility concurrency. They have similar APIs so you can use them in similar ways.

Key points:

  • There is no Python thread scheduler. It is up to the host OS. Yes these are “true” threads.

  • Threads works well for I/O bound problems. You can use literally thousands of threads.

  • Limit CPU-bound processing to C extensions because these release the GIL.

  • Do not use threading for CPU bound problems. It will go slower than no threads, especially on multiple cores.

Starting threads is relatively simple, but there are many potential issues.

We already talked about shared data, this can lead to a “race condition”.

  • May produce slightly different results every run.

  • May just flake out mysteriously every once in a while.

  • May run fine when testing, but fail when run on:
    • a slower system

    • a heavily loaded system

    • a larger dataset

  • Thus you must synchronize threads!

Example: A CPU bound problem

Numerically integrate the function \(y =x^2\) from 0 to 10.

../_images/x2.png

Solution

Parallel Execution Example

Consider the following code in: integrate.py

def f(x):
    return x**2

def integrate(f, a, b, N):
    s = 0
    dx = (b-a)/N
    for i in range(N):
        s += f(a+i*dx)
    return s * dx

We can do better than this.

Break down the problem into chunks that can be parallelized and then add the results together.

The Threading Module

Starting threads doesn’t take much:

import sys
import threading
import time

def func():
    for i in range(5):
        print("hello from thread %s" % threading.current_thread().name)
        time.sleep(1)

threads = []
for i in range(3):
    thread = threading.Thread(target=func, args=())
    thread.start()
    threads.append(thread)
  • The process will exit when the last non-daemon thread exits.

  • A thread can be specified as a daemon thread by setting its daemon attribute: thread.daemon = True.

  • Daemon threads get cut off at program exit, without any opportunity for cleanup. But you don’t have to track and manage them. These are useful for things like garbage collection or network keepalives.

  • You can block and wait for a thread to exit with thread.join().

Subclassing Thread

You can add threading capability to your own classes by subclassing Thread and implementing the run method.

import threading

class MyThread(threading.Thread):

    def run(self):
        print("hello from %s" % threading.current_thread().name)

thread = MyThread()
thread.start()

Race Conditions

In the last example we saw threads competing for access to stdout.

Worse, if competing threads try to update the same value, we might get an unexpected race condition.

Race conditions occur when multiple statements need to execute atomically, but get interrupted midway.

race_condition.py

No Race Condition

Thread 1

Thread 2

Integer value

0

read value

0

increase value

0

write back

1

read value

1

increase value

1

write back

2

Race Condition!

Thread 1

Thread 2

Integer value

0

read value

0

read value

0

increase value

0

increase value

0

write back

1

write back

1

See: http://en.wikipedia.org/wiki/Race_condition

Deadlocks

Synchronization and Critical Sections are used to control race conditions.

But they introduce other potential problems, like deadlocks.

“A deadlock is a situation in which two or more competing actions are each waiting for the other to finish, and thus neither ever does.”

When two trains approach each other at a crossing, both shall come to a full stop and neither shall start up again until the other has gone

See also Livelock:

Two people meet in a narrow corridor, and each tries to be polite by moving aside to let the other pass, but they end up swaying from side to side without making any progress because they both repeatedly move the same way at the same time.

Locks

Lock objects allow threads to control access to a resource until they’re done with it.

This is known as mutual exclusion, often called “mutex”.

A lock has two states: locked and unlocked

If multiple threads have access to the same lock, they can police themselves by calling its .acquire() and .release() methods.

If a lock is locked, .acquire() will block until it becomes unlocked.

These threads will wait in line politely for access to the statements in f().

Mutex Locks (threading.Lock)

  • Probably most common

  • Only one thread can modify shared data at any given time

  • Thread determines when unlocked

  • Must put lock/unlock around critical code in ALL threads

  • Difficult to manage

Easiest with context manager:

x = 0
x_lock = threading.Lock()

# Example critical section
with x_lock:
    ...
    # statements using x

Only use one lock per thread or you risk mysterious deadlocks that are hard to debug.

Or, you can use RLock for code-based locking.

import threading
import time

lock = threading.Lock()

def f():
    lock.acquire()
    print("%s got lock" % threading.current_thread().name)
    time.sleep(1)
    lock.release()

threading.Thread(target=f).start()
threading.Thread(target=f).start()
threading.Thread(target=f).start()

Non-blocking Locking

.acquire() will return True if it successfully acquires a lock.

Its first argument is a boolean which specifies whether a lock should block or not. The default is True.

import threading

lock = threading.Lock()
lock.acquire()

if not lock.acquire(False):
    print("couldn't get lock")

lock.release()

if lock.acquire(False):
    print("got lock")

threading.RLock - Reentrant Lock

Useful for recursive algorithms. A thread-specific count of the locks is maintained. A reentrant lock can be acquired multiple times by the same thread.

However, .release() must be called the same number of times as .acquire() by that thread.

threading.Semaphore

This is like an RLock, but in reverse.

A Semaphore is given an initial counter value, defaulting to 1.

Each call to .acquire() decrements the counter, .release() increments it.

If .acquire() is called on a Semaphore with a counter of 0, it will block until the Semaphore counter is greater than 0.

This is useful for controlling the maximum number of threads allowed to access a resource simultaneously.

See: http://en.wikipedia.org/wiki/Semaphore_(programming)

Events (threading.Event)

  • Threads can wait for particular event

  • Setting an event unblocks all waiting threads

These are commonly used to create barriers to access or to have one thread notify another thread that something has happened.

Condition (threading.Condition)

  • Combination of locking/signaling

  • The lock protects code that establishes a “condition” (e.g., data available)

  • The signal notifies threads that “condition” has changed

These are commonly used in producer/consumer patterns.

Locking Exercise

stdout_writer.py

Multiple threads in the script write to stdout, and their output gets jumbled

  1. Add a locking mechanism to give each thread exclusive access to stdout.

  2. Try adding a Semaphore to allow two threads access at once.

Managing Thread Results

We need a thread safe way of storing results from multiple threads of execution. That is provided by the Queue module.

Queues allow multiple producers and multiple consumers to exchange data safely.

Size of the queue is managed with the maxsize argument.

It will block consumers if empty and block producers if full.

If maxsize is less than or equal to zero, then the queue size is infinite.

from Queue import Queue

q = Queue(maxsize=10)
q.put(37337)
block = True
timeout = 2
print(q.get(block, timeout))

See:

Queues (queue)

  • Easier to use than many of above

  • Do not need locks

  • Has signaling

Common use: producer/consumer patterns

from Queue import Queue
data_q = Queue()

# producer thread:
for item in produce_items():
    data_q.put(item)

# consumer thread:
while True:
    item = q.get()
    consume_item(item)

Scheduling (sched)

  • Schedules based on time, either absolute or delay

  • Low level, so it has many of the traps of the threading synchronization primitives

Timed Events (threading.timer)

Run a function at some time in the future:

import threading

def called_once():
    """
    this function is designed to be called once in the future
    """
    print("I just got called! It's now: {}".format(time.asctime()))

# setting it up to be called
t = Timer(interval=3, function=called_once)
t.start()

# you can cancel it if you want:
t.cancel()

simple_timer.py

Other Queue types

Queue.LifoQueue

  • Last In, First Out

Queue.PriorityQueue

  • Lowest valued entries are retrieved first

One pattern for PriorityQueue is to insert entries of form data by inserting the tuple:

(priority_number, data)

Threading Example with a Queue

integrate_main.py

#!/usr/bin/env python

import threading
import queue

# from integrate.integrate import integrate, f
from integrate import f, integrate_numpy as integrate
from decorators import timer

@timer
def threading_integrate(f, a, b, N, thread_count=2):
    """break work into N chunks"""
    N_chunk = int(float(N) / thread_count)
    dx = float(b - a) / thread_count

    results = queue.Queue()

    def worker(*args):
        results.put(integrate(*args))

    for i in range(thread_count):
        x0 = dx * i
        x1 = x0 + dx
        thread = threading.Thread(target=worker, args=(f, x0, x1, N_chunk))
        thread.start()
        print("Thread %s started" % thread.name)

    return sum((results.get() for i in range(thread_count)))


if __name__ == "__main__":

    # parameters of the integration
    a = 0.0
    b = 10.0
    N = 10**8
    thread_count = 8

    print("Numerical solution with N=%(N)d : %(x)f" %
          {'N': N, 'x': threading_integrate(f, a, b, N, thread_count=thread_count)})

Threading on a CPU Bound Problem

Try running the code in integrate_threads.py

It has a couple of tunable parameters:

a = 0.0  # the start of the integration
b = 10.0  # the end point of the integration
N = 10**8 # the number of steps to use in the integration
thread_count = 8  # the number of threads to use

What happens when you change the thread count? What thread count gives the maximum speed?

Multiprocessing

  • Processes are completely isolated

  • No locking and no GIL!

  • But, instead of locking we have to solve the messaging problem

multiprocessing provides an API very similar to threading, so the transition is easy.

Use multiprocessing.Process instead of threading.Thread.

import multiprocessing
import os
import time

def func():
    print(f"hello from process {os.getpid()}")
    time.sleep(1)

proc = multiprocessing.Process(target=func, args=())
proc.start()
proc = multiprocessing.Process(target=func, args=())
proc.start()

Differences with Threading

Multiprocessing has its own multiprocessing.Queue which handles inter-process communication.

It also has its own versions of Lock, RLock, Semaphore.

from multiprocessing import Queue, Lock

You can use multiprocessing.Pipe for 2-way process communication:

from multiprocessing import Pipe
parent_conn, child_conn = Pipe()
child_conn.send("foo")
print(parent_conn.recv())

Messaging

Pipes (multiprocessing.Pipe)

  • Returns a pair of connected objects

  • Largely mimics Unix pipes, but higher level

  • Send pickled objects or buffers

Queues (multiprocessing.Queue)

  • Same interface as queue.Queue

  • Implemented on top of pipes

  • Means you can pretty easily port threaded programs using queues to multiprocessing
    • Queue is the only shared data

    • Data is all pickled and unpickled to pass between processes which adds significant overhead

Other Features of the multiprocessing Package

  • Pools

  • Shared objects and arrays

  • Synchronization primitives

  • Managed objects

  • Connections

Pooling

A processing pool contains worker processes with only a configured number running at one time.

from multiprocessing import Pool
pool = Pool(processes=4)

The Pool module has several methods for adding jobs to the pool

apply_async(func[, args[, kwargs[, callback]]])

map_async(func, iterable[, chunksize[, callback]])

Pooling example

from multiprocessing import Pool

def f(x):
    return x*x

if __name__ == '__main__':
    pool = Pool(processes=4)

    result = pool.apply_async(f, (10,))
    print(result.get(timeout=1))
    print(pool.map(f, range(10)))

    it = pool.imap(f, range(10))
    print(it.next())
    print(it.next())
    print(it.next(timeout=1))

    import time
    result = pool.apply_async(time.sleep, (10,))
    print(result.get(timeout=1))

See: http://docs.python.org/3/library/multiprocessing.html#module-multiprocessing.pool

ThreadPool

Threading also has a pool. Confusingly, it lives in the multiprocessing module.

from multiprocessing.pool import ThreadPool
pool = ThreadPool(processes=4)

Otherwise you can use it the same way.