Year-End Discount: 10% OFF 1-year and 20% OFF 2-year subscriptions!

Home/Blog/An Introduction to Scaling Distributed Python Applications

An Introduction to Scaling Distributed Python Applications

Feb 15, 2021 - 7 min read
Amanda Fawcett

Python is often dismissed when it comes to building scalable, distributed applications. The trick is knowing the right implementation and tools for writing Python distributed applications that scale horizontally.

With the right methods, technologies, and practices, you can make Python applications fast and able to grow in order to handle more work or requirements.

In this tutorial, we will introduce you to scaling in Python. We’ll learn the key things you need to know when building a scalable, distributed system in Python.

This guide at a glance:

Learn how to scale in Python

Learn how write Python applications that scale horizontally. You’ll cover everything from REST APIs, deployment to PaaS, and functional programming.

The Hacker’s Guide to Scaling in Python

What is scaling?

Scalability is a somewhat vague term. A scalable system is able to grow to accommodate required growth, changes, or requirements. Scaling refers to the methods, technologies, and practices that allow an app to grow.

A key part of scaling is building distributed systems. This means that you distribute workload across multiple workers and with multiple processing units. Workers divide tasks across multiple processors or computers.

Spreading workload over multiple hosts makes it possible to achieve horizontal scalability, which is the ability to add more nodes. It also helps with fault tolerance. If a node fails, another can pick up the traffic.

Before we look at the methods of building scalable systems in Python, let’s go over the fundamental properties distributed systems.


Single-threaded application

This is a type of system that implies no distribution. This is the simplest kind of application. However, they are limited by the power of using a single processor.

Multi-threaded application

Most computers are equipped with this type of system. Multi-threading applications are more error-prone, but they offer few failure scenarios, as no network is involved.

Network distributed application

This type of system is for applications that need to scale significantly. They are the most complicated applications to write, as they require a network.


Scaling across processors is done with multithreading. This means we are running code in parallel with threads, which are contained in a single process. Code will run in parallel only if there is more than one CPU available. Multithreading involves many traps and issues, such as Python’s Global Interpreter Lock (GIL).

CPU scaling in Python

Using multiple CPUs is one of the best options for scalability in Python. To do so, we must use concurrency and parallelism, which can be tricky to implement properly. Python offers two options for spreading your workload across multiple local CPUs: threads and processes.

Threads in Python

Threads are a good way to run a function concurrently. If there are multiple CPUs available, threads can be scheduled on multiple processing units. Scheduling is determined by the operating system.

There is only one thread, the main, by default. This is the thread that runs your Python application. To start another thread, Python offers a threading module.

import threading
def print_something(something):
t = threading.Thread(target=print_something, args=("hello",))
print("thread started")

Once started, the main thread waits for the second thread to complete by calling its join method. But, if you do not join all your threads, it is possible that the main thread finishes before the other threads can join, and your program will appear to be blocked.

To prevent this, you can configure your threads as daemons. When a thread is a daemon, it is like a background thread and will be terminated once the main thread exits. Note that we don’t need to use the join method.

import threading
def print_something(something):
t = threading.Thread(target=print_something, args=("hello",))
t.daemon = True
print("thread started")

Processes in Python

Multithreading is not perfect for scalability due to the Global Interpreter Lock (GIL). We can also use processes instead of threads as an alternative. The multiprocessing package is a good, high-level option for processes. It provides an interface that starts new processes. Each process is a new, independent instance, so each process has its own independent global state.

import random
import multiprocessing
def compute(results):
[random.randint(1, 100) for i in range(1000000)]))
if __name__ == "__main__":
with multiprocessing.Manager() as manager:
results = manager.list()
workers = [multiprocessing.Process(target=compute, args=(results,))
for x in range(8)]
for worker in workers:
for worker in workers:
print("Results: %s" % results)

When work can be parallelized for a certain amount of time, it’s better to use multiprocessing and fork jobs. This spreads the workload among several CPU cores.

We can also use multiprocessing.Pool, which is a multiprocessing library that provides a pool mechanism. With multiprocessing.Pool, we don’t need to manage the processes manually. It also make processes reusable.

import multiprocessing
import random
def compute(n):
return sum(
[random.randint(1, 100) for i in range(1000000)])
if __name__ == "__main__":
# Start 8 workers
pool = multiprocessing.Pool(processes=8)
print("Results: %s" %, range(8)))

Keep the learning going.

Learn how to scale in Python without scrubbing through videos or documentation. Educative’s text-based courses are easy to skim and feature live coding environments - making learning quick and efficient.

The Hacker’s Guide to Scaling Python

Daemon Processes in Python

As we learned, using multiple processes to schedule jobs is more efficient in Python. Another good option is using daemons, which are long-running, background processes that are responsible for scheduling tasks regularly or processing jobs from a queue.

We can use cotyledon, a Python library for building long-running processes. It can be leveraged to build long-running, background, job workers.

Below, we create a class named PrinterService to implement the method for cotyledon.Service: run. This contains the main loop and terminate. This library does most of its work behind scenes, such os.fork calls and setting up modes for daemons.

Cotyledon uses several threads internally. This is why the threading.Event object is used to synchronize our run and terminate methods.

import threading
import time
import cotyledon
class PrinterService(cotyledon.Service):
name = "printer"
def __init__(self, worker_id):
super(PrinterService, self).__init__(worker_id)
self._shutdown = threading.Event()
def run(self):
while not self._shutdown.is_set():
print("Doing stuff")
def terminate(self):
# Create a manager
manager = cotyledon.ServiceManager()
# Add 2 PrinterService to run
manager.add(PrinterService, 2)
# Run all of that

Cotyledon runs a master process that is responsible for handling all its children. It then starts the two instances of PrinterService, and gives new process names so they’re easy to track. With Cotyledon, if one of the processes crashes, it is automatically relaunched.

Note: Cotyledon also offers features for reloading a program configuration or dynamically changing the number of workers for a class.

Event loops and Asyncio in Python

An event loop is a type of control flow for a program where messages are pushed into a queue. The queue is then consumed by the event loop, dispatching them to appropriate functions.


A very simple event loop could like this in Python:

while True: message = get_message() if message == quit: break

Asyncio is a new, state-of-the-art event loop provided in Python 3. Asyncio ) stands for asynchronous input output. It refers to a programming paradigm that achieves high concurrency using a single thread or event loop. This is a good alternative to multithreading for the following reasons:

  • It’s difficult to write code that is thread safe. With asynchronous code, you know exactly where the code will shift between tasks.
  • Threads consume a lot of data. With async code, all the code shares the same small stack and the stack.
  • Threads are OS structures so they require more memory. This is not the case for ssynico.

Asyncio is based on the concept of event loops. When asyncio creates an event loop, the application registers the functions to call back when a specific event happens. This is a type of function called a coroutine. It works similar to a generator, as it gives back the control to a caller with a yield statement.

import asyncio
async def hello_world():
print("hello world!")
return 42
hello_world_coroutine = hello_world()
event_loop = asyncio.get_event_loop()
print("entering event loop")
result = event_loop.run_until_complete(hello_world_coroutine)

Above, the coroutine hello_world is defined as a function, but that the keyword used to start its definition is async def. This coroutine will print a message and returns a result. The event loop runs the coroutine and is terminated as when the coroutine returns.

Next steps for your learning

Congrats on making it to the end! You should now have a good introduction to the tools we can use to scale in Python. We can leverage these tools to build distributed systems effectively. But there is still more to learn. Next, you’ll want to learn about:

  • Run coroutine cooperatively
  • aiohttp library
  • Queue-based distribtuion
  • Lock management
  • Deploying on PaaS

To get started with these concepts, check out Educative’s comprehensive course The Hacker’s Guide to Scaling Python. You’ll cover everything from concurrency to queue-based distribution, lock management, and group memberships. At the end, you’ll get hands on building a REST API in Python and deploying an app to a PaaS.

By the end, you’ll be more productive with Python, and you’ll be able to write distributed applications.

Happy learning!

Continue reading about Python

WRITTEN BYAmanda Fawcett

Join a community of more than 1.6 million readers. A free, bi-monthly email with a roundup of Educative's top articles and coding tips.