Bytewax is a Python framework that simplifies event and stream processing. Because Bytewax couples the stream and event processing capabilities of Flink, Spark, and Kafka Streams with the friendly and familiar interface of Python, you can re-use the Python libraries you already know and love. Connect data sources, run stateful transformations and write to various different downstream systems with built-in connectors or existing Python libraries.
Bytewax is a Python framework and Rust distributed processing engine that uses a dataflow computational model to provide parallelizable stream processing and event processing capabilities similar to Flink, Spark, and Kafka Streams. You can use Bytewax for a variety of workloads from moving data à la Kafka Connect style all the way to advanced online machine learning workloads. Bytewax is not limited to streaming applications but excels anywhere that data can be distributed at the input and output.
Bytewax has an accompanying command line interface, waxctl, which supports the deployment of dataflows on cloud vms or kuberentes. You can download it here.
pip install bytewax
A Bytewax dataflow is Python code that will represent an input, a series of processing steps, and an output. The inputs could range from a Kafka stream to a WebSocket and the outputs could vary from a data lake to a key-value store.
from bytewax.dataflow import Dataflow
from bytewax.connectors.kafka import KafkaInput
# Bytewax has input and output helpers for common input and output data sources
# but you can also create your own with the ManualOutputConfig.
At a high-level, the dataflow compute model is one in which a program execution is conceptualized as data flowing through a series of operator-based steps. Operators like map
and filter
are the processing primitives of Bytewax. Each of them gives you a “shape” of data transformation, and you give them regular Python functions to customize them to a specific task you need. See the documentation for a list of the available operators
import json
def deserialize(key_bytes__payload_bytes):
_, payload_bytes = key_bytes__payload_bytes
event_data = json.loads(payload_bytes) if payload_bytes else None
return event_data["user_id"], event_data
def anonymize_email(user_id__event_data):
user_id, event_data = user_id__event_data
event_data["email"] = "@".join(["******", event_data["email"].split("@")[-1]])
return user_id, event_data
def remove_bytewax(user_id__event_data):
user_id, event_data = user_id__event_data
return "bytewax" not in event_data["email"]
flow = Dataflow()
flow.input("inp", KafkaInput(brokers=["localhost:9092"], topic="web_events"))
flow.map(deserialize)
flow.map(anonymize_email)
flow.filter(remove_bytewax)
Bytewax is a stateful stream processing framework, which means that some operations remember information across multiple events. Windows and aggregations are also stateful, and can be reconstructed in the event of failure. Bytewax can be configured with different state recovery mechanisms to durably persist state in order to recover from failure.
There are multiple stateful operators available like reduce
, stateful_map
and fold_window
. The complete list can be found in the API documentation for all operators. Below we use the fold_window
operator with a tumbling window based on system time to gather events and calculate the number of times events have occurred on a per-user basis.
from datetime import datetime, timedelta, timezone
from collections import defaultdict
from bytewax.window import TumblingWindow, SystemClockConfig
def build():
return defaultdict(lambda: 0)
def count_events(results, event):
results[event["type"]] += 1
return results
cc = SystemClockConfig()
align_to = datetime(2023, 1, 1, tzinfo=timezone.utc)
wc = TumblingWindow(length=timedelta(seconds=5), align_to=align_to)
flow.fold_window("session_state_recovery", cc, wc, build, count_events)
Output mechanisms in Bytewax are managed in the output operator. There are a number of helpers that allow you to easily connect and write to other systems (output docs). If there isn’t a helper built, it is easy to build a custom version, which we will do below. Similar the input, Bytewax output can be parallelized and the client connection will occur on the worker.
import json
import psycopg2
from bytewax.outputs import PartitionedOutput, StatefulSink
class PsqlSink(StatefulSink):
def __init__(self):
self.conn = psycopg2.connect("dbname=website user=bytewax")
self.conn.set_session(autocommit=True)
self.cur = self.conn.cursor()
def write(self, user_id__user_data):
user_id, user_data = user_id__user_data
query_string = """
INSERT INTO events (user_id, data)
VALUES (%s, %s)
ON CONFLICT (user_id)
DO UPDATE SET data = %s;
"""
self.cur.execute(
query_string, (user_id, json.dumps(user_data), json.dumps(user_data))
)
def snapshot(self):
pass
def close(self):
self.conn.close()
class PsqlOutput(PartitionedOutput):
def list_parts(self):
return {"single"}
def assign_part(self, item_key):
return "single"
def build_part(for_part, resume_state):
return PsqlSink()
flow.output("out", PsqlOutput())
Bytewax dataflows can be executed on a single host with multiple Python processes, or on multiple hosts. When processing data in a distributed fashion, Bytewax will ensure that all items with the same key are routed to the same host.
python -m bytewax.run my_dataflow:flow
It can also be run in a Docker container as described further in the documentation.
The recommended way to run dataflows at scale is to leverage the kubernetes ecosystem. To help manage deployment, we built waxctl, which allows you to easily deploy dataflows that will run at huge scale across multiple compute nodes.
waxctl df deploy my_dataflow.py --name my-dataflow
At a high level, Bytewax provides a few major benefits:
- The operators in Bytewax are largely “data-parallel”, meaning they can operate on independent parts of the data concurrently.
- Bytewax offers the ability to express higher-level control constructs, like iteration.
- Bytewax allows you to develop and run your code locally, and then easily scale that code to multiple workers or processes without changes.
- Bytewax can be used in both a streaming and batch context
- Ability to leverage the Python ecosystem directly
Slack Is the main forum for communication and discussion.
GitHub Issues is reserved only for actual issues. Please use the slack community for discussions.
Install the latest release with pip:
pip install bytewax
To build a specific branch, you will need to use Maturin and have Rust installed on your machine. Once those have been installed run
maturin develop -E dev
Important: If you are testing with a maturin built version from source, you should use maturin build --release
since maturin develop
will be slower.
For a more complete example, and documentation on the available operators, check out the User Guide and the /examples folder.
Bytewax is licensed under the Apache-2.0 license.
Contributions are welcome! This community and project would not be what it is without the contributors. All contributions, from bug reports to new features, are welcome and encouraged. Please view the contribution guidelines before getting started.
With ❤️ Bytewax