Extract and load your largest data streams with a framework you can trust for billion records.
- Natively fault-tolerant: Bizon uses a checkpointing mechanism to keep track of the progress and recover from the last checkpoint.
- High throughput: Bizon is designed to handle high throughput and can process billions of records.
- Queue system agnostic: Bizon is agnostic of the queuing system, you can use any queuing system among Python Queue, RabbitMQ, Kafka or Redpanda. Thanks to the
bizon.engine.queue.Queue
interface, adapters can be written for any queuing system. - Pipeline metrics: Bizon provides exhaustive pipeline metrics and implement OpenTelemetry for tracing. You can monitor:
- ETAs for completion
- Number of records processed
- Completion percentage
- Latency Source <> Destination
- Lightweight & lean: Bizon is lightweight, minimal codebase and only uses few dependencies:
requests
for HTTP requestspyyaml
for configurationsqlalchemy
for database / warehouse connectionspolars
for memory efficient data buffering and vectorized processingpyarrow
for Parquet file format
pip install bizon
bizon source list
bizon stream list <source_name>
Create a file named config.yml
in your working directory with the following content:
name: demo-creatures-pipeline
source:
name: dummy
stream: creatures
authentication:
type: api_key
params:
token: dummy_key
destination:
name: logger
config:
dummy: dummy
Run the pipeline with the following command:
bizon run config.yml
Backend is the interface used by Bizon to store its state. It can be configured in the backend
section of the configuration file. The following backends are supported:
sqlite
: In-memory SQLite database, useful for testing and development.bigquery
: Google BigQuery backend, perfect for light setup & production.postgres
: PostgreSQL backend, for production use and frequent cursor updates.
Queue is the interface used by Bizon to exchange data between Source
and Destination
. It can be configured in the queue
section of the configuration file. The following queues are supported:
python_queue
: Python Queue, useful for testing and development.rabbitmq
: RabbitMQ, for production use and high throughput.kafka
: Apache Kafka, for production use and high throughput and strong persistence.
Queue configuration can be set to python_queue
and backend configuration to sqlite
.
This will allow you to test the pipeline without any external dependencies.
To test the pipeline with Kafka, you can use docker compose
to setup Kafka or Redpanda locally.
Kafka
docker compose --file ./scripts/kafka-compose.yml up # Kafka
docker compose --file ./scripts/redpanda-compose.yml up # Redpanda
In your YAML configuration, set the queue
configuration to Kafka under engine
:
engine:
queue:
type: kafka
config:
queue:
bootstrap_server: localhost:9092 # Kafka:9092 & Redpanda: 19092
RabbitMQ
docker compose --file ./scripts/rabbitmq-compose.yml up
In your YAML configuration, set the queue
configuration to Kafka under engine
:
engine:
queue:
type: rabbitmq
config:
queue:
host: localhost
queue_name: bizon