DataBrew - Blink
OpenSource data streaming & processing engine to build event-driven systems
Blink ETL gives you the ability to create truly flexible data pipelines to implement even-driven architecture, database replication, and data lakes
Table of Contents
- Installation
- Running Blink locally
- Docs
Getting started
Let's give Blink a try. Check out different options to start Blink on your local machine
Run with Docker on Linux
Currently, we offer only Linux-based arm64 build for docker
Create a config with the name blink.yaml
and run the docker image
docker run -v ./blink.yaml:/app/blink.yaml usedatabrew/blink start
Install Golang binary
go install github.com/usedatabrew/blink@v1.16.8
Running Blink locally
To run Blink locally - you have to create a config file that will define streams
service:
pipeline_id: 1
source:
driver: playground
config:
data_type: market
publish_interval: 1
historical_batch: false
stream_schema:
- stream: market
columns:
- name: company
nativeConnectorType: String
databrewType: String
nullable: false
pk: false
- name: currency
nativeConnectorType: String
databrewType: String
nullable: false
pk: false
processors:
- driver: sql
config:
query: "select * from streams.market where currency = 'USD'"
sink:
driver: stdout
config: {}
blink start -c blink-config.yaml