This is a demostration of the Dataflow, examples are under examples
folder
- Data streaming: Send data to Pub/Sub, then process them and import records to BigQuery
- Windowed aggregation: Aggregate purchase value by users from Pub/Sub event stream
- This doc from Google is a useful reference too https://cloud.google.com/pubsub/docs/stream-messages-dataflow
- (Optional, for data_streaming only) Create a BQ table with definition stated in
bq_table_def.json
- Spin up the Dataflow job
- Publish message to the designated Pub/Sub topic
- If all is good, output requirements.txt, then deploy to GCP with
./deploy.sh