Showcases for different data tech stacks
Data Lakehouse architecture with storage/compute separation and Apache Iceberg open storage format, fully provisioned using Terraform. Build using GCP-native serverless services (such as Dataproc and BigQuery/BigLake) for an event-driven, scalable ingestion.
API-callable Data Pipeline ingesting data from another API (open meeo) using Apache Beam, storing the results in a DuckDB database, deployable as container images via docker compose.