You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A production-ready PySpark project template with medallion architecture, Python packaging, unit tests, integration tests, CI/CD automation, Databricks Asset Bundles, and DQX data quality framework.
Production-grade Databricks infrastructure templates for Azure. Deploy in 20 minutes with VNet injection, Unity Catalog, managed identity. Perfect for learning and prototyping. Free and open source.
Production-ready support ticket classification using Unity Catalog AI Functions, Vector Search, and RAG. Features 6-phase workflow, knowledge base integration, and Streamlit dashboard.
databricks-dab-lab is an end-to-end lab that shows how to deploy Databricks Asset Bundles (DABs) with GitHub Actions, using Terraform to provision an Azure Databricks workspace + cluster, then deploying and running three jobs in sequence (data setup → ETL → ML training).
Real Estate ELT pipeline using Databricks Asset Bundles on GCP. Ingests, transforms, and analyzes property data via Delta Live Tables. Follows medallion architecture (Bronze/Silver/Gold), modular Python design, CI/CD automation with GitHub Actions, and full Unit and Integration tests coverage.
End-to-end Azure Data Engineering project using ADF for incremental ingestion, Databricks (DLT) for Medallion Architecture, and Delta Lake for CDC (SCD Type 1). Managed via Databricks Asset Bundles (DABs) for professional CI/CD. Focuses on real-time streaming, scalability, and Star Schema modeling.
Production-ready Databricks Asset Bundle for cross-region ML model serving using Delta Sharing. Deploy models and feature tables across workspaces with zero-copy data access and automated online feature store sync.
Enterprise Medallion Lakehouse processing IoT telemetry for heavy machinery. Built with PySpark, Azure Databricks, and Power BI. Features FinOps IaC, CI/CD via Databricks Asset Bundles, and a DirectQuery predictive maintenance semantic model.
Restaurant Chain Analytics Lakehouse with Data Vault 2.1 architecture and dimensional modeling using star schema via Lakeflow Spark Declarative Pipelines in Azure Databricks