Resources and insights
Our Blog
Explore insights and practical tips on mastering Databricks Data Intelligence Platform and the full spectrum of today's modern data ecosystem.
Most teams that move to Databricks get the hard part right. They migrate the processing engine, rebuild the transformation logic, and stand up Unity Catalog. Then they leave Azure Data Factory running in the background: connected to everything, owned by nobody, and quietly accumulating cost and complexity. In this entry, that’s the gap we address.
Explore More Content
ML & AI
Genie Code Analysis: Two Weeks Later
Databricks Genie Code hit a 77.1% task success rate in production data science workflows — more than double what general-purpose coding agents achieve. But that performance is entirely conditional on the quality of your Unity Catalog metadata. SunnyData's two-week evaluation breaks down what works, what doesn't, and the governance layer you need before you go live.
Lakeflow Connect Free Tier: $35/Day Back in Your Budget
Databricks' permanent Lakeflow Connect free tier delivers 100 DBUs per workspace per day — covering up to 100 million records of ingestion at no additional compute cost. For enterprise teams running multiple workspaces, that's over $255,000 in avoided annual costs. This post breaks down the economics, architecture, and what it means for teams still paying a third-party ETL tax.
Temp Tables Are Here, and They're Going to Change How You Use SQL
Learn how temporary tables in Databricks SQL warehouses enable materialized data, DML operations, and session-scoped ETL workflows. Complete with practical examples.
Excel never dies (and neither does SharePoint)
Learn how Databricks' native Excel import handles multi-sheet workbooks, streaming autoloader, and SharePoint integration. Includes code examples.
Migrate Tables: From External to Managed Tables
Migrate external tables to managed tables in Databricks Unity Catalog using ALTER TABLE SET MANAGED while maintaining business continuity.
Defining Table Relationships in Databricks Genie
Master data relationships in Databricks Genie with and without foreign keys. Complete guide to using Unity Catalog constraints and defining custom joins for better AI insights.
Request Access Through Unity Catalog
Implement Databricks Unity Catalog access requests to streamline data permissions. Step-by-step guide covering BROWSE privileges, automated workflows, and ITSM integrations for better data governance.
Grant individual permission to secrets read in Unity Catalog
Implement granular secret access control in Unity Catalog that goes beyond traditional Key Vault-level permissions. This advanced approach uses Unity Catalog UDFs with service credentials to create secret-specific access controls, allowing you to grant users access to individual secrets rather than entire Key Vaults.
Query Your Lakehouse In Under 1 ms.: From OLAP to OLTP
Databricks Lakehouse excels at analytical workloads but struggles with single-row lookups that customer-facing applications demand. This blog shows how to solve the OLAP vs OLTP dilemma by adding PostgreSQL-based OLTP capabilities to your existing Lakehouse architecture
Add External Data Sources to Unity Catalog Lineage
Enhance your Unity Catalog lineage by incorporating external data sources such as Kafka streams and IoT devices. This blog covers both UI-based and programmatic methods for creating complete data lineage visibility in your Databricks environment.