Stream data from BigQuery to Databricks
Sync your BigQuery data with Databricks in minutes using Estuary for real-time, no-code integration and seamless data pipelines.
- No credit card required
- 30-day free trial


- 200+Of connectors
- 5500+Active users
- <100 msEnd-to-end latency
- 7+GB/secSingle dataflow

BigQuery connector details
The BigQuery Batch connector captures data from your BigQuery datasets into Flow collections by periodically running SQL queries and translating the results into JSON documents. It’s ideal for analytical workloads or datasets that don’t require continuous change tracking.
- Periodically executes queries from BigQuery tables or views and loads results into Flow collections
- Supports cursor-based incremental updates for efficient polling and reduced reprocessing
- Customizable polling intervals (default: every 24 hours) for flexible scheduling
- Uses a Google Cloud Service Account with BigQuery User and Data Viewer roles for secure access
- Works seamlessly within Estuary’s Private and BYOC environments for compliance and governance
💡 Tip: Specify a timestamp or monotonically increasing ID column as your cursor to reduce data volume and capture only new or updated records efficiently.

Databricks connector details
Estuary’s Databricks connector materializes Flow collections into tables in a Databricks SQL Warehouse. Data changes are first written to a Unity Catalog Volume and then applied transactionally to Databricks tables, ensuring reliable delivery of real-time updates. This connector supports both standard merges and delta updates for high-performance CDC workflows.
- Real-time CDC data sync from Flow collections into Databricks tables
- Choice of standard (merge) or delta updates to optimize performance and costs
- Native integration with Databricks Unity Catalog for secure, managed storage
- Flexible sync scheduling to balance freshness and warehouse costs

See how Launchmetrics uses Databricks
How to integrate BigQuery with Databricks in 3 simple steps using Estuary
Connect BigQuery as Your Real-Time Data Source
Set up a real-time source connector for BigQuery in minutes. Estuary captures change data (CDC), events, or snapshots — no custom pipelines, agents or manual configs needed.
Configure Databricks as Your Target
Choose Databricks as your target system. Estuary intelligently maps schemas, supports both batch and streaming loads, and adapts to schema changes automatically.
Deploy and Monitor Your End-to-End Data Pipeline
Launch your pipeline and monitor it from a single UI. Estuary guarantees exactly-once delivery, handles backfills and replays, and scales with your data — without engineering overhead.
Estuary in action
See how to build end-to-end pipelines using no-code connectors in minutes. Estuary does the rest.
Why Estuary is the best choice for data integration
Estuary combines the most real-time, streaming change data capture (CDC), and batch connectors together into a unified modern data pipeline:

What customers are saying
Increase productivity 4x
With Estuary companies increase productivity 4x and deliver new projects in days, not months. Spend much less time on troubleshooting, and much more on building new features faster. Estuary decouples sources and destinations so you can add and change systems without impacting others, and share data across analytics, apps, and AI.
Spend 2-5x less
Estuary customers not only do 4x more. They also spend 2-5x less on ETL and ELT. Estuary's unique ability to mix and match streaming and batch loading has also helped customers save as much as 40% on data warehouse compute costs.

BigQuery to Databricks pricing estimate
Estimated monthly cost to move 800 GB from BigQuery to Databricks is approximately $1,000.
Data moved
Choose how much data you want to move from BigQuery to Databricks each month.
GB
Choose number of sources and destinations.
Why pay more?
Move the same data for a fraction of the cost.



Frequently Asked Questions
- Set Up Capture: In Estuary, go to Sources, click + NEW CAPTURE, and select the BigQuery connector.
- Enter Details: Add your BigQuery connection details and click SAVE AND PUBLISH.
- Materialize Data: Go to Destinations, choose your target system, link the BigQuery capture, and publish.
What is BigQuery?
How do I Transfer Data from BigQuery?
What are the pricing options for Estuary?
Estuary offers competitive and transparent pricing, with a free tier that includes 2 connector instances and up to 10 GB of data transfer per month. Explore our pricing options to see which plan fits your data integration needs.
Getting started with Estuary
Free account
Getting started with Estuary is simple. Sign up for a free account.
Sign upDocs
Make sure you read through the documentation, especially the get started section.
Learn moreCommunity
I highly recommend you also join the Slack community. It's the easiest way to get support while you're getting started.
Join Slack CommunityEstuary 101
I highly recommend you also join the Slack community. It's the easiest way to get support while you're getting started.
Watch

Related integrations with BigQuery
DataOps made simple
Add advanced capabilities like schema inference and evolution with a few clicks. Or automate your data pipeline and integrate into your existing DataOps using Estuary's rich CLI.




































