Close Menu
  • Home
  • AI
  • Big Data
  • Cloud Computing
  • iOS Development
  • IoT
  • IT/ Cybersecurity
  • Tech
    • Nanotechnology
    • Green Technology
    • Apple
    • Software Development
    • Software Engineering

Subscribe to Updates

Get the latest technology news from Bigteetechhub about IT, Cybersecurity and Big Data.

    What's Hot

    AND logic nanoparticle for precision immunotherapy of metastatic cancers

    March 9, 2026

    Taiwan Excellence showcases AI breakthroughs at Embedded World 2026

    March 9, 2026

    What cybersecurity actually does for your business

    March 9, 2026
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    Big Tee Tech Hub
    • Home
    • AI
    • Big Data
    • Cloud Computing
    • iOS Development
    • IoT
    • IT/ Cybersecurity
    • Tech
      • Nanotechnology
      • Green Technology
      • Apple
      • Software Development
      • Software Engineering
    Big Tee Tech Hub
    Home»Big Data»Building a near real-time application with Zerobus Ingest and Lakebase
    Big Data

    Building a near real-time application with Zerobus Ingest and Lakebase

    big tee tech hubBy big tee tech hubMarch 8, 2026008 Mins Read
    Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email Telegram WhatsApp
    Follow Us
    Google News Flipboard
    Building a near real-time application with Zerobus Ingest and Lakebase
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link


    Event data from IoT, clickstream, and application telemetry powers critical real-time analytics and AI when combined with the Databricks Data Intelligence Platform. Traditionally, ingesting this data required several data hops (message bus, Spark jobs) between the data source and the lakehouse. This adds operational overhead, data duplication, requires specialized expertise, and it’s generally inefficient when the lakehouse is the only destination for this data.

    Once this data lands in the lakehouse, it is transformed and curated for downstream analytical use cases. However, teams need to serve this analytical data for operational use cases, and building these custom applications can be a laborious process. They need to provision and maintain essential infrastructure components like a dedicated OLTP database instance (with networking, monitoring, backups, and more). Additionally, they need to manage the reverse ETL process for the analytical data into the database to resurface it in a real-time application. Customers also often build additional pipelines to push data from the lakehouse into these external operational databases. These pipelines add to the infrastructure that developers need to set up and maintain, altogether diverting their attention from the main goal: building the applications for their business.

    So how does Databricks simplify both ingesting data into the lakehouse and serving gold data to support operational workloads?

    Enter Zerobus Ingest and Lakebase.

    About Zerobus Ingest

    Zerobus Ingest, part of Lakeflow Connect, is a set of APIs that provide a streamlined way to push event data directly into the lakehouse. Eliminating the single-sink message bus layer entirely, Zerobus Ingest reduces infrastructure, simplifies operations, and delivers near real-time ingestion at scale. As such, Zerobus Ingest makes it easier than ever to unlock the value of your data.

    The data-producing application must specify a target table to write data to, ensure that the messages map correctly to the table’s schema, and then initiate a stream to send data to Databricks. On the Databricks side, the API validates the schemas of the message and the table, writes the data to the target table, and sends an acknowledgment to the client that the data has been persisted.

    Key benefits of Zerobus Ingest:

    • Streamlined architecture: eliminates the need for complex workflows and data duplication.
    • Performance at scale: supports near real-time ingestion (up to 5 secs) and allows thousands of clients writing to the same table (up to 100MB/sec throughput per client).
    • Integration with the Data Intelligence Platform: accelerates time to value by enabling teams to apply analytics and AI tools, such as MLflow for fraud detection, directly on their data.

    Zerobus Ingest Capability

    Specifications

    Ingestion latency

    Near real-time (≤5 seconds)

    Max throughput per client

    Up to 100 MB/sec

    Concurrent clients

    Thousands per table

    Continuous sync lag (Delta → Lakebase)

    10–15 seconds

    Real-time foreach writer latency

    200–300 milliseconds

    About Lakebase

    Lakebase is a fully managed, serverless, scalable, Postgres database built into the Databricks Platform, designed for low-latency operational and transactional workloads that run directly on the same data powering analytical and AI use cases. 

    The complete separation of compute and storage delivers rapid provisioning and elastic autoscaling. Lakebase’s integration with the Databricks Platform is a major differentiator from traditional databases because Lakebase makes Lakehouse data directly available to both real-time applications and AI without the need for complex custom data pipelines. It is built to deliver database creation, query latency, and concurrency requirements to power enterprise applications and agentic workloads. Lastly, it allows developers to easily version control and branch databases like code.

    Key benefits of Lakebase:

    • Automatic data synchronization: Ability to easily sync data from the Lakehouse (analytical layer) to Lakebase on a snapshot, scheduled, or continuous basis, without the need for complex external pipelines
    • Integration with the Databricks Platform: Lakebase integrates with Unity Catalog, Lakeflow Connect, Spark Declarative Pipelines, Databricks Apps, and more.
    • Integrated permissions and governance: Consistent role and permissions management for operational and analytical data. Native Postgres permissions can still be maintained via the Postgres protocol.

    Together, these tools allow customers to ingest data from several systems directly into Delta tables and implement reverse ETL use cases at scale. Next, we will explore how to use these technologies to implement a near real-time application!

    How to Build a Near Real-time Application

    As a practical example, let’s help ‘Data Diners,’ a food delivery company, empower their management staff with an application to monitor driver activity and order deliveries in real-time. Currently, they lack this visibility, which limits their ability to mitigate issues as they arise during deliveries.

    Why is a real-time application valuable? 

    • Operational awareness: Management can instantly see where each driver is and how their current deliveries are progressing. That means fewer blind spots with late orders or when a driver needs assistance.
    • Issue mitigation: Live location and status data enable dispatchers to reroute drivers, adjust priorities, or proactively contact customers in the event of delays, reducing failed or late deliveries.

    Let’s see how to build this with Zerobus Ingest, Lakebase, and Databricks Apps on the Data Intelligence Platform!

    Overview of Application Architecture

    Application Architecture: Data Producer, Zerobus Ingest, Delta, Lakebase, Databricks Apps

    This end-to-end architecture follows four stages: (1) A data producer uses the Zerobus SDK to write events directly to a Delta table in Databricks Unity Catalog. (2) A continuous sync pipeline pushes updated records from the Delta table to a Lakebase Postgres instance. (3) A FastAPI backend connects to Lakebase via WebSockets to stream real-time updates. (4) A front-end application built on Databricks Apps visualizes the live data for end users.

    Starting with our data producer, the data diner app on the driver’s phone will emit GPS telemetry data about the driver’s location (latitude and longitude coordinates) en route to deliver orders. This data will be sent to an API gateway, which ultimately sends the data to the next service in the ingestion architecture.

    With the Zerobus SDK, we can quickly write a client to forward events from the API gateway to our target table. With the target table being updated in near real time, we can then create a continuous sync pipeline to update our lakebase tables. Finally, by leveraging Databricks Apps, we can deploy a FastAPI backend that uses WebSockets to stream real-time updates from Postgres, along with a front-end application to visualize the live data flow.

    Before the introduction of the Zerobus SDK, the streaming architecture would have included multiple hops before it landed in the target table. Our API gateway would have needed to offload the data to a staging area like Kafka, and we would need Spark Structured Streaming to write the transactions into the target table. All of this adds unnecessary complexity, especially given that the sole destination is the lakehouse. The architecture above instead demonstrates how the Databricks Data Intelligence Platform simplifies end-to-end enterprise application development — from data ingestion to real-time analytics and implementation of interactive applications.

    Getting Started

    Prerequisites: What You Need

    Step 1: Create a target table in Databricks Unity Catalog

    The event data produced by the client applications will live in a Delta table. Use the code below to create that target table in your desired catalog and schema.

    Step 2: Authenticate using OAUTH

    Step 3: Create the Zerobus client and ingest data into the target table

    The code below pushes the telemetry events data into Databricks using the Zerobus API. 

    Change Data Feed (CDF) limitation and workaround

    As of today, Zerobus Ingest does not support CDF. CDF allows Databricks to record change events for new data written to a delta table. These change events could be inserts, deletes, or updates. These change events can then be used to update the synced tables in Lakebase. To sync data to Lakebase and continue with our project, we will write the data in the target table to a new table and enable CDF on that table.

    Step 4: Provision Lakebase and sync data to database instance

    To power the app, we will sync data from this new, CDF-enabled table into a Lakebase instance. We will sync this table continuously to support our near real-time dashboard.

    Create synched table into a Lakebase instance

    In the UI, we select:

    • Sync Mode: Continuous for low-latency updates
    • Primary Key: table_primary_key

    This ensures the app reflects the latest data with minimal delay.

    Note: You can also create the sync pipeline programmatically using the Databricks SDK.

    Real-time mode via foreach writer

    Continuous syncs from Delta to Lakebase has a 10-15-second lag, so if you need lower latency, consider using real-time mode via ForeachWriter writer to sync data directly from a DataFrame to a Lakebase table. This will sync the data within milliseconds.

    Refer to the Lakebase ForeachWriter code on Github.

    Step 5: Build the app with FastAPI or another framework of choice

    Screenshot of RideShare360 application

    With your data synced to Lakebase, you can now deploy your code to build your app. In this example, the app fetches events data from Lakebase and uses it to update a near real-time application to track a driver’s activity while en route to making food deliveries. Read the Get Started with Databricks Apps docs to learn more about building apps on Databricks. 

    Additional Resources

    Check out more tutorials, demos and solution accelerators to build your own applications for your specific needs. 

    • Build an End-to-End Application: A real-time sailing simulator tracks a fleet of sailboats using Python SDK and the REST API, with Databricks Apps and Databricks Asset Bundles. Read the blog.
    • Build a Digital Twins Solution: Learn how to maximize operational efficiency, accelerate real-time insight and predictive maintenance with Databricks Apps and Lakebase. Read the blog.

    Learn more about Zerobus Ingest, Lakebase, and Databricks Apps in the technical documentation. You can also take a look at the Databricks Apps Cookbook and Cookbook Resource Collection.

    Conclusion

    IoT, clickstream, telemetry, and similar applications generate billions of data points every day, which are used to power critical real-time applications across several industries. As such, simplifying ingestion from these systems is paramount. Zerobus Ingest provides a streamlined way to push event data directly from these systems into the lakehouse while ensuring high performance. It pairs nicely with Lakebase to simplify end-to-end enterprise application development.



    Source link

    application Building Ingest Lakebase RealTime Zerobus
    Follow on Google News Follow on Flipboard
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
    tonirufai
    big tee tech hub
    • Website

    Related Posts

    Native Splunk Integration in Cisco Nexus One: Real-Time Insights

    March 9, 2026

    Zero-ETL integrations with Amazon OpenSearch Service

    March 9, 2026

    Unique Capabilities of Edge Computing in IoT

    March 7, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Editors Picks

    AND logic nanoparticle for precision immunotherapy of metastatic cancers

    March 9, 2026

    Taiwan Excellence showcases AI breakthroughs at Embedded World 2026

    March 9, 2026

    What cybersecurity actually does for your business

    March 9, 2026

    Native Splunk Integration in Cisco Nexus One: Real-Time Insights

    March 9, 2026
    About Us
    About Us

    Welcome To big tee tech hub. Big tee tech hub is a Professional seo tools Platform. Here we will provide you only interesting content, which you will like very much. We’re dedicated to providing you the best of seo tools, with a focus on dependability and tools. We’re working to turn our passion for seo tools into a booming online website. We hope you enjoy our seo tools as much as we enjoy offering them to you.

    Don't Miss!

    AND logic nanoparticle for precision immunotherapy of metastatic cancers

    March 9, 2026

    Taiwan Excellence showcases AI breakthroughs at Embedded World 2026

    March 9, 2026

    Subscribe to Updates

    Get the latest technology news from Bigteetechhub about IT, Cybersecurity and Big Data.

      • About Us
      • Contact Us
      • Disclaimer
      • Privacy Policy
      • Terms and Conditions
      © 2026 bigteetechhub.All Right Reserved

      Type above and press Enter to search. Press Esc to cancel.