StreamNative Introduces Lakestream Architecture and Launches Native Kafka Service

Read Announcement > Sign Up Now >
StreamNative Logo

Stream Directly to Your Lakehouse — No ETL Required

Ursa Engine writes every event directly to Iceberg and Delta Lake tables — queryable in seconds, not hours.

Organizations across industries use StreamNative

CHALLENGES

Why Batch Lakehouse Ingestion Falls Short

ETL Bottleneck to Lakehouse

Batch connectors and staging tables add hours of delay between event and insight. Analysts work on yesterday's data.

Format Lock-In

Proprietary storage formats prevent query engine portability. Switching analytics tools means costly migration.

Separate Streaming & Storage

Running Kafka for streaming and a separate system for lakehouse ingestion creates operational overhead and data duplication.

Stale Analytics

Dashboards and ML models train on data that's hours or days old. Business decisions are based on outdated information.

THE WAY OUT

The Streaming Lakehouse Advantage

Every event written to a topic simultaneously lands as a row in your lakehouse — no connectors, no staging, no batch windows.

Streaming Lakehouse architecture diagram showing data flow from external sources through Kafka streaming into the Streaming Augmented Lakehouse

Direct-to-Lakehouse Streaming

Every event in Kafka or Pulsar simultaneously lands as a row in Iceberg or Delta tables. No staging, no connectors.

Open Table Formats

Iceberg and Delta Lake tables work with Databricks, Snowflake, Trino, and Spark. No format lock-in.

Sub-Minute Data Freshness

BI dashboards and ML pipelines see fresh data within seconds, not hours.

HOW IT WORKS

Anatomy of a Streaming Lakehouse

A three-layer model: Data - Metadata - Protocol

Data Layer

Events land durably in a write-ahead log and compact into Parquet with atomic catalog updates. Query engines see fresh + historical data through a union read path. Choose latency- or cost-optimized mode per stream.

LEARN MORE

Metadata Layer

A streaming-aware catalog tracks schemas and a streaming offset index that maps offsets to WAL/Parquet files. That enables high performance ingestion and unified governance across streams and tables.

LEARN MORE

Protocol Layer

Stateless services speak Kafka or Pulsar protocols and translate client calls to storage operations. Because brokers are stateless, you scale compute and storage independently, add capacity in seconds, and maintain native client support.

LEARN MORE

RELATED TOOL

Powered by StreamNative Ursa

Ursa is the leaderless, diskless engine behind StreamNative — delivering direct-to-table writes, native Kafka support, and elastic scaling from a single unified platform.

Leaderless & Diskless

Ursa decouples compute from disk-free, leaderless brokers to enable instant failover, elastic scaling, and lower costs.

Stream-as-Table Storage

Ursa ensures exactly-once ingestion and unified data consistency through range-based indexing and atomic Parquet commits.

Native Kafka

Ursa provides native Kafka via stateless brokers with tunable storage for optimized latency or cost.

Compare Streaming Architectures

How Streaming Lakehouse stacks up against Kafka > ETL > Lakehouse, Kafka Tiered Storage, and Streaming Databases across data copies, freshness, compatibility, analytics, and ops.

Data Copies
1 (open table)
Freshness to Query
Seconds
Ingestion Protocol
Kafka or Pulsar
Query Engine
Any SQL on Iceberg/Delta
Ops
Single system

Why Streaming Augmented Lakehouse (SAL) wins

One system, one copy. SAL writes streams directly to Iceberg/Delta as Parquet using a unified catalog. By serving streams and tables from the same bytes, it eliminates connectors, slashes costs, and enables instant analytics.

FAQs

Yes. StreamNative provides native Apache Kafka service powered by Ursa. Your existing Kafka producers and consumers connect without code changes — just point them at your StreamNative cluster endpoint.

No. With Ursa's stream-table duality, every stream is simultaneously a lakehouse table. Ursa writes every event directly to Iceberg or Delta Lake tables as it arrives — no connectors to deploy, manage, or monitor.

StreamNative supports Apache Iceberg and Delta Lake. Data lands in open Parquet files under your catalog, queryable from Databricks, Snowflake, Trino, Spark, and any engine that reads these formats.

Data is queryable within seconds of being produced. Ursa writes to the write-ahead log and compacts into Parquet with atomic catalog updates, giving you sub-minute freshness for analytics and ML.

For cost-optimized topics, Ursa's leaderless, diskless brokers are stateless — compute and storage scale independently. Adding capacity takes seconds, failover takes seconds, and data is durably stored in object storage with eleven nines of durability. Latency-optimized topics retain Kafka's full replication model for the lowest latency.

Stream Directly to Your Lakehouse Today

  • Unify customer events across every channel on one platform.
  • Land campaign data directly in your lakehouse for real-time analytics.
  • Scale elastically for campaigns, launches, and seasonal peaks.