Lakehouse Days: July 2025, Bengaluru

Want to see e6data in action?

Learn how data teams power their workloads.

Get Demo
Get Demo

Join us for an exclusive in-person event on “From Stream to Lakehouse: Real-Time Audits, Unified AI, and Next-Gen Iceberg Catalogs,” hosted by e6data in Bengaluru!

Lakehouse Days - powered by AWS, is designed specifically for data engineers, data architects, and senior software engineers who constantly seek to optimize their data architecture to make it more price-performant while delivering the best user experience.

In this edition, we have three incredible talks lined up:

  • Confluent will showcase how Kafka is being used for capturing, processing, and delivering auditlogs across all customers of Confluent Cloud. Additionally, using Apache Iceberg (Confluent Tableflow) for archival and data analytics use cases.
  • AWS will present Amazon SageMaker Unified Studio and Lakehouse, a single pane that unifies data prep, ML, and generative-AI workflows across S3 and Redshift.
  • e6data will present the case for new-age Iceberg catalogs like Polaris, Gravitino, Lakekeeper, Project Nessie, and Unity Catalog. Further, let’s delve into Iceberg’s data streaming problems and e6data’s solution to this.

Lakehouse Days, powered by AWS, is designed to enable fellow data geeks to meet, network, and engage in insightful discussions on the complex world of data.

Register now!

This is an exclusive, invitation-only event. Please RSVP to reserve your spot through this link: https://lu.ma/kzs7jh63?utm_source=website

Venue - Amazon Development Center, Aquila, Bagmane Constellation Business Park Block-7, Bagmane Constellation Service Rd, Ferns City, Doddanekkundi, Bengaluru, Karnataka 560048

​Date and time - Jul 12, 2025, from 09:30 AM to 2:00 PM

Meet the Speakers

Ankit Garg, Senior Software Engineer, Confluent
Devanshu Bagadia
, Software Engineer, Confluent

Topic: Real-time use case of data streaming platforms

Summary: Real-time streaming use-case: Confluent cloud auditlogs: How Kafka is being used for capturing, processing, and delivering auditlogs across all customers of confluent cloud. Using Apache Iceberg (Confluent Tableflow) for archival and data analytics use case. 

Time: 9:30 - 10:15 AM IST

Ravi Kompella, Analytics Specialist, AWS

Topic: Simplifying Data and AI with Amazon SageMaker Unified Studio and Amazon SageMaker Lakehouse
Summary: Data workers often face complex tools and fragmented workflows. This talk introduces Amazon SageMaker Unified Studio and Amazon SageMaker Lakehouse, designed to dramatically simplify your data and AI journey. Unified Studio provides a single, intuitive environment for all your analytics and ML tasks, from data prep to generative AI, eliminating context-switching. Lakehouse unifies data access across S3 data lakes and Redshift, enabling you to build powerful AI solutions on a single, consistent data copy. Discover how these services streamline your tasks, accelerate innovation, and unlock your data's full potential.

Time: 10:30 - 11:15 AM IST

Ankur Rajan, Senior Software Engineer, e6data

Topic: Emerging Catalogs and Streaming Ingest Problems in Iceberg

Summary: The competition for Apache Iceberg catalogs certainly remains fierce and relevant. Several new catalogs have emerged in recent times, including Apache Polaris™ (incubating), Apache Gravitino, Lakekeeper, and Project Nessie, among others. For data engineers, these developments are fascinating. But most of the developer community still banks on “classic” catalogs, such as Hadoop Catalog, Hive Metastore, and AWS Glue, but the landscape is expanding rapidly. This session will delve into these catalogs with hands-on demos, discuss streaming ingest problems in Iceberg, and provide e6data’s solution to these issues. 

Time: 11:30 AM - 12:15 PM IST

Share on

Build future-proof data products

Try e6data for your heavy workloads!

Get Started for Free
Get Started for Free
Frequently asked questions (FAQs)
How do I integrate e6data with my existing data infrastructure?

We are universally interoperable and open-source friendly. We can integrate across any object store, table format, data catalog, governance tools, BI tools, and other data applications.

How does billing work?

We use a usage-based pricing model based on vCPU consumption. Your billing is determined by the number of vCPUs used, ensuring you only pay for the compute power you actually consume.

What kind of file formats does e6data support?

We support all types of file formats, like Parquet, ORC, JSON, CSV, AVRO, and others.

What kind of performance improvements can I expect with e6data?

e6data promises a 5 to 10 times faster querying speed across any concurrency at over 50% lower total cost of ownership across the workloads as compared to any compute engine in the market.

What kinds of deployment models are available at e6data ?

We support serverless and in-VPC deployment models. 

How does e6data handle data governance rules?

We can integrate with your existing governance tool, and also have an in-house offering for data governance, access control, and security.