Dremio today announced that the metadata catalog at the heart of its Apache Iceberg-based data lakehouse now supports other popular metadata catalog services, including Snowflake’s Apache Polaris-based catalog and Databricks Unity Catalog. The lakehouse provider says the move in its Project Nessie-based metadata catalog will bolster architectural flexibility in the cloud, on-prem, and everywhere in between.
Before metadata catalogs suddenly jumped into the big data consciousness earlier this year, Dremio had been quietly backing its own metadata catalog, dubbed Project Nessie, to provide the necessary housekeeping that a lakehouse based on Apache Iceberg tables requires.
So when Snowflake announced the open source Polaris metadata catalog during its user conference in early June, Dremio executives applauded the announcement and the openness that it could foster in the big data community. Seeing close alignment between Polaris and Nessie, which began development in 2020, Dremio executives pledged to work with the Polaris community to merge the two projects.
The Nessie-Polaris merger has yet to happen, but it is still in the plans. “Our goal is to merge the capabilities of Project Nessie into Apache Polaris (Incubating) to create a single, unified catalog,” says James Rowland-Jones, vice president of product at Dremio. “We believe this will become the default catalog for the open-source community. Dremio will continue to focus on seamless enterprise services built around it.”
In the meantime, Dremio is moving forward with development its own catalog service for technical metadata, dubbed the Dremio Enterprise Data Catalog. Specifically, Dremio today announced several new capabilities in the metadata catalog, which is based on Nessie.
The new bits include integration with the Snowflake metadata catalog service based on Apache Polaris as well as hooking into Unity Catalog, the metadata catalog that Databricks built for managing data stored in Delta Lake tables (Unity Catalog does quite a bit more, including lineage tracking, semantic modeling, security, governance, and functions as a regular, user-focused data catalog, but that’s another story).
Dremio’s move is noteworthy for a couple of reasons. For starters, with its acquisition of Iceberg maker Tabular for between $1 billion and $2 billion and its commitments to essentially merge the Delta Lake and Iceberg specs, Databricks helped to ease CFOs who were worried that they would pick the “wrong” format.
However, while Databricks committed earlier this year to supporting Iceberg tables with a future release of Unity Catalog, that support is not available yet. Dremio’s support for Unity Catalog ensures that Databricks customers who use its metadata catalog can achieve that interoperability with Polaris today.
“Flexibility is essential for modern organizations looking to maximize the value of their data,” said Tomer Shiran, Founder of Dremio. “With expanded Iceberg catalog support across all environments, Dremio empowers businesses to deploy their lakehouse architecture wherever it’s most effective. We’re 100% committed to giving customers the freedom to choose the best tools and infrastructure while reducing fears of vendor lock-in.”
Dremio’s product, which is officially called the Dremio Enterprise Data Catalog for Apache Iceberg, supports all Iceberg engines through the Iceberg REST API. In addition to supporting Dremio’s own SQL query engine, it supports other Iceberg-compatible query engines, including Apache Spark, Flink, and others.
Dremio’s catalog automates many of the housekeeping tasks that are required to keep an Iceber-based data lakehouse running at peak efficiency. That includes things like table optimization routines, such as compaction and garbage collection. It also provides “Git”-like branching and version control, enabling users to access data as it existed at particular moments in time (so-called “time travelling”). The catalog also provides centralized data governance and role-based access control (RBAC), ensuring fine-grained access to data and preventing user access to of sensitive data.
Kevin Petrie, vice president of research at BARC, says Dremio’s move helps enterprises deal with the “extraordinary pressure to access, prepare, and govern distributed datasets for consumption by analytics and AI applications.”
“To meet this demand, they need to catalog diverse data and metadata across data centers, regions, and clouds,” Petrie said in Dremio’s press release. “Dremio is taking a logical step to enable this with an open catalog that is based on Apache Iceberg, the emerging standard for flexible table formats, and by integrating with an ecosystem of popular platforms.”
Related Items:
Polaris Catalog, To Be Merged With Nessie, Now Available on GitHub
What the Big Fuss Over Table Formats and Metadata Catalogs Is All About
October 29, 2024
- Cockroach Labs’ 2025 Resilience Report Unveils Critical Outage Surge and Unprepared Enterprises Worldwide
- Dremio Unveils Full Flexibility with Data Catalog for Apache Iceberg Across All Environments
- Cisco Introduces NVIDIA-Powered AI Servers and PODs for Scalable AI Workloads
- Quest Uncovers Top Data Intelligence Strategies for AI Readiness and Governance in 2024
- GMI Cloud Raises $82M to Drive Global Access to Advanced GPUs and Cloud Infrastructure
- Timescale Brings PostgreSQL into the GenAI Era with pgai Vectorizer
October 28, 2024
- Solix Technologies, Inc. to Deliver Unsurpassed Data Governance, Privacy and Security with 3rd Generation Cloud Data Platform
- OSI Announces the Release of the Industry’s First Open Source AI Definition
October 25, 2024
- MotherDuck Unveils Beta pg_duckdb Extension, Bringing DuckDB Analytics Directly to PostgreSQL
- CTERA Named a Leader in GigaOm Radar for Cloud-Native Globally Distributed File Systems
- Oracle and Yurts Collaborate to Bring Secure GenAI Solutions to Defense and Intelligence Sectors
October 24, 2024
- Acryl Data Announces Inaugural Metadata and AI Summit 2024 to Explore Enterprise AI Solutions
- ServiceNow Launches Workflow Data Fabric, Powering AI Agents with Unified Business Data
- Ataccama Debuts AI Agent for Data Management
- Cockroach Labs Partners with AWS to Enhance Cloud Migrations and GenAI Capabilities
- Freshworks Unveils Autonomous AI Agent for Customer and Employee Service
- Starburst Announces 100GB/s Streaming Ingest from Apache Kafka to Apache Iceberg Tables
- Datorios Announces New Apache Flink Observability Capabilities for Responsible Agentic AI
- Cisco Unveils AI-Powered Webex Solutions to Enhance Customer Experiences
- Kipi.ai Delivers Marketing Mix Modeling and Analytics App on Snowflake AI Data Cloud