

(Shutterstock AI Generator/Shutterstock)
If you’re in the hunt for an enterprise data catalog, you may want to keep Atlan on your speed dial, as the young company is quickly gathering momentum. In fact, Atlan recently nabbed the number one overall rating from Forrester in the space. But as Atlan’s CEO tells us, there’s a lot more innovation yet to come from data catalogs.
In its recent Forrester Wave for Enterprise Data Catalogs, Forrester gave the Atlan Enterprise Data Catalog a score of 4.20 and gave the company’s strategy a score of 4.50, both of which were higher than all 11 other data catalog vendors in the report. The only category in which Atlan did not nab first place was market presence, where the six-year-old company scored a 2.00, well behind bigger, older, and more established players in this market.
“Even as a recent entrant in this market, Atlan’s Third-Gen Data Catalog is quickly outpacing established players by adeptly anticipating and addressing strategic customer needs through automation,” Forrester lead analyst Jayesh Chaurasia wrote. “Atlan is a visionary player with a clear, ambitious goal: to become the data and AI control plane enabling complex business use cases.”
The Forrester analysts liked several aspects of Atlan’s current offering, including its capability to enable data democratization and self-service through automated metadata tracking, its use of GenAI to assist with discovery, its end-to-end lineage tracking, “and a Netflix-like personalized experience to all business and technical personas.”
Atlan scored low in revenue, which isn’t surprising for a relatively new entrant into the space. However, Forrester gave it a decent score for the number of customers, which shows the company is gaining ground.
Three Types of Data Catalogs
In a recent interview with Datanami, Atlan CEO and co-founder Prukalpa Shankar discussed the state of the data catalog market in 2024, the challenges the company had to overcome to reach its current state, and where she sees the future taking data catalogs.
“I think maybe the most overused word of the last couple of years has been catalog,” Shankar said. “When someone asks me something, I say, what do you mean when you say the word ‘catalog,’ because it means slightly different things.”
There are three types of data catalos, Shankar said, starting with catalogs that store and expose technical metadata to enable applications to share data, such as AWS‘s Amazon Glue (Snowflake’s new Polaris catalog and Databricks’ Unity Catalog would also fit the bill here).
“Metadata is becoming incredibly important for driving downstream use cases,” Shankar said. “And so we’re seeing vendors across the space opening up their metadata APIs, almost like, in the application world, what single sign-on was for SaaS applications. We’re seeing metadata becoming the single sign-on for the data world.”
The second type of data catalog is the data dictionary version, which gets closer to the users and requires a better user experience, Shankar said. Tableau led the way with its Tableau Data Catalog, which allows users to discover what various metrics meant within the context of the BI environment so they can make sense of it. Another product like this is dbt Labs Explorer, she said.
“The third is what we think of ourselves, which is more of the control plane version of a catalog,” Shankar said. “The foundation of the control plane is in your metadata layer, which is being able to bring together metadata from all of these ecosystems, stitch it together, make it intelligent, make sense of it, but then drive the use cases” across these ecosystems.
Data Control Planes
The control plane version of a data catalog that Atlan builds must be able to handle a large diversity of data, users, and tools. Data of all types; users like data analysts, data engineers, and data scientists; and tools ranging from BI products to ETL and data transformation tools to data warehouses and date lakes, all must work with this system.
As Forrester points out, Atlan has done a good job of handling the current ecosystem of data, tools, and users. The company has tapped AI and machine learning to automate metadata tracking where it can, thereby lifting the burden of manually stitching and staging data off the shoulders of data stewards.
“Three years ago, I had written this article called Data Catalog 3.0…that said metadata is becoming big data and we need to think about the foundational computation systems of metadata the way we thought about big data,” she said. “The interesting thing, three years later is, I don’t think it’s becoming big data. It is big data. We have customers who, in their starting week, are bringing in millions of assets into [the catalog] The scale of what we are dealing with from a metadata perspective is a whole different scale than what existed five or six years ago.”
The automation of metadata tracking is important now, but it will become even more important in the future, as the volume and variety of use cases that data catalogs must manage expands ever outward and upward.
“In two years from now, our data consumers will be LLMs [large language models] and in this LLM stack, there’s a whole different world of problems that we’re dealing with,” Shankar said. “We’re probably not going to only stick to a single foundational model. We’re going to have significant multiple deployments across architectures. We’ll deal with unstructured data. And the only thing that that’s preventing us from getting to that world is the concept of AI-ready data.”
Solving Data Management
The foundational challenges in data management haven’t changed in more than 25 years, Shankar said. Getting the right data to the right place at the right time remains the ultimate goal. But of course, the type of data, and the places that people want to consume it–not to mention the timeline (i.e. now)–have changed a lot, which is part and parcel of the challenge faced not just by data catalog vendors like Atlan, but the data management field as a whole.
Recent industry events, such as the emergence of Apache Iceberg as a standard for table formats and the Iceberg REST API for connecting to metadata catalogs, such as Snowflake’s Polaris and Databricks’ Unity Catalog, are good for customers. Shankar hopes that drives the discussion toward greater openness higher up the data catalog stack, and eventually into the control plane.
“I’m very bullish about the version of the world that is moving to more and more open standards,” Shankar said. “There have been now foundational improvements I think from a from a data lake layer, with open standards from your data itself, so you can bring your own compute. I think the same will happen in the metadata layer.”
Customers naturally want to avoid lock-in, whether it’s a cloud lock-in, database lock-in, table format lock-in, or data catalog lock-in. Even if the Atlan product is not open source, Shankar said that Atlan strives to be open with its platform, and to open up access to its metadata. “The more players start opening up metadata, the more customers start asking for it,” she said.
Atlan uses a graph database to help it make sense of the different types of metadata that it tracks. That includes table metadata, operational metadata from data pipelines, lineage metadata from SQL transformations, and compliance metadata, which is tracked as tags. By collecting and tracking all this metadata as graph and exposing it through the control plane, Atlan is able to deliver better visibility and access to customers.
“I had a customer the other day who said ‘Data storage is cheap. Data confusion is not,’” Shankar said. “And if you see the evolution, the final leg is that our end users are using data, trusting data, [and embarking upon] data-driven decision making.
“The final leg actually is still very similar to what it used to be 15 years ago, despite having the ecosystem going through three layers of technology transformations,” she continued. “And I think we’re now finally at a point where we can solve for the final leg. I think that’s the last step to the problem statement that needs to be solved.”
Related Items:
Data Catalogs Vs. Metadata Catalogs: What’s the Difference?
Active Metadata – The New Unsung Hero of Successful Generative AI Projects
Atlan Plants Itself in the Middle of the Data Governance Map with $105M Round
March 26, 2025
- Quest Adds GenAI to Toad to Bridge the Skills Gap in Modern Database Management
- SymphonyAI Expands Industrial AI to the Edge with Microsoft Azure IoT Operations
- New Relic Report Reveals Media and Entertainment Sector Looks to Observability to Drive Adoption of AI
- Databricks and Anthropic Sign Deal to Bring Claude Models to Data Intelligence Platform
- Red Hat Boosts Enterprise AI Across the Hybrid Cloud with Red Hat AI
March 25, 2025
- Cognizant Advances Industry AI with NVIDIA-Powered Agents, Digital Twins, and LLMs
- Grafana Labs Unveils 2025 Observability Survey Findings and Open Source Updates at KubeCon Europe
- Algolia Boosts Browse with AI-Powered Collections
- AWS Expands Amazon Q in QuickSight with New AI Scenarios Capability
- Komprise Automates Complex Unstructured Data Migrations
- PEAK:AIO Chosen by Scan to Support Next-Gen GPUaaS Platform
- Snowflake Ventures Deepens Investment in DataOps.live to Advance Data Engineering Automation
- KX Emerges as Standalone Software Company to Make Temporal AI a Commercial Reality
- PAC Storage Unveils 5000 Series Data Storage Solutions
March 24, 2025
- Tessell Introduces Fully Managed Database Service on Google Cloud
- Datavault AI Joins IBM Partner Plus to Transform AI-Driven Data Monetization
- Cerabyte Unveils Immutable Data Storage for Government Customers
- Provenir Highlights AI-Driven Risk Decisioning in Datos Insights Report
- Algolia Showcases Powerful AI-Driven Search at ShopTalk Spring 2025
- StarTree Awarded 2025 Confluent Data Flow ISV Partner of the Year – APAC
- PayPal Feeds the DL Beast with Huge Vault of Fraud Data
- OpenTelemetry Is Too Complicated, VictoriaMetrics Says
- When Will Large Vision Models Have Their ChatGPT Moment?
- The Future of AI Agents is Event-Driven
- Accelerating Agentic AI Productivity with Enterprise Frameworks
- Your Next Big Job in Tech: AI Engineer
- Data Warehousing for the (AI) Win
- Nvidia Touts Next Generation GPU Superchip and New Photonic Switches
- Krishna Subramanian, Komprise Co-Founder, Stops By the Big Data Debrief
- Alation Aims to Automate Data Management Drudgery with AI
- More Features…
- IBM to Buy DataStax for Database, GenAI Capabilities
- Clickhouse Acquires HyperDX To Advance Open-Source Observability
- NVIDIA GTC 2025: What to Expect From the Ultimate AI Event?
- Excessive Cloud Spending In the Spotlight
- EDB Says It Tops Oracle, Other Databases in Benchmarks
- Databricks Unveils LakeFlow: A Unified and Intelligent Tool for Data Engineering
- Google Launches Data Science Agent for Colab
- Meet MATA, an AI Research Assistant for Scientific Data
- CDOAs Are Struggling To Measure Data, Analytics, And AI Impact: Gartner Report
- Big Data Heads to the Moon
- More News In Brief…
- Gartner Predicts 40% of Generative AI Solutions Will Be Multimodal By 2027
- Snowflake Ventures Invests in Anomalo for Advanced Data Quality Monitoring in the AI Data Cloud
- NVIDIA Unveils AI Data Platform for Accelerated AI Query Workloads in Enterprise Storage
- Accenture Invests in OPAQUE to Advance Confidential AI and Data Solutions
- Qlik Study: 94% of Businesses Boost AI Investment, But Only 21% Have Fully Operationalized It
- Seagate Unveils IronWolf Pro 24TB Hard Drive for SMBs and Enterprises
- Gartner Identifies Top Trends in Data and Analytics for 2025
- Qlik Survey Finds AI at Risk as Poor Data Quality Undermines Investments
- Palantir and Databricks Announce Strategic Product Partnership to Deliver Secure and Efficient AI to Customers
- Cisco Expands Partnership with NVIDIA to Accelerate Enterprise AI Adoption
- More This Just In…