
The End of the Data Broker Model and the Rise of the Data Marketplace

(PopTika/Shutterstock)
Most commodity markets have evolved from traditional broker models to more transparent, efficient transactions with lower fees and higher impact for both buyers and sellers. Travel agents, stock brokers, and even real estate agents (to name just a few) have been either completely replaced or greatly disrupted by digital marketplaces. Gone are the days of paying brokers high fixed or percentage fees for the straightforward tasks of connecting buyers and sellers to transact on largely standardized products or commodities.
Take for example the changes made in financial services: the rapid growth of passively managed, low fee Exchange Traded Funds (ETFs) vs. the more traditional actively managed mutual fund, with fund managers acting as intermediaries in the construction and management of the fund’s underlying assets.
I have seen this transition in digital advertising as well – the rise and decline of digital advertising networks. As with other ad markets, digital networks delivered scale, performance optimization, and rudimentary quality controls to digital ad buyers with non-transparent percentage fees on advertising spend. As ad formats commoditized, ad buyers found not just scale, but dramatically more transparency, quality control, and lower, more transparent fees through ad exchanges. The digital networks that captured the lion’s share of billions in digital ad spend outside of Google and Facebook in 2015 are essentially gone five years later.
This same phenomenon is now underway for buyers and sellers of data assets, and not just for martech and advertising use cases. According to the latest report by The Eckerson Group, The Rise of the Data Exchanges:
“Data exchanges are emerging as a key component of the data economy. By connecting data suppliers and consumers through a seamless experience, data exchanges have created a modern-day goldrush to help modern, digital organizations address their insatiable hunger for data. And, by eliminating the friction from finding, acquiring, and integrating data, these exchanges make it easy for any company to monetize its data assets and create new revenue streams.”
The rise of these data exchanges or marketplaces, and the decline of the broker model for data will benefit not just buyers and sellers, but more importantly, consumers that own the data they create through their digital interactions.
Before outlining the specific problems with the broker model, it’s important to ask what the key drivers are for buyers and sellers; how do most data transactions work today? Most transactions are managed by an intermediary broker through a process like this:
1) Broker asks the buyer what they are trying to achieve, e.g. better personalize an app experience, estimate the future value of a customer, understand store location visitation patterns, improve the relevance of an ad
2) Broker aggregates data assets (e.g. location data, mobile device IDs related to app installs, behavioral data derived from browsing behaviors of mobile app IDs) from a few sellers
3) Broker delivers that aggregated snapshot back to the buyer mapped to a list of emails, mobile IDs, geo locations, or similar
This method of transacting has several challenges for buyers and sellers, each of which are addressable if you fast forward beyond the data broker model.
Problem #1: Speed. Almost all the problems buyers are looking to solve are ongoing problems. Data assembled as a snapshot is stale the day after it’s delivered, and erodes quickly from then on. If you want to continue to personalize your app, you need current data all the time.
Solution: A robust data marketplace automatically manages updates to underlying data so the buyer has effectively “always on” access to the latest and greatest. There should be no need to rerun queries, normalization, etc. every time you refresh. The additional upside is that as soon as the time is cut down between the buying and selling of data there no longer is the opportunity cost that is associated with time. That opens up the ability to test, play, experiment, and change strategies as the business warrants.
Problem #2: Quality. Accuracy, completeness, recency, is generally opaque to the buyer. The buyer does not know what she’s getting even after she gets it. If the downstream results of the exercise are poor due to any of those quality issues she’ll have a hard time figuring that out, perhaps assuming the exercise had limited potential ROI based on the results.
Solution: An effective data marketplace or exchange scores the quality of underlying data assets to make critical quality attributes like accuracy and recency transparent to the buyer along with the identity of the sellers.
Problem #3: Discovery. In the broker model, buyers rely on the intermediary to find the right data on their behalf, without understanding the comprehensive catalog of what data sellers and solutions may be available. What may be “good enough” from the broker’s perspective this translates to missed opportunities for both buyers and sellers.
Solution: A true data marketplace enables buyers to easily and transparently search and browse the universe of data products that may help them improve results for their use case.
Problem #4: Non-transparent economics. Data brokers may charge arbitrary and/or high fees that obscure the underlying economics to both buyer and seller. This limits the potential value for both buyer and seller, which limits the overall potential for the marketplace.
Solution: An efficient data marketplace charges reasonable, transparent fees, and only for the data the buyer truly wants to buy. The buyer chooses the most promising data for their current application, for economics that are transparent and repeatable.
Problem #5: Compliance. Brokers have limited tools and incentives to help sellers and buyers understand source data compliance with applicable privacy and other regulatory issues, particularly as policies and regulations evolve over time. This is a huge problem for consumers. Any given data source may have unknown compliance, or worse yet, be out of compliance with critical consumer privacy protections. Consumers lose transparency and control over their data with no way to get it back.
Solution: a privacy-centric data marketplace not only promotes consumer choice and transparency with features like central opt outs, it enables all participants to track policies as they change over time, and automatically flags any data that has an unknown state of compliance.
Beyond the critical improvements needed for consumer privacy protection, the downstream effect of the problems together is the negative impact on buyer ROI. One-off transactions with low quality, potentially non-compliant data on top of unclear underlying economics means most broker managed transactions are unprofitable. Consumers do not get the more personalized app experience, the more relevant content or ad that the data buyer was trying to achieve. How much more is the buyer going to invest in these transactions?
By addressing these problems, a data marketplace or exchange that replaces brokers delivers better ROI for buyers, more demand for sellers, and better outcomes for consumers along with the critical improvements to transparency and choice mentioned above. As a board member for a company involved in creating data marketplaces, I’ve had the first-hand opportunity to see how an effective data marketplace can function and deliver these benefits for all participants in the marketplace.
Data brokers are going the way of travel agents. They’ll be around for a few more years, but if you’re looking for sustained results you’re much better off using a marketplace that delivers quality data quickly, transparently, and with verifiable protection for consumers’ privacy.
About the author: Ryan Christensen is a tech executive with 25 years of experience in enterprise, advertising, and marketplace technology businesses. Ryan was most recently SVP of Platform Product at Xandr (AT&T) after serving as the COO of AppNexus, which was acquired by AT&T in 2018. He currently sits on the Board of Narrative.io, a software company simplifying the acquisition and distribution of data. He can be reached on Twitter via @r_y_a_n_c and at https://www.linkedin.com/in/ryanechristensen/.
Related Items:
Taking the Pain Out of Buying and Selling Data
Easy Money? Considerations for Building Out Your Data Monetization Practice
February 7, 2025
- Palantir Wins Dresner Advisory Services 2024 Application Innovation and Technology Innovation Awards in Multiple Categories
- MOSTLY AI Unveils Open-Source Toolkit for Synthetic Data Generation
- Pecan AI Launches AI-Driven Co-Pilot for Predictive Analytics, Expanding Access to ML
- European Union: First Rules of the Artificial Intelligence Act Are Now Applicable
February 6, 2025
- SoftServe Survey Finds 58 Percent of Leaders Report Using Inaccurate Data
- Oracle Recognized as a Leader in the 2025 IDC MarketScape
- Cognida.ai Secures $15M Series A from Nexus Venture Partners
- TrueFoundry Secures $19 Million Series A Funding to Transform AI Deployment
- Glean Achieves $100M ARR in Three Years
- Moveworks Launches Quick GPT
- Qumulo Announces Cloud Data Fabric File System
- Aerospike Unveils Database 8 with Distributed ACID Transactions for OLTP
February 5, 2025
- Lightning AI Brings DeepSeek to Private Enterprise Clouds with AI Hub
- PEAK:AIO Powers AI Data for University of Strathclyde’s MediForge Hub
- dbt Labs Surpasses $100M ARR, Expands Global Customer Base
- Qlik Connect 2025 Brings AI, Iceberg, and Automation to the Forefront
- Immuta Finds Legacy Data Provisioning Systems Are Hindering AI Adoption
- Arcitecta Named a Leader in Coldago Research’s Map 2024 for Unstructured Data Management
- Hydrolix Releases Apache Spark Connector for Databricks Integration
February 4, 2025
- OpenTelemetry Is Too Complicated, VictoriaMetrics Says
- PayPal Feeds the DL Beast with Huge Vault of Fraud Data
- Inside Nvidia’s New Desktop AI Box, ‘Project DIGITS’
- The Top 2025 GenAI Predictions, Part 2
- What Are Reasoning Models and Why You Should Care
- Slicing and Dicing the Data Governance Market
- The Top 2025 Generative AI Predictions: Part 1
- Why Data Lakehouses Are Poised for Major Growth in 2025
- Your APIs are a Security Risk: How to Secure Your Data in an Evolving Digital Landscape
- Top-Down or Bottom-Up Data Model Design: Which is Best?
- More Features…
- Meet MATA, an AI Research Assistant for Scientific Data
- Dataiku Report Predicts Key AI Trends for 2025
- AI Agent Claims 80% Reduction in Time to Complete Data Tasks
- Qlik and dbt Labs Make Big Data Integration Acquisitions
- Bloomberg Survey Reveals Data Challenges for Investment Research
- Collibra Bolsters Position in Fast-Moving AI Governance Field
- Observo AI Raises $15M for Agentic AI-Powered Data Pipelines
- Anaconda’s Commercial Fee Is Paying Off, CEO Says
- Google Cloud’s 2023 Data and AI Trends Report Reveals a Changing Landscape
- Mathematica Helps Crack Zodiac Killer’s Code
- More News In Brief…
- Informatica Reveals Surge in GenAI Investments as Nearly All Data Leaders Race Ahead
- Gartner Predicts 40% of Generative AI Solutions Will Be Multimodal By 2027
- PEAK:AIO Powers AI Data for University of Strathclyde’s MediForge Hub
- Hightouch Partners with Databricks to Launch Self-Service Platform for Retail Media Networks
- Dremio’s New Report Shows Data Lakehouses Accelerating AI Readiness for 85% of Firms
- Scale Computing’s Edge Computing for In-Store Retail Solutions Showcased at NRF ’25
- dbt Labs Acquires SDF Labs to Elevate SQL Capabilities and Developer Efficiency
- Reltio Is Recognized as Representative Vendor in 2024 Gartner Market Guide for Master Data Management Solutions
- Seagate Unveils IronWolf Pro 24TB Hard Drive for SMBs and Enterprises
- TigerGraph Launches Savanna Cloud Platform to Scale Graph Analytics for AI
- More This Just In…