

(Summit Art Creations/Shutterstock)
Kinetica got its start building a GPU-powered database to serve fast SQL queries and visualizations for US government and military clients. But with a pair of announcements at Nvidia’s GTC show last week, the company is showing it’s prepared for the coming wave of generative AI applications, particularly those utilizing retrieval augmented generation (RAG) techniques to tap unique data sources.
Companies today are hunting for ways to leverage the power of large language models (LLMs) with their own proprietary data. Some companies are sending their data to OpenAI’s cloud or other cloud-based AI providers, while others are building their own LLMs.
However, many more companies are adopting the RAG approach, which has surfaced as perhaps the best middle ground between that doesn’t require building your own model (time-consuming and expensive) or sending your data to the cloud (not good privacy and security-wise).
With RAG, relevant data is injected directly into the context window before being sent off to the LLM for execution, thereby providing more personalization and context in the LLMs response. Along with prompt engineering, RAG has emerged as a low-risk and fruitful method for juicing GenAI returns.

The VRAM boost in Nvidia’s Blackwell GPU will help Kinetica keep the processor fed with data, Negahban said
Kinetica is also now getting into the RAG game with its database by essentially turning it into a vector database that can store and serve vector embeddings to LLMs, as well as by performing vector similarity search to optimize the data it sends to the LLM.
According to its announcement last week, Kinetica is able to serve vector embeddings 5x faster than other databases, a number it claims came from the VectorDBBench benchmark. The company claims its able to achieve that speed by leveraging Nvidia’s RAPIDS RAFT technology.
That GPU-based speed advantage will help Kinetica customers by enabling them to scan more of their data, including real-time data that has just been added to the database, without doing a lot of extra work, said Nima Negahban, co0founder and CEO of Kinetica.
“It’s hard for an LLM or a traditional RAG stack to be able to answer a question about something that’s happening right now, unless they’ve done a lot of pre-planning for specific data types,” Negahban told Datanami at the GTC conference last week, “whereas with Kinetica, we’ll be able to help you by looking at all the relational data, generate the SQL on the fly, and ultimately what we put just back in the context for the LLM is a simple text payload that the LLM will be able to understand to use to give the answer to the question.”
This essentially gives users the capability to talk to their complete corpus of relational enterprise data, without doing any preplanning.
“That’s the big advantage,” he continued, “because the traditional RAG pipelines right now, that part of it still requires a good amount of work as far as you have to have the right embedding model, you have to test it, you have to make sure it’s working for your use case.”
Kinetica can also talk to other databases and function as a generative federated query engine, as well as do the traditional vectorization of data that customers put inside of Kinetica, Negahban said. The database is designed to be used for operational data, such as time-series, telemetry, or teleco data. Thanks to the support for NVIDIA NeMo Retriever microservices, the company is able to position that data in a RAG workflow.
But for Kinetica, it all comes back to the GPU. Without the extreme computational power of the GPU, the company has just another RAG offering.
“Basically you need that GPU-accelerated engine to make it all work at the end of the day, because it’s got to have the speed,” said Negahban, a 2018 Datanami Person to Watch. “And we then put all that orchestration on top of it as far as being able to have the metadata necessary, being able to connect to other databases, having all that to make it easy for the end user, so basically they can start taking advantage of all that relational enterprise data in their LLM interaction.”
Related Items:
Bank Replaces Hundreds of Spark Streaming Nodes with Kinetica
Kinetica Aims to Broaden Appeal of GPU Computing
Preventing the Next 9/11 Goal of NORAD’s New Streaming Data Warehouse
February 14, 2025
- Clarifai Unveils Control Center for Enhanced AI Visibility and Decision-Making
- EDB Strengthens Partner Program to Accelerate Postgres and AI Adoption Worldwide
- Workday Introduces Agent System of Record for AI Workforce Management
- Fujitsu Unveils Generative AI Cloud Platform with Data Security Focus
- NTT DATA Highlights AI Responsibility Gap as Leadership Fails to Keep Pace
- Gurobi AI Modeling Empowers Users with Accessible Optimization Resources
February 13, 2025
- SingleStore Unveils No-Code Solution Designed to Cut Data Migration from Days to Hours
- Databricks Announces Launch of SAP Databricks
- SAP Debuts Business Data Cloud with Databricks to Turbocharge Business AI
- Data Science Salon Kickstarts 2025 with DSS ATX Conference, Featuring AI Startup Showcase
- Hydrolix Achieves Amazon CloudFront Ready Designation
- Astronomer Launches Astro Observe to Unify Data Observability and Orchestration
- EU Launches InvestAI Initiative to Build AI Gigafactories Across Europe
- HPE Announces Shipment of Its First NVIDIA Grace Blackwell System
- IDC Celebrates 60 Years of Tech Intelligence at Directions 2025
- Lucidity Gains $21M to Scale AI-Driven Cloud Storage Optimization
- Glean Launches Open Security and Governance Partner Program for Enterprise AI
February 12, 2025
- OpenTelemetry Is Too Complicated, VictoriaMetrics Says
- What Are Reasoning Models and Why You Should Care
- Three Ways Data Products Empower Internal Users
- Memgraph Bolsters AI Development with GraphRAG Support
- Keeping Data Private and Secure with Agentic AI
- Three Data Challenges Leaders Need To Overcome to Successfully Implement AI
- PayPal Feeds the DL Beast with Huge Vault of Fraud Data
- Top-Down or Bottom-Up Data Model Design: Which is Best?
- Inside Nvidia’s New Desktop AI Box, ‘Project DIGITS’
- Data Catalogs Vs. Metadata Catalogs: What’s the Difference?
- More Features…
- Meet MATA, an AI Research Assistant for Scientific Data
- AI Agent Claims 80% Reduction in Time to Complete Data Tasks
- DataRobot Expands AI Capabilities with Agnostiq Acquisition
- Collibra Bolsters Position in Fast-Moving AI Governance Field
- Snowflake Unleashes AI Agents to Unlock Enterprise Data
- Observo AI Raises $15M for Agentic AI-Powered Data Pipelines
- Anaconda’s Commercial Fee Is Paying Off, CEO Says
- Microsoft Open Sources Code Behind PostgreSQL-Based MongoDB Clone
- Confluent and Databricks Join Forces to Bridge AI’s Data Gap
- Mathematica Helps Crack Zodiac Killer’s Code
- More News In Brief…
- Informatica Reveals Surge in GenAI Investments as Nearly All Data Leaders Race Ahead
- Gartner Predicts 40% of Generative AI Solutions Will Be Multimodal By 2027
- PEAK:AIO Powers AI Data for University of Strathclyde’s MediForge Hub
- DataRobot Acquires Agnostiq to Accelerate Agentic AI Application Development
- TigerGraph Launches Savanna Cloud Platform to Scale Graph Analytics for AI
- EY and Microsoft Unveil AI Skills Passport to Bridge Workforce AI Training Gap
- Alluxio Enhances Enterprise AI with Version 3.5 for Faster Model Training
- DeepSeek-R1 models now available on AWS
- Lightning AI Brings DeepSeek to Private Enterprise Clouds with AI Hub
- Seagate Unveils IronWolf Pro 24TB Hard Drive for SMBs and Enterprises
- More This Just In…