Oracle Launches HeatWave GenAI with In-Database LLMs and Vector Store

Oracle introduces HeatWave GenAI, featuring the industry's first in-database large language models and an automated vector store. This innovation enables enterprises to develop AI applications without data migration, AI expertise, or additional costs, outperforming competitors like Snowflake and Google BigQuery.
Oracle Launches HeatWave GenAI with In-Database LLMs and Vector Store
Image Credit: Oracle

Customers can now build generative AI applications seamlessly without AI expertise, data movement, or additional costs. HeatWave GenAI surpasses Snowflake, Google BigQuery, and Databricks in vector processing speeds, achieving 30x, 18x, and 15x faster performance, respectively.

Introduction of HeatWave GenAI


Oracle has announced the availability of HeatWave GenAI, featuring the industry’s first in-database large language models (LLMs) and an automated in-database vector store. This new offering also includes scale-out vector processing and contextual natural language interactions informed by unstructured content. HeatWave GenAI enables enterprises to leverage generative AI directly within their database, eliminating the need for AI expertise or data migration to separate vector databases. It is now available across all Oracle Cloud regions, Oracle Cloud Infrastructure (OCI) Dedicated Region, and multiple clouds at no extra cost to HeatWave customers.

Key Features of Oracle HeatWave GenAI

Developers can now create a vector store for unstructured enterprise content using a single SQL command, thanks to built-in embedding models. Users can perform natural language searches in one step, utilizing either in-database or external LLMs. With HeatWave’s high scalability and performance, there is no need to provision GPUs, simplifying application complexity, enhancing performance, and improving data security while lowering costs.

Enhancing Enterprise AI with HeatWave

“HeatWave’s innovation continues with the integration of HeatWave GenAI,” said Edward Screven, Oracle’s Chief Corporate Architect. “These AI enhancements enable developers to quickly build rich generative AI applications without AI expertise or data movement. Users can interact intuitively with enterprise data to get accurate business insights swiftly.”

Industry Reactions to HeatWave GenAI

Vijay Sundhar, CEO of SmarterD, commented, “HeatWave GenAI simplifies generative AI use, significantly reducing application complexity, inference latency, and costs. This democratization of AI will enhance our productivity and enrich our applications.”

Innovative In-Database LLMs and Vector Store by Oracle

Oracle’s in-database LLMs reduce the complexity and cost of developing generative AI applications. Customers can perform data searches, content generation, and retrieval-augmented generation (RAG) within HeatWave’s vector store. Additionally, HeatWave GenAI integrates with OCI Generative AI services to access pre-trained models from leading LLM providers.

The automated in-database vector store allows businesses to utilize generative AI with their documents without transferring data to a separate database. The process, including document discovery, parsing, embedding generation, and insertion into the vector store, is fully automated within the database, making HeatWave Vector Store efficient and user-friendly.

Oracle HeatWave’s Advanced Vector Processing

HeatWave’s scale-out vector processing supports fast and accurate semantic search results. It introduces a native VECTOR data type and an optimized distance function, enabling semantic queries using standard SQL. HeatWave’s in-memory hybrid columnar representation and scale-out architecture ensure near-memory bandwidth execution and parallel processing across up to 512 nodes.

HeatWave Chat: Simplifying User Interactions

HeatWave Chat, a Visual Code plug-in for MySQL Shell, provides a graphical interface for HeatWave GenAI. It allows developers to ask questions in natural language or SQL, maintain context, and verify answer sources. The integrated Lakehouse Navigator facilitates the creation of vector stores from object storage, enhancing the user experience.

Impressive Benchmark Performance of HeatWave GenAI

HeatWave GenAI demonstrates impressive performance in creating vector stores and processing vector queries. It is 23x faster than Amazon Bedrock for creating vector stores and up to 80x faster than Amazon Aurora PostgreSQL for similarity searches, delivering accurate results with predictable response times.

HeatWave GenAI: Customer and Analyst Endorsements

Safarath Shafi, CEO of EatEasy, praised HeatWave’s in-database AutoML and LLMs for their differentiated capabilities, enabling new customer offerings and improving performance and quality of LLM results.

Eric Aguilar, founder of Aiwifi, highlighted HeatWave’s simplicity, security, and cost-effectiveness in leveraging generative AI for enterprise needs.

Holger Mueller, VP at Constellation Research, emphasized HeatWave’s integration of automated in-database vector stores and LLMs, which allows developers to create innovative applications without moving data, ensuring high performance and cost efficiency.

Oracle HeatWave: Integrated AI and Analytics Solution

HeatWave is the only cloud service offering integrated generative AI and machine learning for transactions and lakehouse-scale analytics. It is a key component of Oracle’s distributed cloud strategy, available on OCI, AWS, Microsoft Azure via Oracle Interconnect for Azure, and in customer data centers with OCI Dedicated Region and Oracle Alloy.

Read the HeatWave technical blog


Recent Content

Looking to learn AI in 2025 without breaking the bank? This blog breaks down the best free AI courses and certifications from top platforms like Google, IBM, and Harvard. Whether you’re a beginner, teacher, or tech professional, you’ll find career-relevant learning paths, direct course links, and tips to get certified and start building AI projects today.
Explore the transformative potential of Open Radio Access Networks (O-RAN) as it integrates AI, enhances security, and fosters interoperability to reshape mobile network infrastructure. In this article, we explore the advancements and challenges of O-RAN, revealing how it sets the stage for future mobile communications with smarter, more secure, and highly adaptable network solutions. Dive into the strategic implications for the telecommunications industry and learn why O-RAN is critical for the next generation of digital connectivity.
Nvidia’s Open Power AI Consortium is pioneering the integration of AI in energy management, collaborating with industry giants to enhance grid efficiency and sustainability. This initiative not only caters to the rising demands of data centers but also promotes the use of renewable energy, illustrating a significant shift towards environmentally sustainable practices. Discover how this synergy between technology and energy sectors is setting new benchmarks in innovative and sustainable energy solutions.
SK Telecom’s AI assistant, adot, now features Google’s Gemini 2.0 Flash, unlocking real-time Google search, source verification, and support for 12 large language models. The integration boosts user trust, expands adoption from 3.2M to 8M users, and sets a new standard in AI transparency and multi-model flexibility for digital assistants in the telecom sector.
SoftBank has launched the Large Telecom Model (LTM), a domain-specific, AI-powered foundation model built to automate telecom network operations. From base station optimization to RAN performance enhancement, LTM enables real-time decision-making across large-scale mobile networks. Developed with NVIDIA and trained on SoftBank’s operational data, the model supports rapid configuration, predictive insights, and integration with SoftBank’s AITRAS orchestration platform. LTM marks a major step in SoftBank’s AI-first strategy to build autonomous, scalable, and intelligent telecom infrastructure.
Telecom providers have spent over $300 billion since 2018 on 5G, fiber, and cloud-based infrastructure—but returns are shrinking. The missing link? Network observability. Without real-time visibility, telecoms can’t optimize performance, preempt outages, or respond to security threats effectively. This article explores why observability must become a core priority for both operators and regulators, especially as networks grow more dynamic, virtualized, and AI-driven.

Download Magazine

With Subscription
Whitepaper
5G network rollouts are now sprouting around the globe as operators get to grips with the potential of new enterprise applications. Yet behind the scenes, several factors still could strongly impact just how transformative this technology will be in years to come. Ultimately, it will all boil down to one...
NetInsight Logo
Whitepaper
System integrators play a crucial role in the network ecosystem by bringing together various components and technologies from the diverse network ecosystem players to build, deploy, and operate comprehensive end-to-end solutions that meet the specific needs of their clients....
Tech Mahindra Logo

It seems we can't find what you're looking for.

Subscribe To Our Newsletter

Scroll to Top