Oracle Launches HeatWave GenAI with In-Database LLMs and Vector Store

Oracle introduces HeatWave GenAI, featuring the industry's first in-database large language models and an automated vector store. This innovation enables enterprises to develop AI applications without data migration, AI expertise, or additional costs, outperforming competitors like Snowflake and Google BigQuery.
Oracle Launches HeatWave GenAI with In-Database LLMs and Vector Store
Image Credit: Oracle

Customers can now build generative AI applications seamlessly without AI expertise, data movement, or additional costs. HeatWave GenAI surpasses Snowflake, Google BigQuery, and Databricks in vector processing speeds, achieving 30x, 18x, and 15x faster performance, respectively.

Introduction of HeatWave GenAI


Oracle has announced the availability of HeatWave GenAI, featuring the industry’s first in-database large language models (LLMs) and an automated in-database vector store. This new offering also includes scale-out vector processing and contextual natural language interactions informed by unstructured content. HeatWave GenAI enables enterprises to leverage generative AI directly within their database, eliminating the need for AI expertise or data migration to separate vector databases. It is now available across all Oracle Cloud regions, Oracle Cloud Infrastructure (OCI) Dedicated Region, and multiple clouds at no extra cost to HeatWave customers.

Key Features of Oracle HeatWave GenAI

Developers can now create a vector store for unstructured enterprise content using a single SQL command, thanks to built-in embedding models. Users can perform natural language searches in one step, utilizing either in-database or external LLMs. With HeatWaveโ€™s high scalability and performance, there is no need to provision GPUs, simplifying application complexity, enhancing performance, and improving data security while lowering costs.

Enhancing Enterprise AI with HeatWave

โ€œHeatWaveโ€™s innovation continues with the integration of HeatWave GenAI,โ€ said Edward Screven, Oracleโ€™s Chief Corporate Architect. โ€œThese AI enhancements enable developers to quickly build rich generative AI applications without AI expertise or data movement. Users can interact intuitively with enterprise data to get accurate business insights swiftly.โ€

Industry Reactions to HeatWave GenAI

Vijay Sundhar, CEO of SmarterD, commented, โ€œHeatWave GenAI simplifies generative AI use, significantly reducing application complexity, inference latency, and costs. This democratization of AI will enhance our productivity and enrich our applications.โ€

Innovative In-Database LLMs and Vector Store by Oracle

Oracleโ€™s in-database LLMs reduce the complexity and cost of developing generative AI applications. Customers can perform data searches, content generation, and retrieval-augmented generation (RAG) within HeatWaveโ€™s vector store. Additionally, HeatWave GenAI integrates with OCI Generative AI services to access pre-trained models from leading LLM providers.

The automated in-database vector store allows businesses to utilize generative AI with their documents without transferring data to a separate database. The process, including document discovery, parsing, embedding generation, and insertion into the vector store, is fully automated within the database, making HeatWave Vector Store efficient and user-friendly.

Oracle HeatWaveโ€™s Advanced Vector Processing

HeatWaveโ€™s scale-out vector processing supports fast and accurate semantic search results. It introduces a native VECTOR data type and an optimized distance function, enabling semantic queries using standard SQL. HeatWaveโ€™s in-memory hybrid columnar representation and scale-out architecture ensure near-memory bandwidth execution and parallel processing across up to 512 nodes.

HeatWave Chat: Simplifying User Interactions

HeatWave Chat, a Visual Code plug-in for MySQL Shell, provides a graphical interface for HeatWave GenAI. It allows developers to ask questions in natural language or SQL, maintain context, and verify answer sources. The integrated Lakehouse Navigator facilitates the creation of vector stores from object storage, enhancing the user experience.

Impressive Benchmark Performance of HeatWave GenAI

HeatWave GenAI demonstrates impressive performance in creating vector stores and processing vector queries. It is 23x faster than Amazon Bedrock for creating vector stores and up to 80x faster than Amazon Aurora PostgreSQL for similarity searches, delivering accurate results with predictable response times.

HeatWave GenAI: Customer and Analyst Endorsements

Safarath Shafi, CEO of EatEasy, praised HeatWaveโ€™s in-database AutoML and LLMs for their differentiated capabilities, enabling new customer offerings and improving performance and quality of LLM results.

Eric Aguilar, founder of Aiwifi, highlighted HeatWaveโ€™s simplicity, security, and cost-effectiveness in leveraging generative AI for enterprise needs.

Holger Mueller, VP at Constellation Research, emphasized HeatWaveโ€™s integration of automated in-database vector stores and LLMs, which allows developers to create innovative applications without moving data, ensuring high performance and cost efficiency.

Oracle HeatWave: Integrated AI and Analytics Solution

HeatWave is the only cloud service offering integrated generative AI and machine learning for transactions and lakehouse-scale analytics. It is a key component of Oracleโ€™s distributed cloud strategy, available on OCI, AWS, Microsoft Azure via Oracle Interconnect for Azure, and in customer data centers with OCI Dedicated Region and Oracle Alloy.

Read theย HeatWave technical blog


Recent Content

The integration of tariffs and the EU AI Act creates a challenging environment for the advancement of AI and automation. Tariffs, by increasing the cost of essential hardware components, and the EU AI Act, by increasing compliance costs, can significantly raise the barrier to entry for new AI and automation ventures. European companies developing these technologies may face a double disadvantage: higher input costs due to tariffs and higher compliance costs due to the AI Act, making them less competitive globally. This combined pressure could discourage investment in AI and automation within the EU, hindering innovation and slowing adoption rates. The resulting slower adoption could limit the availability of crucial real-world data for training and improving AI algorithms, further impacting progress.
NVIDIA has launched a major U.S. manufacturing expansion for its next-gen AI infrastructure. Blackwell chips will now be produced at TSMCโ€™s Arizona facilities, with AI supercomputers assembled in Texas by Foxconn and Wistron. Backed by partners like Amkor and SPIL, NVIDIA is localizing its AI supply chain from silicon to system integrationโ€”laying the foundation for โ€œAI factoriesโ€ powered by robotics, Omniverse digital twins, and real-time automation. By 2029, NVIDIA aims to manufacture up to $500B in AI infrastructure domestically.
Samsung has launched two new rugged devicesโ€”the Galaxy XCover7 Pro smartphone and the Tab Active5 Pro tabletโ€”designed for high-intensity fieldwork in sectors like logistics, healthcare, and manufacturing. These devices offer military-grade durability, advanced 5G connectivity, and enterprise-ready security with Samsung Knox Vault. Features like hot-swappable batteries, gloved-touch sensitivity, and AI-powered tools enhance productivity and reliability in harsh environments.
Nokia, Digita, and CoreGo have partnered to roll out private 5G networks and edge computing solutions at high-traffic event venues. Using Nokia’s Digital Automation Cloud (DAC) and CoreGoโ€™s payment and access tech, the trio delivers real-time data flow, reliable connectivity, and enhanced guest experience across Finland and international locationsโ€”serving over 2 million attendees to date.
OpenAI is developing a prototype social platform featuring an AI-powered content feed, potentially placing it in direct competition with Elon Musk’s X and Metaโ€™s AI initiatives. Spearheaded by Sam Altman, the project aims to harness user-generated content and real-time interaction to train advanced AI systemsโ€”an approach already used by rivals like Grok and Llama.
AI Pulse: Telecomโ€™s Next Frontier is a definitive guide to how AI is reshaping the telecom landscape โ€” strategically, structurally, and commercially. Spanning over 130 pages, this MWC 2025 special edition explores AIโ€™s growing maturity in telecom, offering a comprehensive look at the technologies and trends driving transformation.

Explore strategic AI pillarsโ€”from AI Ops and Edge AI to LLMs, AI-as-a-Service, and governanceโ€”and learn how telcos are building AI-native architectures and monetization models. Discover insights from 30+ global CxOs, unpacking shifts in leadership thinking around purpose, innovation, and competitive advantage.

The edition also examines connected industries at the intersection of Private 5G, AI, and Satelliteโ€”fueling transformation in smart manufacturing, mobility, fintech, ports, sports, and more. From fan engagement to digital finance, from smart cities to the industrial metaverse, this is the roadmap to telecomโ€™s next eraโ€”where intelligence is the new infrastructure, and telcos become the enablers of everything connected.

Download Magazine

With Subscription
Whitepaper
Telecom networks are facing unprecedented complexity with 5G, IoT, and cloud services. Traditional service assurance methods are becoming obsolete, making AI-driven, real-time analytics essential for competitive advantage. This independent industry whitepaper explores how DPUs, GPUs, and Generative AI (GenAI) are enabling predictive automation, reducing operational costs, and improving service quality....
Whitepaper
Explore the collaboration between Purdue Research Foundation, Purdue University, Ericsson, and Saab at the Aviation Innovation Hub. Discover how private 5G networks, real-time analytics, and sustainable innovations are shaping the "Airport of the Future" for a smarter, safer, and greener aviation industry....
Article & Insights
This article explores the deployment of 5G NR Transparent Non-Terrestrial Networks (NTNs), detailing the architecture's advantages and challenges. It highlights how this "bent-pipe" NTN approach integrates ground-based gNodeB components with NGSO satellite constellations to expand global connectivity. Key challenges like moving beam management, interference mitigation, and latency are discussed, underscoring...

Subscribe To Our Newsletter

Scroll to Top