Spice Cloud v1.7.0: DataFusion v49, Full-Text Search Updates & More
Spice Cloud & Spice.ai Enterprise 1.7.0 are now live, bringing performance upgrades with DataFusion v49, real-time full-text search indexing, EmbeddingGemma support, and improvements across search, embeddings, and API integrations. Spice Cloud customers will automatically upgrade to v1.7.0 on deployment, while Spice.ai Enterprise customers can consume the Enterprise v1.7.0 image from the Spice AWS Marketplace listing.
What’s New in v1.7.0
DataFusion v49 Upgrade
Spice now runs on DataFusion v49, delivering lower latency and improved query optimization.

DataFusion v49 highlights include:
- Dynamic filters and pushdown to skip unnecessary reads in
ORDER BY & LIMITqueries - Compressed spill files to reduce disk usage during large sorts and aggregations
- Support for ordered-set aggregates with
WITHIN GROUP - New
REGEXP_INSTRfunction to identify regex match positions
EmbeddingGemma Support
Spice now supports EmbeddingGemma, Google’s latest embedding model for text and documents. It delivers high-quality embeddings for semantic search, retrieval, and recommendation tasks. Configure it directly in your Spicepod via HuggingFace.
Embedding Request Caching
Repeated embedding requests can now be cached in the Spice runtime. This reduces both latency and costs, with configurable cache size and TTL options. Check out the caching documentation for more details.
Real-Time Indexing for Full Text Search
Full-text indexing now supports real-time changes from CDC streams such as Debezium. New events are searchable as they arrive, ensuring continuously fresh results.
OpenAI Responses API Tool Calls with Streaming
The OpenAI Responses API in Spice now supports tool calls with streaming. Results from tools like web_search and code_interpreter are streamed as they’re generated, enabling more responsive agent and application experiences.
Bug & Stability Fixes
v1.7.0 includes numerous fixes and improvements:
- CDC streams readiness and full-text indexing reliability
- Vector search pipeline and
vector_searchUDTF fixes - Kafka schema inference, consumer group persistence, and cooperative mode
- Error reporting improvements (e.g., ThrottlingException handling)
- Iceberg connector support for
LIMITpushdown - S3 Vector ingestion reliability and tracing fixes
v1.7 Release Community Call
We’ll walk through highlights of v1.7 live on our Release Community Call. Join us to see the new functionality in action and bring your questions! Register here.
.png)
Resources to Get Started with Spice
- Sign up for Spice Cloud for free, or get started with Spice Open Source
- Explore the Spice cookbooks and docs
- Get a demo if you'd like to see the product live or have any questions
Explore more lorem ipsum sit amet consectetur
Vorem ipsum dolor sit amet, consectetur adipiscing elit. Nunc vulputate libero et velit interdum, ac aliquet odio mattis.
Spice Cloud v1.9.0: Introducing the Spice Cayenne Data Accelerator
Spice Cloud & Spice.ai Enterprise 1.9.0 are live! Our mission at Spice is to make building data-intensive applications and AI systems easier, faster, and more secure. With v1.9.0, we’re taking a big step forward. This release introduces Spice Cayenne, our new premier data accelerator based on Vortex, upgrades to DataFusion v50 and DuckDB v1.4.2, new HTTP Data Connector support for querying API endpoints […]

Spice Cloud v1.8.0: Iceberg Write Support, Acceleration Snapshots & More
Spice Cloud & Spice.ai Enterprise 1.8.0 are live! v1.8.0 includes Iceberg write support, acceleration snapshots, partitioned S3 Vector indexes, a new AI SQL function for LLM integration, and an updated Spice.js SDK. V1.8.0 also introduces developer experience upgrades, including a redesigned Spice Cloud dashboard with tabbed navigation: Spice Cloud customers will automatically upgrade to v1.8.0 on deployment, while […]

True Hybrid Search: Vector, Full-Text, and SQL in One Runtime
TL;DR Show Me the (Data)! It’s well established (and maybe even trite) to say that enterprises are going all-in on artificial intelligence, with more than $40 billion directed toward generative AI projects in recent years. The initial results have been underwhelming. A recent study from the Massachusetts Institute of Technology’s NANDA initiative concluded that despite the enormous allocation of […]

See Spice in action
Get a guided walkthrough of how development teams use Spice to query, accelerate, and integrate AI for mission-critical workloads.
Get a demo

