包阅导读总结
1. 关键词:Redis Query Engine、Vector Database、Throughput、Performance、Scaling
2. 总结:Redis 推出了更快的 Query Engine,在速度和扩展性方面表现出色,比其他向量数据库和模仿者具有更高的性能和吞吐量,今秋将在 Redis Cloud 可用,详情见工程博客。
3. 主要内容:
– Redis 一直致力于提供快速的数据和 AI 基础设施
– 最新的 Redis Query Engine 加速了查询、搜索和向量工作负载
– 新引擎使 Redis 比其他基准测试的向量数据库更快
– 与其他向量数据库对比
– 测试了其他七种顶级向量数据库,Redis 表现更优
– 包括纯向量数据库、通用数据库及 Redis 模仿者
– Redis 的高性能实现
– 关键设计选择包括单线程和低时间复杂度操作
– 新引擎通过多线程提高查询吞吐量
– 新 Redis Query Engine 可用性
– 已在 Redis Software 通用可用,今秋将在 Redis Cloud 可用
– 鼓励试用,了解更多可查看工程博客及获取企业支持
思维导图:
文章地址:https://redis.io/blog/announcing-faster-redis-query-engine-and-our-vector-database-leads-benchmarks/
文章来源:redis.io
作者:Jim Allen Wallace
发布时间:2024/6/20 13:00
语言:英文
总字数:926字
预计阅读时间:4分钟
评分:91分
标签:技术,Redis,查询引擎,向量数据库,性能
以下为原文内容
本内容来源于用户推荐转载,旨在分享知识与观点,如有侵权请联系删除 联系邮箱 media@ilingban.com
Redis has always been built for speed. Our mission is to provide the fastest data and AI infrastructure available for enterprises at scale. Our latest enhancement to our Redis Query Engine accelerates current query, search, and vector workloads, unlocking more throughput at high speed. You can build faster apps, and scale with better performance for your end users.
Today, we’re announcing the general availability of the faster Redis Query Engine on Redis Software. The new Redis Query Engine boosts current Redis query throughput 16X and this makes Redis faster than any other vector database we benchmarked against.
Redis offers better performance than any other vendor.
We tested the other top seven vector databases and Redis outperformed them across a wide variety of scenarios. We used industry-standard benchmarks where we used identical hardware and tested four data sets which simulate diverse use cases. Read our engineering blog to learn more about how we did it.
Our customers are building innovative apps that require more throughput while keeping the latency low. With GenAI, customers are deploying chatbots that need real-time RAG and must process multiple steps and retrieve data from vector databases, instantly. Meanwhile, Large Language Models (LLMs) continue to get faster, increasing the need for other components to be fast too. OpenAI released GPT-4o and Google introduced Gemini 1.5 Flash, to address apps’ needs for real-time performance. Apps require responses within a window of milliseconds at scale and under load, so we needed to boost throughput to support customers’ changing needs.
Redis Query Engine uses multi-threading to improve query throughput.
Redis has achieved its high performance from a few key design choices. One principle is that Redis has historically been single-threaded. Redis’ architecture uses a shared-nothing model, which avoids the bottlenecks and locking issues associated with multithreaded access to memory. Another design principle is that each operation needs to have a low time complexity. So even with a single thread, Redis delivers a very high throughput.
As customers increasingly leverage Redis for their real-time data platforms, there’s a growing demand to utilize its unique advanced query capabilities. Delivering more relevant data to fast apps requires more sophisticated and intricate queries. As the volume of data increases to hundreds of millions of documents, execution time for some queries correspondingly increases. This results in more time spent executing complex queries, reducing the availability of the main thread, and limiting throughput. While horizontal scaling can increase throughput for Redis operations, it is not the case for Redis queries.We must scale vertically.

By enabling queries to access the index concurrently, effectively allowing Redis to scale vertically, we enable scaling both the throughput of Redis operations and queries. The new Redis Query Engine unlocks the next level of performance for larger datasets with higher throughput needs. In regards to speed, our latency maintains submillisecond responses like developers expect from Redis operations, and queries average latency below two-digit milliseconds.
We compared Redis with three segments of vector database providers.
Our three segments included pure vector database providers, general-purpose databases with vector capabilities, and Redis imitators. We found that the speed, extensibility, and enterprise features varied significantly across the three groups. The pure vector databases were better at performance, but worse at scaling. The general-purpose databases were much worse at performance, but better at other integrations. And Redis imitators were significantly slower without the latest improvements in speed. Many users assume that Redis imitators are equally fast, but we want to break that illusion for you, to save you time and money, by comparing them to Redis.
Redis is faster than pure vector databases and also scales better.
One of the challenges we hear from customers is that they need a solution that’s built for scale and provides enterprise-grade services. We experienced this ourselves when we benchmarked other pure vector databases since we had difficulties keeping availability of the service under high load. This led us to move from their cloud service to an on-premises deployment to run the benchmarks. While these vector databases came the closest in performance, in our testing, they had a few issues getting set up and running within their cloud service. We documented it in our detailed benchmark report.
Redis is MUCH faster than general-purpose databases that added vector support.
There’s been a surge of vector database options in the last two years as RAG has become a top framework for enhancing GenAI apps. Many of these databases with new support for vectors are already widely adopted, so also using them for vectors can be easy. What we found is that Redis does here what Redis does best. We’re fast. Redis is widely used because we speed up data access from traditional databases. We find that customers also prefer Redis because we speed up GenAI apps the same way.
Redis imitators can’t keep up.
Compared to other Redis imitators, such as Amazon MemoryDB and Google Cloud MemoryStore for Redis, Redis demonstrates significantly higher throughput. Redis has 3.9 times more throughput for lower-dimensional datasets compared with Amazon MemoryDB and has 77% more throughput than Google Cloud MemoryStore for Redis. Don’t settle for an imitator.
See it for yourself.
Try out the higher throughput with the new Redis Query Engine. It’s now generally available in Redis Software and will be coming to Redis Cloud this fall. To learn more, check out our engineering blog. To get faster query speeds for your app today, download Redis Software and contact your rep for enterprise support.