We are thrilled to announce Dragonfly Search, enabling both vector search and faceted search in our robust and performant in-memory data store.
In this blog post, we navigate the journey of optimizing Dragonfly for BullMQ, achieving an exceptional 30x throughput increase. This endeavor highlights our deep commitment to the community and demonstrates our dedication to expanding and adapting Dragonfly to the open-source ecosystem.
In this blog post, we discover how Dragonfly boosts e-commerce platforms with efficient caching, personalized content, and managing high-traffic events. Learn about Dragonfly's capabilities for ensuring smooth operations during peak times.
We are thrilled to announce the General Availability of the Dragonfly Kubernetes Operator.
In this blog post, we build a real-time ad server application using Bun, ElysiaJS, and Dragonfly, showcasing seamless integration and exceptional developer experience. Dive into hands-on examples and enhance your understanding of these cutting-edge technologies.
In this post, we explore the seamless integration of Dragonfly as a drop-in replacement for Redis as the backing in-memory store of BullMQ, a robust background job processing library for Node.js.
In the Developing with Dragonfly series of blog posts, we will explore different techniques and best practices for developing applications with Dragonfly. In this blog post, we explore the 3 common caching problems and best practices to mitigate them.
The landscape of open source software is evolving in order to continue to thrive. Business Source Licenses are necessary for sustained innovation and are being embraced by leading open source innovators. This post breaks down why Dragonfly decided on BSL from the start.
In the Developing with Dragonfly series of blog posts, we will explore different techniques and best practices for developing applications with Dragonfly. We start with one of the most common usages: Cache-Aside.
Wrestling with BigKeys is hardly an issue when using Dragonfly, the groundbreaking in-memory data store. You can use BigKeys in Dragonfly if they are necessary to your server applications, fearlessly.
In this blog post, we explore the seamless integration of Dragonfly as a drop-in replacement for Redis in Feast—an acclaimed feature storage and server project widely recognized in the machine learning domain. With just some simple steps, we unveil the outstanding compatibility of Dragonfly as an online store solution for Feast, ensuring a smooth transition without compromising existing functionalities.
This post explores the limitations of horizontal scaling in terms of cluster reliability, load distribution, and cloud over-commitment. It also outlines design decisions that were made to allow Dragonfly, a drop-in Redis replacement, to scale vertically in order to handle heavy workloads and large data volumes on a single instance. By adopting Dragonfly's vertical scaling capabilities, organizations can achieve improved performance, cost savings, and operational efficiency in their distributed systems.
This blog post discusses zero-downtime migration from Redis to Dragonfly using Redis Sentinel. It covers the benefits of Dragonfly, migration techniques, and the role of Redis Sentinel in ensuring seamless migration without service interruption. The post provides steps to configure Dragonfly as a replica and promote it using Sentinel.
This blog post covers monitoring in-memory datastores, focusing on Dragonfly. Learn how to use Prometheus and Grafana to monitor Dragonfly metrics and visualize them on a dashboard. Explore memory consumption, client-side metrics, server metrics and more.
This blog post discusses the limitations of Lua scripting in Redis and introduces Dragonfly as a drop-in replacement for Redis. Dragonfly addresses the challenges of long-running scripts and scalability, offering a vertically scalable, multi-threaded, and asynchronous architecture that improves performance and efficiency for Lua scripting in Redis.
Dragonfly, an in-memory database that can be a drop-in Redis replacement, now supports replication for high availability in its version 1.0 release.
In this blog post, you will learn how to migrate data from a Redis Cluster to a single-node Dragonfly instance. We will use a sample application to demonstrate the migration process and cover everything step by step.
Dragonfly offers a rate limiting API via the CL.THROTTLE command. This post outlines the rate limit algorithm background and how you can use it in your application.
We are thrilled to announce the latest addition to our in-memory data store - the Kubernetes operator for Dragonfly!
In this blog post, you will learn how to use Redis Lists to build a background processing pipeline with Dragonfly.
We are pleased to announce that Dragonfly 1.0, the most performant in-memory datastore for cloud workloads, is now generally available.
In this article, we will explain the main reasons why your Redis instance might fail, and provide advice to avoid this.
A thorough benchmark comparison of throughput, latency, and memory utilization between Redis and Dragonfly.
2022 saw the emergence of a new technology and database project Dragonfly as well as the founding of a new company (DragonflyDB) to shepard and evolve it.
Balance is essential in life. When our focus is limited to improving a single aspect of our life, we weaken the whole system.
Infrastructure should be boring. Boring is good. Boring means that it just works, and you don’t have to worry about it. A year ago, we went on a quest to build a boring in-memory store.
Dragonfly crossed the 10K GitHub stars milestone in just 75 days. What an incredible start for our journey!
I talked in my previous post about Redis eviction policies. In this post, I would like to describe the design behind Dragonfly cache.
We could not have predicted the events of the last days. In a single week, Dragonfly transformed from a dream to reality.
Let’s talk about the simplicity of Redis. Redis was initially designed as a simple store, and it seems that its APIs achieved this goal.
Following my previous post, we are going start with the “hottest potato” - single-threaded vs multi-threaded argument.
During the last 13 years, Redis has become a truly ubiquitous memory store that has won the hearts of numerous dev-ops and software engineers.