In the short amount of time that I've used AWS Bedrock and other related AWS AI/ML tools for professional work, I've learned that AWS does not particularly excel in the vector search field. There's a few things that I have used with AWS for AI/ML purposes, with varying degrees of results. The common thread between all of the options is that while they're good, they're not great. For example: it's not obvious that a lot of the Bedrock models (outside of Claude) are pretty mediocre. OpenSearch, while technically being usable for RAG, is more of a general search engine and can fit other purposes like log analytics and full-text search (and therefore does not cater to the particularities of RAG systems). Although I can tolerate Amazon for work, I would much prefer a simpler solution that lets you keep working instead of having to constantly fiddle with knobs.
Enter Qdrant, a vector storage database that aims to excel at high-performance vector search at scale. Compared to other solutions (and of course the previously mentioned AWS offering), I found their offering to be pretty compelling:
fastembed
library)I realise that playing the Rust card here, as someone who is already heavily invested in Rust, is not particularly imaginative. However, the fact remains that Rust is likely to continue to see future use in database programs. SurrealDB and TiKV (a part of the TiDB project by PingCap) both use Rust under the hood, with Databend and ScyllaDB using Rust in some part. It should be mentioned of course, that Qdrant has database client libraries in languages other than Rust (notably Python).
In my opinion, Qdrant makes vector search extremely easy (and is really good at it!). All you need to do is configure and create a collection (selecting things like the embedding dimension size and search method, as well as whether to store embeddings on disk or in memory). Then you insert embeddings alongside relevant JSON payloads and you're basically ready to go! If you have used Weaviate or Pinecone, it is mostly the same process. They also have a number of useful examples to help you get started, utilising a range of other AI/ML services and libraries respectively.
Because Qdrant is simple and requires minimal configuration, you can get started extremely quickly with no need to navigate the complexities of something like AWS (IAM roles, anyone?). The APIs are also pretty intuitive - although the APIs are there if you need the raw endpoint, I've found the client libraries to be quite comprehensive. Batch inserting, collection management and complex filtering for vector retrieval can all be done from the library. The Rust one is faster of course, but that's really the only difference.
You can also store multiple embeddings under a single embedding record so that you can search for all of them at the same time. For example, if you have two relevant snippets (like a code snippet and a JSON payload snippet that contains metadata about the code snippet) that go together and you want your application to search for cosine similarity for both of them, it is quite simple to just put them together in an array and then upsert the relevant embeddings.
Their free tier is also pretty generous and I have found that even after inserting quite a few embeddings, I'm still within the free tier bounds. As previously mentioned of course, you can also run Qdrant's Docker image should you want to try things locally. However, for web applications that I've created in the past, it works pretty well for the use cases I have. If I were to use Qdrant for a business use case though, I'd be happy to do so. Judging by this Reddit thread, Qdrant is the cheapest in terms of pricing and it's not close, except for the pricing at 20M vectors where it comes out a bit more expensive than Pinecone normally, but is much cheaper for high perf).
The Qdrant team are also quite active on their Discord server. Whether this is a benefit or not mostly depends on whether or not you use Discord, but if you have any questions you can always drop a mention and they tend to respond pretty quickly. They also hold some of their events there so if you want insider insight, that's the place to be.
This answer boils down to a few things, currently:
Pinecone being closed-source is neither here nor there, but having to register just to do anything is already a non-starter for me personally.
Although vector databases are quite specialised, they're very useful for the right use case - and Qdrant is no exception. For engineers who primarily work with (or want to work with!) Rust, using Qdrant is mostly a no-brainer as it's super easy to get started. Interested? Check out Qdrant and try building something cool with it!
You can also, of course, check out the Shuttle articles I've written on Qdrant if you want to follow a tutorial: