the-case-against-vector-databases
the-case-against-vector-databases
vec3.ai 2
Even VCs like Sequoia admit:
infrastructure is getting overbuilt.
They see this with GPUs - vector dbs are much easier to build in comparison
vec3.ai 3
VCs see vector databases as an investment in “picks and shovels” for AI,
with a proven business model (database).
💸
With so much funding (thus sponsored content)
and with all the hype around AI
- it's easy to assume you need a vector db in your AI project.
vec3.ai 4
Why you don’t need a vector db:
2. You don't have enough data to use it anyway (see next slides)
3. Information retrieval is not your core focus, and it’s better to
integrate with out-of-the-box solution
vec3.ai 5
Way too fast…
vec3.ai 6
Alternatives sufficient for most data needs:
1. LLM alone can fit all your data in - no need for vector search
2. Exhaustive vector search (brute-force)
vec3.ai 7
Vector database comes with a cost.
vec3.ai 8
Hidden costs of vector databases:
1. Yet another database to maintain
2. Need to sync data with other dbs
vec3.ai 9
👉 Always verify your users' needs first
before proceeding with an ambitious implementation :)
vec3.ai 10
Vector search is an optimization.
As engineers often say “premature optimization is the root of all evil”.
vec3.ai 11
What vector search tries to optimize
🔎
keyword search
🧠LLMs
vec3.ai 13
Why keyword search rocks! 🎸
1. Often performs better than vector search
2. Generalizes well to unseen, out-of-domain data
3. Search mechanics:
vec3.ai 14
Vector search needs an embedding model
Vector search relies on a neural net that encodes the data into vectors:
• a generic model can perform worse on data from a narrow domain
vec3.ai 15
Embeddings are inherently limited
vec3.ai 16
Most advanced solutions usually combine
different techniques
🔎
keyword search
+ 🧠LLMs
+ 🗂️ ↗
vector search
= HYBRID SEARCH
1. Make sure to start simple, with the right components.
2. Focus on optimizing whatever provides the best boost to overall accuracy.
vec3.ai 17
Vector dbs x AutoGPT: an overkill solution
Let's say LLM call takes 10 seconds, You reach 100k embeddings after 11 days.
1 embedding is generated every 10s. Even then, brute-force vector search
(np.dot) takes milliseconds
np.dot() np.dot()
1 ms <100 ms Embeddings np.dot time AutoGPT time AutoGPT cost
then after 100k calls…
10 s 10 s 1 <1ms 10s $0.27
LLM call LLM call LLM call LLM call LLM call LLM call 10k <10ms 27h $2.7k
Startups usually build value where no one else already had (blue ocean strategy).
Most novel value can be added through adopting generative LLMs.
vec3.ai 19
AI strategy - seek 10x improvement
In the example, improving company search
keyword search 50%
with vector search yields 25 pp gain.
vector search (fine-tuned) 60%
hybrid search 75% Building sales automation with LLMs would
sales automation 30% be much more disruptive compared to
previous methods.
sales automation with LLM 80%
0% 20% 40% 60% 80% 100%
% of automation, made-up data serving as an example
Some use cases might benefit much more from the current LLM revolution.
vec3.ai 20
Future research 🔬
There's only so much meaning you can squeeze into a vector.
On the other hand, generative LLMs will keep getting better.
vec3.ai 21
Think about your users' needs.
Simpler = better.
vec3.ai 22
If you found this page helpful, go ahead and share it with friends.
@Dariusz Semba
vec3.ai 23
Sources / further reading
1. Vector Search with OpenAI Embeddings: Lucene Is All You Need paper
2. SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking paper
3. On Hybrid Search by Qdrant
4. Beware Tunnel Vision in AI Retrieval by Colin Harman
5. Emerging Architectures for LLM Applications by a16z
7. Auto-GPT Unmasked: The Hype and Hard Truths of Its Production Pitfalls by Jina.AI
8. Why AutoGPT engineers ditched vector databases by Dariusz Semba
9. Introducing Natural Language Search for Podcast Episodes by Spotify
10.Why You Shouldn’t Invest In Vector Databases? by Yingjun Wu
vec3.ai 24