Mehdi Jamei’s Post

View profile for Mehdi Jamei, graphic

Director of AI @ System Inc. | PhD, genAI/ML

I've noticed more companies stepping back from building RAG systems. Reconstructing a decades-old discipline like information retrieval isn't easy, even with advanced LLMs. Additionally, evaluating and monitoring RAG pipelines remains challenging and unresolved. An often overlooked point: you don't necessarily need vector search to build a #RAG. Traditional methods like #BM25 can be more effective, consistently performant, and much cheaper. While retrieving certain facts from a collection of documents is fairly straightforward, the most ambitious promise of RAG—answering open-ended questions grounded in a large body of multimodal content—remains unfulfilled. This is challenging because relevant documents might not be semantically close to the question in the embedded space. A potential solution could be using LLMs' internal knowledge to "seed" various searches, combine the results, and then use LLM reasoning to sift through a large context to refine and answer the question. Is anyone exploring this approach?

  • No alternative text description for this image
Ajit Patil

Executive Leader | Data Polymath | Payments | AI Strategy | Engineering | Product | Innovation | People

2w

Would using LLM for problems which cannot be solved by traditional means would be a good start. Also LLM needs to be an enabler as you pointed out instead of the end state

Like
Reply

To view or add a comment, sign in

Explore topics