J. Michael Dockery’s Post

View profile for J. Michael Dockery, graphic

LegalTech AI | Angel Investor | Former BigLaw Partner

Was a bit surprised to see Wired touting RAG as "taking hold in Silicon Valley" like it's the hot new thing - RAG has been around for ages (at least in LLM terms) and I actually think its application space is currently shrinking as context windows grow. Their "reduce AI hallucinations" thing is not really the whole story - I mean, yes, RAG can definitely help in the sense that asking a broad question to the base model (with no input other than the prompt) is likely to result in hallucinations, whereas feeding in concrete data (with RAG or otherwise) will ground the response in that data and therefore produce fewer hallucinations. But RAG is really just a tool for feeding that data - and, yes, for massive search tasks (like broad legal research requests across entire databases), RAG may be the best (if not only) way to do it. But for smaller tasks (like analyzing a single document) it's often better to just feed the input in directly, which the larger context windows of the newest models increasingly allow. I feel like a few months ago (around the time Opus was released) people were asking whether RAG would be obsolete - that was an overreaction, but still it doesn't really fit this narrative of RAG as some hot new thing in the LLM world. Maybe I missed the pendulum swinging the other way again? Not sure. https://lnkd.in/eUEwBmFf

Reduce AI Hallucinations With This Neat Software Trick

Reduce AI Hallucinations With This Neat Software Trick

wired.com

To view or add a comment, sign in

Explore topics