Discussion about this post

User's avatar
Daniel Manzke's avatar

You are tackling an interesting topic which I hit every day too. Especially if I read the next “RAG is dead”.

The main problem we have right now when working with knowledge (especially in enterprise), you don’t want to send the document every time to the llm to answer one question.

That’s just horrible inefficient.

There must be a way that keeps the document close to the llm (like your disk idea), so I can send several questions including chat history.

You mentioned curse how they do it. Do you have any interesting links?

Expand full comment
3 more comments...

No posts