Question from Ignite-Best practices around RAG with embeddings #9842
Closed
chrisbardon
started this conversation in
General
Replies: 2 comments
-
Tagging @westey-m for visibility. |
Beta Was this translation helpful? Give feedback.
0 replies
-
@chrisbardon, in general I haven't come across a lot of better alternatives, but here are couple of my thoughts:
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I was talking to @moonbox3 at Ignite last week, and I was asking about MS recommendations around RAG best practices, and whether there was a better approach than the basic semantic embeddings we've all been using. The embedding model is going to calculate based on a single version of semantics, but my question was whether that was reliable in all cases. Obviously something like a hybrid search improves this with things like traditional keywords (like in Azure AI Search), but the question was really whether there was a better way to use semantics in things like a classifier taxonomy, or to use multiple embedding models to extract different representations of semantics. I did see a really good session on advanced RAG and llamaindex that gave me some ideas, but the central problem still remains that a semantic embedding really only represents one thing, so the best practice seems to be to make those things as small as possible (single topic document chunks).
I've got some ideas of things to try, but don't necessarily want to reinvent the wheel, so I thought I'd ask around to see if there were any recommendations from the team first.
Beta Was this translation helpful? Give feedback.
All reactions