Thank you for the video, that was a great paper to go through. I find RAG research techniques have so much insight to how we can develop and identify our own cognitive impediments to our own judgement. The Comprehensiveness, Diversity of perspective, Empowerment and Directness is such a good mental model to use in our own human judgement.
same here brother. There are so many techniques, everyday I learn something new which is both good and terrifying ha. What stack are you thinking of using ? We are researching DataStax, Pinecone, Weaviate and are learning to build agents with LangGraph.
great presentation Vivek. Some questions: - is graphRAG production ready ? if not, would it be difficult to upgrade RAG methods once we are in production ? - is there a RAG provider/stack that you prefer ? (datastax, pinecone, weaviate + a bunch of others who are all competing for attention) - what are your thoughts on LangChain vs LangGraph ?
can you pontificate on the combination of upcoming transformer inference ASICs with deep agentic workflows employing GraphRAG style strategies? Seems like we will be close to our personal assistants writing a PhD thesis in the background whenever we ask a question. SOHU is reporting 500,000 tokens per second with Llama3 70B....
wouldn't it cut to the chase to train an llm on your own data? theres your graph use one of these OpenAI's GPT-3/4 Hugging Face Transformers (e.g., GPT-2, GPT-3 via third-party providers) Google's T5 (Text-to-Text Transfer Transformer) Meta's BART and BlenderBot Anthropic's Claude every week update the llm summarization is the death of real data, better off one level of summarization? Just a thought!
Thank you for the video, that was a great paper to go through. I find RAG research techniques have so much insight to how we can develop and identify our own cognitive impediments to our own judgement. The Comprehensiveness, Diversity of perspective, Empowerment and Directness is such a good mental model to use in our own human judgement.
Really great explanation, I’m currently struggling to decide on my own KG structure for a 2000 medical pdf corpus, so this was very helpful
same here brother. There are so many techniques, everyday I learn something new which is both good and terrifying ha. What stack are you thinking of using ? We are researching DataStax, Pinecone, Weaviate and are learning to build agents with LangGraph.
Excellent content summarizing cutting edge approaches, thank you!
Very interesting paper, thanks for covering!
great presentation Vivek. Some questions:
- is graphRAG production ready ? if not, would it be difficult to upgrade RAG methods once we are in production ?
- is there a RAG provider/stack that you prefer ? (datastax, pinecone, weaviate + a bunch of others who are all competing for attention)
- what are your thoughts on LangChain vs LangGraph ?
Thanks so much, great content.
very well "ragged"... both on the local domain (details) and global domain (overview of pros-cons) 😉😎
can you pontificate on the combination of upcoming transformer inference ASICs with deep agentic workflows employing GraphRAG style strategies? Seems like we will be close to our personal assistants writing a PhD thesis in the background whenever we ask a question. SOHU is reporting 500,000 tokens per second with Llama3 70B....
super excellent video
Why not just use RAPTOR instead?
wouldn't it cut to the chase to train an llm on your own data?
theres your graph
use one of these
OpenAI's GPT-3/4
Hugging Face Transformers (e.g., GPT-2, GPT-3 via third-party providers)
Google's T5 (Text-to-Text Transfer Transformer)
Meta's BART and BlenderBot
Anthropic's Claude
every week update the llm
summarization is the death of real data, better off one level of summarization? Just a thought!
Maybe you don't want to send all your valuable business data to third party companies.
@@mccleod6235 thats true but its not necessary, there are models that are open source you can train air gapped from a jetson
"every week update the llm"
yeah no way unless you have tons of cash, compute, and time
maybe maybe not, let you know! your probably right, will see if my idea pans out