I wanted to extend my heartfelt thanks for the excellent session on how Retrieval-Augmented Generation (RAG) can be used to train Large Language Models (LLMs) to build expert systems in the retail, software, automotive, and other sectors. Your explanation was incredibly clear and insightful, making a complex topic easily understandable. I truly felt like Dr. Watson listening to Sherlock Holmes unravel the mysteries of the universe, marveling at the clarity and depth of the information presented. Your efforts in breaking down the concepts and applications of RAG in such a straightforward manner have left me feeling both enlightened and excited about the potential this technology holds for our industry. Thank you once again for your time and for sharing your expertise. I look forward to exploring and implementing these innovative solutions in our own projects
How do you handle terabytes of enterprise data, just do embedding groups? Should you generate sub questions first? How do you handle large amount of users?
@@shubhamsharma5631 I challenge you to find it. That is simply a link to the general github which is convoluted , not the exact notebook which is difficult to find.
Hello Jeff, could be that there is a mistake 24:00 in a for loop instead of “for i, s in enumerate(texts + table_summaries + image_summaries)” should be “for i, s in enumerate(text_summaries + table_summaries + image_summaries)”
Wat about Copyright and Ethical issues? How much do u guys charge for using ur model? And as per IBM and Oracle embeddings are nothing new so why use urs?
im not sure if there are existing libraries to do that, maybe check docs. although here's my intuitive approach. video is basically series of images with some history/context attached to previous and subsequent frames. so if you keep that history across frames intact by either providing previous frames as input, or keep a local vector of it all, you can make it work. not sure if its the best approach, but i m open for discussion
Check out all the AI videos at Google I/O 2024 → goo.gle/io24-ai-yt
IO2024_Multimodal_RAG_Demo.ipynb can't find this notebook
Pretty useless video without sample code.
I wanted to extend my heartfelt thanks for the excellent session on how Retrieval-Augmented Generation (RAG) can be used to train Large Language Models (LLMs) to build expert systems in the retail, software, automotive, and other sectors.
Your explanation was incredibly clear and insightful, making a complex topic easily understandable. I truly felt like Dr. Watson listening to Sherlock Holmes unravel the mysteries of the universe, marveling at the clarity and depth of the information presented.
Your efforts in breaking down the concepts and applications of RAG in such a straightforward manner have left me feeling both enlightened and excited about the potential this technology holds for our industry.
Thank you once again for your time and for sharing your expertise. I look forward to exploring and implementing these innovative solutions in our own projects
Impressive technology. Look forward to using it for my project.
Could'nt find the exact notebook used here.
will there be an opensource version of this, or atleast a paper
WHERE IS THE SAMPLE CODE??????? This is very frustrating to showcase but not share code.
@googledevelopers I second this comment.. could you please share that notebook?
agree
How do you handle terabytes of enterprise data, just do embedding groups? Should you generate sub questions first? How do you handle large amount of users?
Why do you use multimodal embedding model if you summarize images and ground them into text?
When I use RAG, Am I sharing my data with the model/company? or is it private with an extracost?
Rag is an architecture i believe. with out without it - whatever happening to the data same applies
Not necessarily. You can keep the data local. You only use the LLM for it's ability to summarize and generate responses as well as queries
Thank for sharing👍
where is this notebook in the cookbook repo?
Same question
33:18
I have a notebook in Kaggle named Multimodal RAG Gemini - should help, YT removing links for some reason.
@@Chitragar thank you
@@shubhamsharma5631 I challenge you to find it. That is simply a link to the general github which is convoluted , not the exact notebook which is difficult to find.
Hello Jeff, could be that there is a mistake 24:00 in a for loop instead of “for i, s in enumerate(texts + table_summaries + image_summaries)” should be “for i, s in enumerate(text_summaries + table_summaries + image_summaries)”
Nice, but why don't you develop a simple drag-and-drop RAG? e.g. I add a drive folder link and Google generates a RAG chat based on its content.
Is there a link to the python notebook? I'd love to play with it!
Is "unstructured" the best choice here for parsing PDF? Any better alternatives?
Llamaparse
langchain would be a good choice
Can I use the fine tuned Gemini RAG model via API from a mobile app?
Fantastic. Thank You!
where is the notebook?
Can someone please share the link?
Wat about Copyright and Ethical issues? How much do u guys charge for using ur model? And as per IBM and Oracle embeddings are nothing new so why use urs?
This example briefs about text and PDF, do we have any for video how de we use RAG, Vector store for Video can anyone give some reference
im not sure if there are existing libraries to do that, maybe check docs. although here's my intuitive approach. video is basically series of images with some history/context attached to previous and subsequent frames. so if you keep that history across frames intact by either providing previous frames as input, or keep a local vector of it all, you can make it work. not sure if its the best approach, but i m open for discussion
Can you please provide the source code? It would be great help!! Thank you!
Wonderful
dense embeddings are never enough for RAG system
You can get blue driver and get all error codes and example
Where is the code link?
github link please
33:18
@@shubhamsharma5631 can we run the code without subscription?
Ok but someone could literally look any of this up online or look for it in a manual, etc. w/out using AI...
Please share the colab link
Yes please do
33:18
Hmm... interestin...
Happy
Haha we just need your browsing history
🎉
After Muaadh Rilwan's post on LinkedIn
❤
🥺
This shows how garbage Langchain is as a library. Extremely verbose and intransparent.
Is there any framework better than Langchain?
can you tell me more about it please :)
So if I'm using the 2nd way, what's the name of the multidality-modal would be?