A gentle introduction to RAG (using open-source models)

แชร์
ฝัง

ความคิดเห็น • 44

  • @acastanob
    @acastanob 3 หลายเดือนก่อน +3

    Came here out of curiosity, and ended up watching the full video. Thanks for taking the time to explain the very basics. Learned a lot!

  • @felipeclavijo1736
    @felipeclavijo1736 3 หลายเดือนก่อน +3

    YOU ARE JUST INCREDIBLE !!!!!!!!!! keep them coming. you are pretty much my main teacher.

  • @RajaRahamathullah04
    @RajaRahamathullah04 3 หลายเดือนก่อน +1

    This was amazing! Thank you very much for all the hard work - you’re incredible! Keep up the great content. My humble request to continue with UI interface for this RAG application.

  • @ic4roswings
    @ic4roswings 3 หลายเดือนก่อน

    You are the man, I have watched god knows how many videos about rag and i finally get it, Thank you very much

  • @nicoloferrari5909
    @nicoloferrari5909 2 หลายเดือนก่อน

    That's incredible how good you're at explaining this though argument! Thanks a lot for your work. Really appreciate it

  • @nmstoker
    @nmstoker 3 หลายเดือนก่อน

    Great video and nice that it's possible to run entirely locally, all with open source 🎉

  • @gisdevelopr
    @gisdevelopr หลายเดือนก่อน

    Very beautiful explain each step and make it so simple to understand, thanks for providing this video.

  • @iiboo1
    @iiboo1 3 หลายเดือนก่อน

    Fantastic! You've really made my day by explaining it so clearly. Thank you!

  • @NumericLee
    @NumericLee 3 หลายเดือนก่อน +2

    outstanding. for next video, I would love to see how LLMs are applied to mine unstructured data

  • @R-f3u6z
    @R-f3u6z 2 หลายเดือนก่อน

    Thanks for the fantastic explanation!

  • @pedrolima-lr3lu
    @pedrolima-lr3lu หลายเดือนก่อน

    Best explanation !
    Thank You

  • @sam-uw3gf
    @sam-uw3gf 3 หลายเดือนก่อน +2

    Great video

  • @マーベリックテック
    @マーベリックテック 3 หลายเดือนก่อน

    Absolutely loved it, Thank you for your efforts 🙂

  • @iamliam1241
    @iamliam1241 3 หลายเดือนก่อน

    Very clear and useful. Thank you!

  • @davidtindell950
    @davidtindell950 3 หลายเดือนก่อน +1

    Thank You. I have done many different RAG apps. A future vid suggestion: comparing results from completely local RAG to using remote embeddings, say OpenAI. I am finding that the remote embeddings are consistently better quality and the Q&A responses are better. To me, it is pointless to do a RAG app if the embeddings are poor and the answers are mediocre !

  • @sirishkumar-m5z
    @sirishkumar-m5z 3 หลายเดือนก่อน

    RAG is a powerful tool for working with open-source models. It's a good idea to explore alternative tools as well, ensuring you choose the best fit for your specific needs.

  • @andrewsperspective
    @andrewsperspective 3 หลายเดือนก่อน

    This was perfect thank you

  • @mehershahzad-n5s
    @mehershahzad-n5s 3 หลายเดือนก่อน

    You well explained RAG

  • @Lucky-op7qz
    @Lucky-op7qz 3 หลายเดือนก่อน

    superrr amazinggggg,explaination

  • @uwegenosdude
    @uwegenosdude 3 หลายเดือนก่อน +1

    Thanks a lot for your very interesting video. It's great that all the time your code is working out of the box. Only langchain-ollama was missing in the requirements.txt. And unfortunately faiss-gpu is not supported on Windows 11 (AFAIK?). Great stuff you are offering to all of us all the time. Your explanations are always so good to understand. Amazing !! Please keep going !

  • @Nabad077
    @Nabad077 3 หลายเดือนก่อน +1

    You are really really good

    • @underfitted
      @underfitted  3 หลายเดือนก่อน

      🙏🏻

  • @Ali-km8kn
    @Ali-km8kn หลายเดือนก่อน +1

    Great video and explanation! Thank you. I've a question. Will the context variable be inputted to the model through the prompt as embeddings of each page of the four pages or it will be converted back to string? Thank you in advanced.

  • @kamalkhanal409
    @kamalkhanal409 3 หลายเดือนก่อน +1

    Love the video! Could you please create a video showing how to export a Jupyter notebook into a proper project structure and deploy it on the cloud?

  • @bald_ai_dev
    @bald_ai_dev 3 หลายเดือนก่อน +4

    is it better to use specialized embedding models like nomic-embed-text or llama3.1 itself as an embedding model?
    also can you please do a tutorial on some of the major rag ideas like building a self correcting rag (CRAG) and the compare the results with naive rag using an evaluating framework like ragas, giskard etc?

    • @HassanAllaham
      @HassanAllaham 3 หลายเดือนก่อน +2

      This is a very good comment and a very useful request.. I would like the model to respond not just with the answer but also with the source of that answer (file name + page number = make sure the model is not drunk one 🧐). I believe this would be great if we add a re-ranker model 💯

    • @rayofvictory
      @rayofvictory 3 หลายเดือนก่อน

      @@HassanAllaham that will be great! @underfitted can you please chime in?

    • @kamalkhanal409
      @kamalkhanal409 3 หลายเดือนก่อน

      @@HassanAllaham For that, you need to keep the retriever output in a variable or a list while executing.

  • @thedevmachine
    @thedevmachine 4 วันที่ผ่านมา

    Santiago, I loved the video. Very clear explanation. I have a question. As you know, there is a limit to passing a prompt. For example, if I want a summary of a whole document, in theory, I have to pass the whole document to the LLM so it can create a summary of it. But this won’t fit in the context window. Chatgpt has I think 128K limit on the api but OLLAMA does not have this I think. Also I have no idea if 128K is enough for any LLM's. If I already stored a large document on my vector database how could I pass the whole document to LLM to summarize it? I cant just add whole document in the prompt. Thanks

  • @sharangkulkarni1759
    @sharangkulkarni1759 3 หลายเดือนก่อน

    good video

  • @JohnSanabria
    @JohnSanabria 3 หลายเดือนก่อน

    It is possible to run the jupyter notebook on Google Colab? How it could be?

    • @bharathjpv9334
      @bharathjpv9334 2 หลายเดือนก่อน

      Upload it from local

  • @fushumang1716
    @fushumang1716 3 หลายเดือนก่อน

    next video is how to query if documents have images. Can LLMs describe or get context from images

  • @papalevies
    @papalevies 3 หลายเดือนก่อน

    faiss-gpu only supports up to python 3.10, is there an alternative?

  • @AkshayRanchod-p4e
    @AkshayRanchod-p4e 2 หลายเดือนก่อน

    Apparently import langchain_ollama does not exist. I keep getting this error when trying to run the model

  • @treakfreak3466
    @treakfreak3466 2 หลายเดือนก่อน

    Hey there thanks , your videos are really helpful. I am student creating project around rag I want a video how can I make interface oriented or easy full stack rag bot without the large GPU

  • @mdsoykot2932
    @mdsoykot2932 3 หลายเดือนก่อน

    I'm facing an issue trying to install faiss-gpu on a Mac with an M3 Pro chip. Is anyone else having this problem?

  • @bharasiva96
    @bharasiva96 3 หลายเดือนก่อน

    Gravenberch was my motm

  • @ChronicleContent
    @ChronicleContent 3 หลายเดือนก่อน +3

    *"An unnecessarily complicated introduction to RAG that only works locally.". There I fixed it for you.

    • @rayofvictory
      @rayofvictory 3 หลายเดือนก่อน +1

      May I know what is unnecessarily complicated? He is taking the time to go step by step for users to scale this solution for our use cases.

    • @tecnopadre
      @tecnopadre 3 หลายเดือนก่อน +1

      People always have to criticize, whatever it is.​@@rayofvictory

    • @o_glethorpe
      @o_glethorpe 3 หลายเดือนก่อน

      If his explanation is too complex for you maybe this subject is not for you.

    • @ChronicleContent
      @ChronicleContent 3 หลายเดือนก่อน

      @@o_glethorpe I am not talking about me. This is supposed to be introduction. You don't need any of these to create a RAG.

    • @ChronicleContent
      @ChronicleContent 3 หลายเดือนก่อน

      @@rayofvictory you don't need langchain or any other of these tools to create a RAG. Also he mentions "a user employee asks" but all this is local so that's not true.