Generate LLM Embeddings On Your Local Machine

แชร์
ฝัง
  • เผยแพร่เมื่อ 23 ม.ค. 2025

ความคิดเห็น • 22

  • @moumniable
    @moumniable ปีที่แล้ว +6

    i just love how diverse your videos are ! even when i don't particulary look for something your videos drives me to learn more. thanks ❤

  • @rons96
    @rons96 ปีที่แล้ว +10

    Not bad, but if i may say any tip, i would say to use a sentence-transformer from hugging face for embeddings and then use a llama like llm only to customize the answer, because models created just for embeddings seems to be more accurate for this task. Also, langchain module is easier and useful than using numpy and requests, with lot more features. I use this setup most for RAG and seems to work pretty well.

    • @henrischomäcker
      @henrischomäcker ปีที่แล้ว

      Very interesting hint. Why would you suggest to use a sentence-transformer from hugging face for creating the embeddings instead?
      Which method does a sentence-transformer from hugging face use to create the embeddings and which one does ollama use?

    • @hackdonalds
      @hackdonalds 11 หลายเดือนก่อน +2

      I tried llama2 and mistral embeddings through ollama embeddings api. The similarity search results were sht compared to Xenova/all-MiniLM-L6-v2 or gte-small

    • @rons96
      @rons96 10 หลายเดือนก่อน

      @@hackdonalds yes, llama for embeddings is not good, with sentence-transformers i mean that one you mentioned, then use llama to elaborate the sentence. There's another model better for embeddings but it will require more resources and i don't remember the name now.

    • @rahulmakwana663
      @rahulmakwana663 5 หลายเดือนก่อน

      @@rons96instructorembedding

    • @Midicifu
      @Midicifu หลายเดือนก่อน

      Hey thanks for this (you save me some time) I want to use a hugging face model but I am using dotnet the model i want to ise is sentence-transformers/distiluse-base-multilingual-cased-v2, this model has many 539mb files in different extensions, I dont whant to use python i want to keep my project in c# so, how can i do this? Thank you for your time!

  • @godwinntowdanso4111
    @godwinntowdanso4111 7 หลายเดือนก่อน

    Spot on. Simplified presentation

  • @mohammadalibazyar5079
    @mohammadalibazyar5079 10 หลายเดือนก่อน

    thanks, bro... really helpful ❤

  • @EliSpizzichino
    @EliSpizzichino 10 หลายเดือนก่อน

    that's very interesting! I imagine you can build your local knowledge base in this way...
    I need to make one for code-snippets that store knowledge bits find around....
    Is `d` dimension fixed by the model? does it mean I have 4096 bytes to store my embedding?

  • @JuanDiegoSalamanca-oy6xs
    @JuanDiegoSalamanca-oy6xs 9 หลายเดือนก่อน

    if you do it in Colab what url do you use?

  • @ddschaefer
    @ddschaefer ปีที่แล้ว

    Great video! But where comes faiss into play?

  • @peterparker5161
    @peterparker5161 8 หลายเดือนก่อน +2

    I tried this with LLAMA3 8b locally. It can work if the sentences are short enough. But when I started plugins in long paragraphs (youtube transcripts) it becomes basically useless. Transformers that are creating for embedding (BERT for example) seems to work better. They also have lower computational cost compared to LLAMA. I tried again with "nomic-embed-text-v1.f16.gguf" and it works much better.

  • @Darkev77
    @Darkev77 ปีที่แล้ว +1

    Powerful video! Guys, anyone knows how I can generate these embeddings if I were to deploy my app remotely?

  • @iamreallybadatphysics
    @iamreallybadatphysics ปีที่แล้ว

    Great video! 😃

  • @ChristopherBruns-o7o
    @ChristopherBruns-o7o 7 หลายเดือนก่อน

    Cool guide. Very well explained. +1

  • @Usman_ekram
    @Usman_ekram ปีที่แล้ว

    First comment ❤ and ur videos are fantastic

  • @dangalimov7435
    @dangalimov7435 ปีที่แล้ว

    Brilliant!

  • @roberthenry7283
    @roberthenry7283 5 หลายเดือนก่อน +1

    where is the source code

  • @JonathanVisionAssets
    @JonathanVisionAssets ปีที่แล้ว

    🔥

  • @agastronics
    @agastronics ปีที่แล้ว

    comments[-1] #FirstComment😅