What are the different types of models - The Ollama Course

แชร์
ฝัง
  • เผยแพร่เมื่อ 9 พ.ย. 2024

ความคิดเห็น • 25

  • @blocSonic
    @blocSonic 4 วันที่ผ่านมา

    Oh wow! I completely misunderstood what "instruct" models are for and was avoiding them when, in fact, that's what I need. Thank you!

  • @solyarisoftware
    @solyarisoftware หลายเดือนก่อน +1

    Thanks, Matt. Your Ollama course is great because it's easy to follow and addresses problems from your unique point of view. Always upvoted!
    Regarding naming, what you call "source" models are also referred to as "foundation" or "pretrained" models, as far as I know. It's a good distinction between chat-fine-tuned models (sometimes called chat-completion models) and instruct-fine-tuned models (sometimes called text-completion models).
    In general, custom fine-tuning a model involves taking a source model and refining it with custom data. This feature is not currently supported by the present version of Ollama, even though you’ve rightly dedicated one or two videos to how to create a custom fine-tuned model by training an original source model.
    Regarding multimodal models, as you mentioned, Ollama includes some vision LLMs (image input) like LLava and others, I believe. You correctly pointed out that multimodal could also involve audio input (and output), which seems feasible at the moment (I’ll need to double-check by example the new released Mistral Pixtral when available on ollama). BTW I think video processing using Ollama is also of great interest, so it might be worth exploring this topic in future videos.
    Just my two cents-thanks again!

  • @jimlynch9390
    @jimlynch9390 หลายเดือนก่อน +1

    Once more, I learned something. i've asked that question before but never have gotten a satisfactory answer. Thanks, Matt.

  • @NLPprompter
    @NLPprompter หลายเดือนก่อน +1

    Thank you for keep caring poor learner. Thanks Matt

  • @claudioguendelman
    @claudioguendelman หลายเดือนก่อน +4

    Excellent thanks from Chile.

  • @Alex-os5co
    @Alex-os5co หลายเดือนก่อน +1

    Awesome course, thank you! My only request would be to have mentioned the suffix Q_M Q4 etc

  • @tonyhartmann7630
    @tonyhartmann7630 หลายเดือนก่อน +1

    Thanks for the explanation 😊

  • @akamalov
    @akamalov หลายเดือนก่อน +2

    you rock! Thank you!

  • @bernieoconnor9350
    @bernieoconnor9350 หลายเดือนก่อน

    Thank you, Matt. Great info.

  • @therobotocracy
    @therobotocracy หลายเดือนก่อน

    Nice, well done.

  • @bobdowling6932
    @bobdowling6932 หลายเดือนก่อน

    Would you regard the NER models as cases of fine-tuned models?

  • @utvikler-no
    @utvikler-no หลายเดือนก่อน

    Thanks

  • @bustabob08
    @bustabob08 หลายเดือนก่อน

    Tak!

  • @ISK_VAGR
    @ISK_VAGR หลายเดือนก่อน

    Thanks Matt, great video and series! Why don’t LLMs always produce good embeddings? And why do embedding models sometimes underperform in RAG applications? I’ve tested many models, but only five have consistently provided accurate embeddings for paper abstracts, verified by clustering and ground truth.

  • @deepanshusinghal955
    @deepanshusinghal955 หลายเดือนก่อน

    How can I get the fast answers and accurate at the same time using Ollama ?????

  • @muraliytm3316
    @muraliytm3316 หลายเดือนก่อน

    Hi sir your videos are great and very informative and I really like them but could you please explain some of the concepts by sitting before a pc and show them practically, I am really confused what model to download, the benchmarks show good results and when I really use them they are worse and also there are different quantisations like q4,q6,q8,fp16,K_S,K_M,etc which are difficult to understand. Thanks for reading the comment

    • @technovangelist
      @technovangelist  หลายเดือนก่อน +1

      There is another video in the course that shows the quants

  • @kappaj01
    @kappaj01 19 วันที่ผ่านมา

    Question - with Ollama now supporting the GUFF file format from Huggingface, can you run the video models locally with Ollama? Have not tried it yet...

    • @technovangelist
      @technovangelist  19 วันที่ผ่านมา

      Ollama has supported the gguf file format since the day it was created. There was ggml before, but gguf is the only thing ollama can work with. You have been able to download any gguf from huggingface and use it in ollama for at least a year. So not sure what you are asking. Which video model?

    • @kappaj01
      @kappaj01 19 วันที่ผ่านมา

      @@technovangelist Thanks for the response. Around 5:40 you mentioned that Ollama does not support video yet. Is that an Ollama restriction or the model that was downloaded?
      Appologies for being unclear.

    • @technovangelist
      @technovangelist  19 วันที่ผ่านมา

      got it. its mostly a restriction of the available models. There aren't any that do a good job of reviewing a video that are a format that llamacpp or ollama can support. the models I have seen that can do that only support a second or two of video and that would need a lot of memory.

    • @kappaj01
      @kappaj01 19 วันที่ผ่านมา

      @@technovangelist Ahh - thank you very much! Awesome series - keep up the good work. Learning heaps and bounds.

  • @tecnopadre
    @tecnopadre หลายเดือนก่อน

    I always think why isn't a model that only talks and can be trained with information (like FAQ Help desk or company internal bot) and of course it's an small one and answers properly without hallucinations

    • @azrajiel
      @azrajiel หลายเดือนก่อน

      @@tecnopadre there is the so called noun phrase collisions,, which are seemingly a big part in hallucinations, even in rag systems. basically the problem is not inaccurate data but reference nouns that are ambiguous. there are some very interesting articles to google and also some work to eliminate them. basically it can be corrected with the right prompting.