Ollama UI - Your NEW Go-To Local LLM

แชร์
ฝัง
  • เผยแพร่เมื่อ 30 ก.ย. 2024

ความคิดเห็น • 249

  • @AZisk
    @AZisk 4 หลายเดือนก่อน +63

    This is definitely my local go-to now. What an amazing project.

    • @nickiascerinschi206
      @nickiascerinschi206 4 หลายเดือนก่อน

      Могу я там исп llmstudio?

    • @stickmanland
      @stickmanland 4 หลายเดือนก่อน +1

      Wow, you too?

  • @mrthp
    @mrthp 4 หลายเดือนก่อน

    man thank you this is the most comprehensive easiest way to install this. I had ollama before but with open webui it's a game changer. Do i need to run the app docker before each time or it will auto boot when computer starts ?

    • @darkskinnedpimp
      @darkskinnedpimp 4 หลายเดือนก่อน

      You can change the services that start with the computer by typing MSCONFIG after hitting Windows+R. If you see Docker in the list it will boot with computer

  • @galdakaMusic
    @galdakaMusic 4 หลายเดือนก่อน

    Is posible install this in a Rpi5 with Coral and use from multiple sources like Home Assistant or mi custom app?

  • @guinea_horn
    @guinea_horn 4 หลายเดือนก่อน +83

    As a contributor (merged one single PR 😊) but mostly a very early adopter of this project, I'm always stoked to see people talking about open-webui

    • @AZisk
      @AZisk 4 หลายเดือนก่อน +2

      great project! I just covered it on my channel too. congrats!

    • @aladinmovies
      @aladinmovies 4 หลายเดือนก่อน +1

      Nice

    • @fire17102
      @fire17102 4 หลายเดือนก่อน

      What was your PR about? ❤
      Watching this video makes me wanna go add some stuff too,
      Things I caught from Matthew: RAG loading indicator until files are avail. Get new models in UI rather than term.
      Things I wish to add: litellm, groq adding pipelines and flows not just prompts, adding #folder to context, easier installer that checks docker etc with Houston assistant. OpenInterpreter integration. OpenUI integration + Developer Mode, Tester Mode, Rapid Feedback, Obvsf API Pools, Import ChatGPT & Claude History, Stars, Topics (connected threads), Timelines, Search Hist, Actions, Prompt Refinements, Self Improvement, Skill Library, Image Gen & Editing. Song Gen, Incognito thread, global/group chats. Sign in from phone/multi device. Agents and Clients Dashs.

    • @snuwan
      @snuwan 4 หลายเดือนก่อน

      I have been using it for a while. A great UI

    • @jarad4621
      @jarad4621 4 หลายเดือนก่อน

      ​@@fire17102awesome yeah please add all that stuff

  • @retromancer8262
    @retromancer8262 4 หลายเดือนก่อน +21

    You can actually skip the git clone step. everything is contained in the docker image.

  • @annonymous3326
    @annonymous3326 4 หลายเดือนก่อน +4

    Wow I just built a virtual AI girlfriend using ollama. I'm trying it on Llama3 model, and recently migrated to one of the uncensored models. Good-bye wife, hello AI - LOL.

  • @garibacha
    @garibacha 4 หลายเดือนก่อน +15

    I was fishing yesterday... unbelievably, thank you! Ps.: Can you make a video about WebUI + Open Interpreter + LLM Local or LM Studio? Thanks

  • @AITester-j3u
    @AITester-j3u 4 หลายเดือนก่อน +31

    I've been with you for over a year and it's been amazing watching you dominate this LLM news space. For example your snake game has become a standard in the industry now!!! Like you, I'm constantly in the LLM lab and I'm constantly coming across your name with a quote regarding a large language model. Awesome job carving out a niche - MUCH LOVE FROM NEW ORLEANS 🔥💪

  • @jimigoodmojo
    @jimigoodmojo 4 หลายเดือนก่อน +16

    Few more cool features
    1. Image generation. I hooked this up to both local automatic1111 and DallE3 with api key. It's a bit of an odd workflow. You prompt it and the response will have a little pic button under it. I loaded a model finetuned for image prompts so response is cool.
    2. Hook up openai models for chat choices with your api key.
    3. Pull any llava model and you can hit the plus button to load a picture and ask questions about it.

  • @nachos_dimension3552
    @nachos_dimension3552 4 หลายเดือนก่อน +2

    As someone who is completely new, this is so damn confusing LOL first off, what is docker. Second, where are you and is this a windows terminal? Im extremely lost and maybe link your old videos?

    • @boomantv8271
      @boomantv8271 4 หลายเดือนก่อน +1

      I'm not sure if you've gotten any answers to your questions here, but I thought I'd try to help a bit where I can. To your first question, docker is a program that allows you to run software in something called a container. You can think of containers as a space in your RAM that is separate from the rest of the programs that are running on your computer. Docker is nice because it makes it easy for you to run things that are dependent on certain versions of Python , shared libraries, and other stuff without you having to worry about changing versions back and forth to make other things work on your computer. Secondly, from what I could see, buddy is on a Mac computer (the UI he shown at the top where there was a llama kind of was the giveaway) which unless he changed it to bash or something, means he is using a zsh terminal, but that's not important. If you are using a windows computer look up how to get WSL (a.k.a Windows Subsystem for Linux) on your computer and you should be able to follow most of the instructions.
      Helpful tips:
      Linux has different distributions but they fall into 3 main families- Debian(some people say Ubuntu but I'm my own brand of snob here), RedHat, and Arch. This is important to you as far as getting started working with linux because it'll determine your package manager. (Debian uses apt, RedHat uses dnf, and Arch uses pacman) they all have their own quarks as far as what they name things and how they function. But your going to want to ensure you have git installed in you WSL environment as well as having the most up-to-date version of things on there for security reasons to start off.
      Learn about the basics of bash shell.
      Learn the basics of git.
      There will be some differences as far as folder structure once the program is installed, but I think you're smart of enough to figure that out 😉. Hope this helps.
      Now go on out there and cry like the rest of us. 🫡

  • @marcelbloch9516
    @marcelbloch9516 4 หลายเดือนก่อน +3

    if you using docker you dont have to clone that git repo to run it.

  • @spencerezralow7136
    @spencerezralow7136 4 หลายเดือนก่อน +7

    Really awesome. This is something I’ve been looking for for a long time. The one I built myself is terrible.

  • @drivebuss8079
    @drivebuss8079 4 หลายเดือนก่อน +1

    Unsubscribed. This shit has been available for more than a year.

  • @raghuvallikkat3384
    @raghuvallikkat3384 4 หลายเดือนก่อน +2

    do we really need to do the git clone. I think it is enough we do docker run

  • @jlccVPServ
    @jlccVPServ 4 หลายเดือนก่อน +6

    I don’t usually post messages, but your video changed that. Very well done! I followed your steps, and within minutes, I had LLama3 running on Open-webUI with Docker on Ollama on a Windows computer. Thank you, Sir. Keep up the great work!

    • @peterbelanger4094
      @peterbelanger4094 3 หลายเดือนก่อน

      I HATE docker!!! this is all to confusing, grrrr! you all can do this, I'm stupid :(

  • @Maisonier
    @Maisonier 4 หลายเดือนก่อน +3

    How does it compare with Anything LLM?

    • @spillledcarryout
      @spillledcarryout 4 หลายเดือนก่อน +1

      Anything LLM is way clunkier, all models i have tried act like quite dumb

  • @user-tg6vq1kn6v
    @user-tg6vq1kn6v 4 หลายเดือนก่อน +3

    My go to channel for learning how not to write the game snake in python

  • @trezero
    @trezero 4 หลายเดือนก่อน +5

    It would be great to see a comparison between ollama and lm studio explaining the benefits and reasoningg of when to use each. The one thing i havent seen much of is how to leverage (if possible) other models from huggingface within ollama. This is easy to do in LM studio. For most other thigs i prefer ollama but i tend to use LM studio to test new models that ollama might not have readily available.

    • @xTheToolx
      @xTheToolx 4 หลายเดือนก่อน

      Same here. I prefer LM Studio and use it across networks hosted from a server. Not quite the same with ollama. Going to experiment with open-webui and see if it can connect to LMStidio with some code tuning.

    • @jarad4621
      @jarad4621 4 หลายเดือนก่อน

      Ask perplexity pro new model Bout it really good now based on llama 70b

  • @CM-zl2jw
    @CM-zl2jw 4 หลายเดือนก่อน +1

    Thank you. I just NEED to decide on which GPU to buy.
    Btw. Why do 6% of the people in your poll believe GPUs should be regulated?

    • @damienhughes4559
      @damienhughes4559 4 หลายเดือนก่อน +2

      Maybe 6% of his audience are OpenAI employees. 😂

  • @brunodangelo1146
    @brunodangelo1146 4 หลายเดือนก่อน +2

    What's the advantage of this over LMStudio?
    Not asking which one is better, but on what use cases they excel over the other.

    • @juanjesusligero391
      @juanjesusligero391 4 หลายเดือนก่อน +2

      For one thing, Open WebU is open source, which means that you'll be able to use it forever for free, for commercial or personal use (that and all the other advantages of it being FOSS, like more security, a nice and supportive community, etc.). That being said, LMStudio is easier to install.

  • @ronaldgaines336
    @ronaldgaines336 2 หลายเดือนก่อน

    I am new to LLM development. How do I get the documents(PDF) loaded and RAG setup in Ollama UI to be part of my development in AUTOGEN 2.0? If you load the documents with Ollama UI and I use Ollama as a local LLM for AutoGen, will I see the embeddings in my prompt data through AutoGen? Thanks!

  • @Cine95
    @Cine95 4 หลายเดือนก่อน +3

    Does it support image processing or is there anyway in which you could add image processing capabilities

    • @JoeBrigAI
      @JoeBrigAI 4 หลายเดือนก่อน

      Under settings there is a way to link this UI to an existing image generator like Automatic111 or ComfiUI.

    • @jimigoodmojo
      @jimigoodmojo 4 หลายเดือนก่อน

      Llava model allows you to ask questions about images. Hit + next to prompt field to upload pic. Then ask away. Wasn't great generating code from mock, but answered general questions ok.

  • @konradriedel4853
    @konradriedel4853 27 วันที่ผ่านมา

    Thanks for the vid, any Chance you May explain how to deploy such a chatbot for usage of customers or even implement or in Something Like c# Code?

  • @brianlink391
    @brianlink391 4 หลายเดือนก่อน +1

    Hey, seriously? I mean, I know the Walmart version is surprisingly fast, but come on, I'm using an RTX 3060 here! What's up with that?

  • @quatre1559
    @quatre1559 4 หลายเดือนก่อน +1

    can you link the video where you install docker.. i cant seem to find it..

  • @michai333
    @michai333 4 หลายเดือนก่อน +1

    How do you feel this compares to the newest LM Studio? It looks like it’s lacking on the advanced settings.

  • @lepton555
    @lepton555 16 วันที่ผ่านมา

    This thing needs docker engine unavailable for mac. Luckily, I'm a linux guy

  • @Termonia
    @Termonia 4 หลายเดือนก่อน +1

    Matthew, could be awesome a podcast with David Ondrej. He is inviting you in his videos. I'd love you both to talk about this tools. I'm using Anything LLM with LM Studio. What is the difference with Ollama with Ollama UI? thank you for your videos.

  • @SiliconSouthShow
    @SiliconSouthShow 4 หลายเดือนก่อน +1

    HEADS UP your first acct is your admin acct, if you update owu or ollama, or dkr, make sure you follow the instructions or it will lose your user name and password and you wont have admin anymore, and its a pain to straighten it out, doable, I did it, but a pain.

  • @realsoftgames7174
    @realsoftgames7174 8 วันที่ผ่านมา

    Are their any videos showing open we ui with agents, function calling and Web search?

  • @excessivelysalty_81
    @excessivelysalty_81 2 หลายเดือนก่อน +1

    Oh interesting, that documents feature I might have to try attempting to link it to all my Obsidian files. Ollama + Obsidian seems like a good match.

  • @Devsterinator
    @Devsterinator 4 หลายเดือนก่อน

    Does anyone know how to make the open-webui interface accessible from my phone on the same network? I'm used to being able to add --listen to a gradio webui and access it anywhere on the network, but I haven't seen anyone do that with this one.

  • @denijane89
    @denijane89 4 หลายเดือนก่อน +1

    I used yarn in the plane to discuss with phi3 about astrophysics. It was epic. At some point, it becomes slow (llama3 started out slow), so I would read for a minute something else and return to its response and continue the chat. It was great. I so hate the offline time on planes, but having a gpt (or 10) on you computer really helps. Also very funny how my battery would drop on burst every time I press enter, lol. It takes juice generating those responses. But again, it's pretty cool. I think at some point I wanted to install ollama ui but the docker part lost me.

  • @jtmuzix
    @jtmuzix 3 หลายเดือนก่อน

    I've been using text-generation-webui, with extensions, but I might have to switch if I can get this working correctly.

  • @DarinLawsonHosking
    @DarinLawsonHosking 4 หลายเดือนก่อน +1

    This comes so close
    Quick question I am looking for a single installation that allows running LLMs locally and has a front end that includes text-to-speech and speech-to-text, with a backend that allows me to attach from the local network from other machines to provide LLM services the kicker is I would like to to add all interactions into a rag from both front end and back end services to allow "learning"
    The idea is to have 1 pc providing all my AI needs locally for home automation, assistance, and various library/research/content generation services on my local network and continuing to learn about me and my needs as time passes (ultimately to be my interface with the internet in general) and help/lead would be greatly appreciated.

  • @DihelsonMendonca
    @DihelsonMendonca 2 หลายเดือนก่อน

    It´s wonderful, but how can I add Hugging Face models to Open WebUI also ? I have tons of them ! ❤❤❤

  • @matikaevur6299
    @matikaevur6299 4 หลายเดือนก่อน +1

    Thanks! Great video!
    Can you do bit different scenario : LLM engine (Ollama, llama.cpp or something else) running on dedicated hw with GPU acc. and frontend in local machine (vm/container/conda env).
    I'v been using ollama on hw with GPU's and frontends in VM or container since they started supporting OpenAI API .. but very mixed results. Idea is that opensource LLM backends are quite stable now .. few upgrades, new features and bugfixes. Just download new model when something interesting appears.
    But frontends and other implementations are still Wild West and better kept in easily recyclable format :)

  • @aamir122a
    @aamir122a 4 หลายเดือนก่อน +1

    I have tried the blulk load feature in the documents , it does not work , cannot change the embedding models as well , after the restart it keeps defaulting to the original. There is a ticket in the github for exactly this . However I have not heard back from the team.

  • @cyanophage4351
    @cyanophage4351 4 หลายเดือนก่อน +1

    If it's local why does it need a sign in. There should be a way to create a version that doesn't need that

  • @DaveEtchells
    @DaveEtchells 4 หลายเดือนก่อน +1

    Another superb, insanely useful how-to vid!
    A noob question: Setting this up with Docker vs Anaconda, the config with Python version and all the supporting libraries and package becomes the default for that user account on your computer, correct? (I’m on a Mac like you, and don’t think this is an issue at all for me, just wanted to understand the setup better.)

  • @kiranwebros8714
    @kiranwebros8714 4 หลายเดือนก่อน

    Instead of Model centric, I wish an agent centric UI with RAG capability

  • @jawwadhussain8457
    @jawwadhussain8457 หลายเดือนก่อน

    Now what if we made your own model ! and we want to use that model on your website how could we do that , any tools easy for non programers

  • @Phanboy
    @Phanboy 4 หลายเดือนก่อน +1

    Does it have a memory feature? Wonder if it can be used as an ongoing "database" per-se

  • @peterbelanger4094
    @peterbelanger4094 3 หลายเดือนก่อน

    This sucked. appears only for the ultra technical who can use stuff like docker.

  • @WhyAreYouFindingMe
    @WhyAreYouFindingMe 4 หลายเดือนก่อน +2

    Hey! Love your videos! and yes, I'm your first commenter of this video

  • @hicamajig
    @hicamajig 4 หลายเดือนก่อน +3

    Supposedly they are working on implementing a perplexity style search too! Pretty slick

    • @jarad4621
      @jarad4621 4 หลายเดือนก่อน

      Like web?

    • @stickmanland
      @stickmanland 4 หลายเดือนก่อน

      @@jarad4621 Yes web

    • @ihaveacutenose
      @ihaveacutenose 4 หลายเดือนก่อน

      If they get something like that it's game over for the rest of the opensource interfaces!

    • @ihaveacutenose
      @ihaveacutenose 4 หลายเดือนก่อน

      Can you use page-assist with openweb ui?

  • @jasonsalgado4917
    @jasonsalgado4917 4 หลายเดือนก่อน +1

    Why would I use this over autogen? Looks pretty similar

  • @rebelScience
    @rebelScience 2 หลายเดือนก่อน

    Awesome video. Can you, going forward, mention the system resources needed for the local stuff like that, to run it at reasonable speed? Thanks.

  • @aa-xn5hc
    @aa-xn5hc 4 หลายเดือนก่อน +1

    Did not explain how to update and still keep all the settings intact!

  • @brainwithani5693
    @brainwithani5693 4 หลายเดือนก่อน

    I dont think i understand, can this be used as a UI for a particular platform or device, or is it just a standalone thing? I mean, would this work as a UI instead of windows or Android, or is this a toy?

    • @SooDamGood
      @SooDamGood 4 หลายเดือนก่อน

      its local on your pc

  • @MaxKøngerskov
    @MaxKøngerskov หลายเดือนก่อน

    Absolutely amazing, thanks!

  • @orthodox_gentleman
    @orthodox_gentleman 4 หลายเดือนก่อน

    I wonder how it could be best accessed via iPhone GUI?

  • @AnimalCentral-l7p
    @AnimalCentral-l7p 4 หลายเดือนก่อน

    there is no voice calling feature in it like chatgpt has now on the phone right?

  • @mendthedivide
    @mendthedivide 4 หลายเดือนก่อน +2

    its on Pinokio as well

  • @zippytechnologies
    @zippytechnologies 4 หลายเดือนก่อน +1

    Yes I knew talking about anticipation would put you over the top to release😂

  • @torarinvik4920
    @torarinvik4920 4 หลายเดือนก่อน +1

    There is an agentic system that runs on Groq, might be worth checking out, I can't remember the name.

  • @SinOfLustAMV
    @SinOfLustAMV 4 หลายเดือนก่อน

    This is fine. but AnythingLLM FTW!!!! XD

  • @PJ-hi1gz
    @PJ-hi1gz 4 หลายเดือนก่อน +1

    Thank you. Best tutorial on youtube. Very clear.

  • @Bruce-sc1ph
    @Bruce-sc1ph 3 หลายเดือนก่อน

    can I run an remote model with api key on this?

  • @MAFiA303
    @MAFiA303 4 หลายเดือนก่อน

    I wonder if there will be a tool called Barack Ollama

  • @markg5891
    @markg5891 4 หลายเดือนก่อน +1

    Question about these embedding models. I _thought_ that adding your own document makes and "embedding" loaded into the LLM for you to search your document. But what are those embedding models that you can define in the gui settings? I can't quite figure out why i'd want them or what they are used for..

    • @jarad4621
      @jarad4621 4 หลายเดือนก่อน

      I just learn this myself. Apparently embedding is a seperate thing or app or code that does that part so you actually need the database and an embedding model only to be able to save stuff in. vector you don't need the llm yet only once it's saved by the embedder then you can retrieve from the db with your lmm and rag, the main llm doesn't save the data only fetches it later

    • @markg5891
      @markg5891 4 หลายเดือนก่อน

      @@jarad4621 Thank you for your reply! I've read it about 10 times but still don't think i'm following you entirely. I get that you don't train your main LLM for your custom data. That's what the whole RAG thing is about. But with RAG you end up with a vector database that your main LLM can use to provide better results for specific subjects. Thus far i get it.
      But what is that "embeddings llm" like "nomic-embed-text"? I can't figure out how you use it, where you use it or what to do with it.. Just guessing here, or is there some kind of model hierarchy? Does it follow a path like llama -> nomic-embed-text -> own vector db? In which case the "nomic-embed-text" would be the one interacting with my data, not llama like it would be in a classic RAG setup? 99% guesses here, i just don't know.

  • @awesomedata8973
    @awesomedata8973 4 หลายเดือนก่อน +1

    I'm not on a Mac. :(

    • @boomantv8271
      @boomantv8271 4 หลายเดือนก่อน

      *** I posted this for someone else here thought it may help you a little bit too. If anyone wants to correct or add anything please do.***
      I'm not sure if you've gotten any answers to your questions here, but I thought I'd try to help a bit where I can. To your first question, docker is a program that allows you to run software in something called a container. You can think of containers as a space in your RAM that is separate from the rest of the programs that are running on your computer. Docker is nice because it makes it easy for you to run things that are dependent on certain versions of Python , shared libraries, and other stuff without you having to worry about changing versions back and forth to make other things work on your computer. Secondly, from what I could see, buddy is on a Mac computer (the UI he shown at the top where there was a llama kind of was the giveaway) which unless he changed it to bash or something, means he is using a zsh terminal, but that's not important. If you are using a windows computer look up how to get WSL (a.k.a Windows Subsystem for Linux) on your computer and you should be able to follow most of the instructions.
      Helpful tips:
      Linux has different distributions but they fall into 3 main families- Debian(some people say Ubuntu but I'm my own brand of snob here), RedHat, and Arch. This is important to you as far as getting started working with linux because it'll determine your package manager. (Debian uses apt, RedHat uses dnf, and Arch uses pacman) they all have their own quarks as far as what they name things and how they function. But your going to want to ensure you have git installed in you WSL environment as well as having the most up-to-date version of things on there for security reasons to start off.
      Learn about the basics of bash shell.
      Learn the basics of git.
      There will be some differences as far as folder structure once the program is installed, but I think you're smart of enough to figure that out 😉. Hope this helps.

  • @nguyenanhnguyen7658
    @nguyenanhnguyen7658 4 หลายเดือนก่อน

    Ollama UI only accept 1 OpenAI-compatible servẻ end-point (vLLM), while accepting multiple GGUF. Weird !

  • @cyboc
    @cyboc 3 หลายเดือนก่อน

    Great video! I installed Open WebUI on top of Ollama yesterday. It's freaking awesome. Looks and feels just like ChatGPT. Keep up the good work on your videos. I think they are "best of breed" when it comes to AI videos on TH-cam. By the way, the response tokens per second in your video is very high. What is your hardware, especially your GPU(s)?

  • @Electro_69
    @Electro_69 3 หลายเดือนก่อน

    Is there a way to make custom ui?

  • @abdelhakkhalil7684
    @abdelhakkhalil7684 4 หลายเดือนก่อน

    I have already downloaded many models that I use with Oobabooga and LM Studio (Thank you for suggesting LM Studio, it's a good software. Shame it does not support EXL2). But, I don't want to download models specific to Ollama.

  • @mpvincent7
    @mpvincent7 4 หลายเดือนก่อน

    Will try it This coming week! Have you or anyone tried the "Chat with RTX" by NVIDIA? I am trying to go through the setup giving it directories to intake BUT it failed on most expecting zip files... I recall you had covered "Private GPT" before, would it be able to do the same thing intended for Chat with RTX? I bought a new system with a 4060TI 16gig and 32 gig RAM, 1 TB SSD with 2 TB HDD, windows 11... It should run well but currently crawling while Parsing Nodes... any guidence woudl be greatly apreciated!

  • @Yewbzee
    @Yewbzee 4 หลายเดือนก่อน

    Great video thanks for sharing. Sam Altman talks about GPT 4 being an embarrassing model. Hopefully his team see this video and are equally embarrassed about their piss poor UI.

  • @AIVisionaryLab
    @AIVisionaryLab 2 หลายเดือนก่อน

    @Mathew ollama with webui on Linux os

  • @tukanhamen
    @tukanhamen 4 หลายเดือนก่อน

    I really wish people would start being more original with their ui. If I wanted my front end to look just like ChatGPT then I would just use ChatGPT free lol. That’s why I like LMStudio. Only thing holding me back from using it more often is the lack of voice features.

  • @verasalem5071
    @verasalem5071 4 หลายเดือนก่อน

    Hey matt, can you tell us how you setup your terminal? (which terminal software you use, addons, etc). Was also hoping you can create a video testing which LLMs do the best at text summarization for large documents / transcripts

  • @TomM-p3o
    @TomM-p3o 4 หลายเดือนก่อน

    What we need is multi-prompt templates (series of prompts, one at a time), including step repeat. This way we can have the LLMs reflect on its previous answer before executing the next step in the series.

  • @alkeryn1700
    @alkeryn1700 4 หลายเดือนก่อน

    i run it without docker or ollama lol.

  • @therealbobbyb
    @therealbobbyb หลายเดือนก่อน +1

    So helpful! Thank you!!!

  • @cyboc
    @cyboc 3 หลายเดือนก่อน

    I should add that your style is just as engaging as the dude from The Why Files channel. When are you going to get a cool sidekick like HeckleFish?

  • @JoeBrigAI
    @JoeBrigAI 4 หลายเดือนก่อน

    Still not sure about the name, Open WebUI. Very generic and not descriptive of its functionality.

  • @takstravelvlog7357
    @takstravelvlog7357 4 หลายเดือนก่อน

    Thanks for creating this video with clear instructions!!
    After running this UI with llama3, the responses that I get is very slow even to those simple questions, e.g. how are you?. Is having GPU in my machine must to use this with proper speed?

  • @jakovnikolic8622
    @jakovnikolic8622 4 หลายเดือนก่อน

    Your video about this software comes way too late to be honest. The guy behind this project had made this thing awesome like year and a half ago. But it is nice that you did it anyway people need to know about this project and support it since rly rly good. Keep up the good work ;)

  • @laalbujhakkar
    @laalbujhakkar 4 หลายเดือนก่อน

    0:50 "Look how fast that is" well, the response being fast is not a function of the UI, it's a function of the hardware you're running it on. so please.

  • @doubs22
    @doubs22 4 หลายเดือนก่อน +1

    We have been using Chabot-ui and enjoy the built in tools and assistants libraries.
    Keeping an eye on these other repos until a true leader emerges.

    • @jarad4621
      @jarad4621 4 หลายเดือนก่อน

      Does it have web?

    • @doubs22
      @doubs22 4 หลายเดือนก่อน

      @jarad4621 web search? It needs to be added as a tool/skill.

  • @PeterJung-cx1ib
    @PeterJung-cx1ib 2 หลายเดือนก่อน

    What HW are u running the LLM on?

  • @xDARKMATT3Rx
    @xDARKMATT3Rx 4 หลายเดือนก่อน

    Another extremely informative video! I have an idea for a follow-up video: I would love to see a tutorial on how to secure the docker container with https / ssl. I am trying to figure this out through experimentation and have so far been unsuccessful. Thank you again for your great content!

  • @mikelewis1166
    @mikelewis1166 3 หลายเดือนก่อน

    MSTY is way better. I’ve actually switched to MSTY for my work.

  • @hotlineoperator
    @hotlineoperator 4 หลายเดือนก่อน

    Impressive, this is having all - except support for my local language (what ChatGPT do have).

  • @theworddoner
    @theworddoner 4 หลายเดือนก่อน

    I prefer ooba text Gen webui as it’s more customizable. The RAG implementation is not something I’ve tried before. I guess I’ll give it a try.

  • @AdamMoreyPlus
    @AdamMoreyPlus 4 หลายเดือนก่อน

    Great video. Thanks! fyi - In the install section, it's not necessary to clone the repo before running docker.

  • @proflead
    @proflead หลายเดือนก่อน

    Nice solution with the docker!

  • @lwjunior2
    @lwjunior2 4 หลายเดือนก่อน

    Can it handle the processing of documents with tables? Will an OCR like Tesseract be needed? Any suggestions

  • @Gl0we22
    @Gl0we22 4 หลายเดือนก่อน

    have you tried Silly tavern? its more geared towards characters and roleplay but has some nice features and a fully customizable ui

  • @VAS.T
    @VAS.T 4 หลายเดือนก่อน

    what gpu server provider you use for your language model deployment? or do you only need gpu power for training?
    can you create chat agents from those ollama deployed chat models?
    im just getting into stuff and some of your videos are nice to follow

  • @thekingofallblogs
    @thekingofallblogs 4 หลายเดือนก่อน

    in case this helps someone. to run linux images on windows, you need to switch to linux containers, which will use wsl2. you do this by rightclicking on docker running in taskbar, and selecting switch to linux containers.

  • @JohnSmith-vo8is
    @JohnSmith-vo8is 4 หลายเดือนก่อน

    Could this work on Jan?

  • @giox1979
    @giox1979 2 หลายเดือนก่อน

    Thank you for your clear documentation, really helpful to setup a complete system in a few steps. Great job!

  • @johnwalker348
    @johnwalker348 4 หลายเดือนก่อน

    amazing tutorial, thanks!!
    how do you know the chunking methods for the embedding?

  • @OpusG5
    @OpusG5 4 หลายเดือนก่อน

    I was hoping someone else would handle all the heavy lifting with the UI, as it was somewhat of a turn-off for me to read replies in the shell. Thank you for sharing it!

  • @Marques2025
    @Marques2025 2 หลายเดือนก่อน

    Can we generate plots on this

  • @drlordbasil
    @drlordbasil 4 หลายเดือนก่อน

    I clicked hella fast because I thought I missed an update from ollama XD love openui though!

  • @sherpya
    @sherpya 4 หลายเดือนก่อน

    just make sure you don't use update models functionality if you have some models as ollama author pointed, until they implement model hash comparisons

  • @all3n1k
    @all3n1k 4 หลายเดือนก่อน

    we dont need fast inference we need quality inference, im willing to wait minutes for responses for complex questions as long as they are coherent

  • @polishandrew3955
    @polishandrew3955 4 หลายเดือนก่อน

    Anyone know which terminal software Matthew uses for his? That auto complete is a nice feature and would love to add it to mine