Unlimited AI Agents running locally with Ollama & AnythingLLM

แชร์
ฝัง
  • เผยแพร่เมื่อ 6 มิ.ย. 2024
  • Hey everyone,
    Recently in AnythingLLM Desktop, we merged in AI Agents. AI Agents are basically LLMs that do something instead of just replying. We support both tool-call-enabled models like OpenAI but have even now have a no-code way to bring AI agents to every open-source LLMs like with Ollama or LMStudio.
    Now, with no code required, you can take any LLM and get automatic web scraping, web-browsing, chart generation, RAG memory, and summarization all autonomously and running locally.
    If the future of AI is agents, AnythingLLM is where it is going to happen!.
    Download AnythingLLM: useanything.com/download
    Star on Github: github.com/Mintplex-Labs/anyt...
    Chapters:
    0:00 Introduction to adding agents to Ollama
    0:45 What is Ollama?
    1:08 What is LLM Quantization?
    1:28 What is an AI Agent?
    2:54 How to pick the right LLM on Ollama
    5:11 Pulling Ollama models and running the server
    5:45 Downloading AnythingLLM Desktop
    6:17 AnythingLLM - Initial setup
    7:21 Sending our first chat - no RAG
    8:22 Uploading a document privately
    8:43 Sending a chat again but with RAG
    9:10 How to add agent capabilities to Ollama
    10:45 Add live web-searching to Ollama LLMs (Free)
    11:41 Using agents in AnythingLLM demonstration
    13:24 Agent document summarization and long-term memory
    14:35 Why you should use AnythingLLM x Ollama
    15:00 Star on Github, please!
    15:06 Thank you
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 251

  • @sergiofigueiredo1987
    @sergiofigueiredo1987 27 วันที่ผ่านมา +58

    @TimCarambat I had to pause the video just to leave a comment! I'm deeply impressed by the excellence and simplicity of the content presented here. It's truly remarkable to have access to such tools, created by a team that clearly demonstrates passion and a keen ear for what we all think and wish would be great to have, and at every update, distilling all p of these wishes into a few simple clicks within this amazing piece of technology! I'm immensely grateful for the opportunity to experienceh the brilliance of software engineering and development of Anything LLM, especially within the context of open-source communities. Participating in the advancement of genuine and incredible open tools is a privilege. Thank you Tim! I will be promoting this project to the moon and back, because this deserves to be known.

    • @TimCarambat
      @TimCarambat  27 วันที่ผ่านมา +3

      This is so incredibly kind. Sharing with team!

    • @THOOOMEME
      @THOOOMEME 13 วันที่ผ่านมา

      haha I was just about to leave a comment when I read yours. I feel the same. What a champion Tim is. I do not know if I will ever install AnythingLLM but I think I will donate to Tim regardless.

    • @ts757arse
      @ts757arse 10 วันที่ผ่านมา

      Aye, I was interested in anythingLLM a while back but chose another project for my inference server. I've found getting half decent agent capabilities to be a huge time sink for someone with my skill set (I'm a physical security guy, not a programmer) and the results just weren't worth the time invested.
      Even basic agent capabilities with RAG, memory and so on in a package that I can just plug into ollama sounds awesome.
      Prepping the server now. Here's hoping.

  • @surfkid1111
    @surfkid1111 27 วันที่ผ่านมา +24

    You built an amazing piece of software. Thank god that I stumbled across this video.

  • @MartinBlaha
    @MartinBlaha 27 วันที่ผ่านมา

    Thank you! Will test it for sure. I think you guys are on the exact right path 😎👍

  • @jonathan58475
    @jonathan58475 16 วันที่ผ่านมา +2

    Tim, thank you for making the world a better place with this awesome tool! :)

  • @liviuspinu11
    @liviuspinu11 24 วันที่ผ่านมา +5

    Thank you for explaining quantisation in details for niebiews.

  • @michaelklimpel3020
    @michaelklimpel3020 14 วันที่ผ่านมา

    Big thanks man. This video helps alot for me as an beginner to understand how good a local llm is and which Usecases we have. Thumbs up for this great video.

  • @yusufaliyu9759
    @yusufaliyu9759 28 วันที่ผ่านมา +2

    Great this will make LLM more understandable for many ppl.

  • @yasin6904
    @yasin6904 11 วันที่ผ่านมา

    Im a chronic video skipper but watched this back to back. Great explanations and can't wait to try this out! Would love to see more videos, tutorials or even lectures from you. You really have a knack for explaining things!😊

    • @yasin6904
      @yasin6904 10 วันที่ผ่านมา

      PS I've starred on Github!

  • @quinnlintott406
    @quinnlintott406 13 วันที่ผ่านมา

    I had no idea you had a channel talking about your software. Im a big fan of your work!

  • @fxstation1329
    @fxstation1329 18 วันที่ผ่านมา +5

    What I love about your tutorials is that you succinctly explain all the things that come across during the tutorial. Thanks!

  • @tunoajohnson256
    @tunoajohnson256 20 วันที่ผ่านมา

    Awesome vid! Really impressed with how you presented the information. 🙏 thank you

  • @OpenAITutor
    @OpenAITutor 17 วันที่ผ่านมา +1

    Amazing Tim. Keep up the good work.

  • @ilanlee3025
    @ilanlee3025 7 วันที่ผ่านมา

    Good stuff, will try it out. Subscribed. Looking forwards to seeing how this develops.

  • @TheDrMusician
    @TheDrMusician 27 วันที่ผ่านมา +9

    This is by far the easiest and most powerful way to use LLMs locally, full support, like and sub. And many thanks for the amazing work, especially being open source.

  • @jakeparker918
    @jakeparker918 13 วันที่ผ่านมา

    This is so dope. Great no-code solution and it's awesome that it's open source.

  • @SiliconSouthShow
    @SiliconSouthShow 28 วันที่ผ่านมา +7

    Fantastic Tim! Mine doesnt have agent config, guess i need to delete and udate, ill try that, looks great! keep up good work, i love anythingllm i really do!

  • @stanTrX
    @stanTrX 27 วันที่ผ่านมา

    This is the easiest all-in-one platform. Thanks. More videos please ❤

  • @figs3284
    @figs3284 26 วันที่ผ่านมา +1

    Incredible.. gonna make building tools so much easier. Cant wait to see more agent abilities added!

  • @MaliciousCode-gw5tq
    @MaliciousCode-gw5tq 23 วันที่ผ่านมา

    Damm,... finally found the tools that i been looking for..MAN you save my day, i have been crazy stuck finding webui for my ollama remote server..your a gift from heaven keep it up your helping alot of people like us..thank you so much..❤❤❤😂😅😊😊

  • @johnbramich
    @johnbramich 6 วันที่ผ่านมา

    Can't wait to use this. Thank you!

  • @jimg8296
    @jimg8296 26 วันที่ผ่านมา +2

    Anythingllm is awesome. Glad to hear custom agents are on the roadmap. It's the big hole in capability. Also need config to change agent promt. I scan a lot of code and the @ is used often to define decorators.

  • @SiliconSouthShow
    @SiliconSouthShow 28 วันที่ผ่านมา +3

    @TimCarambat
    I'm excited to see the features you talked about work with the ollama like in the video for the agent, as of now, its same as before I updated, but it's exciting to think of the future.

  • @vulcan4d
    @vulcan4d 27 วันที่ผ่านมา +1

    This is awesome work. I looked at the other simple to install Windows front ends and stumbled on this. Pretty cool stuff and I love how you can add documents and external websites to feed it information. An offline LLM is soooooo much more preferred. The only item I don't understand is why you could just ask a regular question once you provided the document, but used @agent when asking to summarize a document.

    • @TimCarambat
      @TimCarambat  27 วันที่ผ่านมา +1

      IMO, i find having a local LLM that even is **only** like 75% as good as on online alternative is just much more rewarding.
      Like i can be on an airplane, open my laptop, and start brainstorming with an AI. Pretty neat.
      Next evolution would be a local AI on your phone but i dont think we have that tech _yet_

  • @d.d.z.
    @d.d.z. 26 วันที่ผ่านมา +1

    You are amazing. Thank you 🎉

  • @mrinalraj4801
    @mrinalraj4801 24 วันที่ผ่านมา

    Great work. Thanks a lot 🙏

  • @akikuro1725
    @akikuro1725 28 วันที่ผ่านมา +3

    Awesome! thank you for this. looking forward to more information/details/examples on using agents w/AnythingLLM!

  • @kangoclap
    @kangoclap 19 วันที่ผ่านมา +1

    looking forward to utilizing AnythingLLM, it looks really awesome! congrats on creating such an impressive application! thank you!

  • @gillopez8660
    @gillopez8660 27 วันที่ผ่านมา

    Wow this is amazing... I'm gonna go star you!

  • @sashkovarha
    @sashkovarha 28 วันที่ผ่านมา +1

    This explained the rag and agents parts I couldn't set up. Great educational content for those who are not programmers. Appreciate your explanations being without that much of "pre-supposed" know-how, that coders have - which is most tutorials on youtube...
    I still didn't get why there's a difference between @agent commands and just regular chat

    • @TimCarambat
      @TimCarambat  28 วันที่ผ่านมา +1

      In a perfect world, they are the same. AnythingLLM originally was only rag. In the near future @agent won't be needed and agent commands will work seamlessly in the chat.
      So @agent is temporary for now so you know for sure you want to possibly use some kind of tool for your prompt. Otherwise, it's just simple rag

  • @spacetimepotato
    @spacetimepotato 5 วันที่ผ่านมา

    There were some concepts I didn't quite understand; for example, tunneling from the Windows PC to the Mac (if it's on your local network, why work with VPN protocols rather than client/server - due to needing a stateful connection vs. 200 response code or something?). But the interface itself is brilliant! And I think that when it becomes agent-swarm-capable it's going to be a much better option for me than Crew AI, as it feels more intuitive, I am just going to need multiple agents working together. I have never installed a local LLM, but you have inspired me to give it a try. Thanks!

  • @flusyrom
    @flusyrom 5 วันที่ผ่านมา

    Funny ! I heard yesterday for the first time about AnythingLLM during an AI-info event.... and discarded the idea of giving it more attention because it was presented as "just another local RAG support". And now I stumble across this video by chance - and the additional agent functionality changes everything ! BTW, very well presented , this feature !
    My immediate idea & feedback: if there was ANY chance to model custom agents in Flowise and re-import the JSON exports of this Flowise flow as input for an AnythingLLM custom agent, you'd save yourself the trouble of designing your own agent editor AND would start with a comparably large installed base. OK, maybe that's just wishful thinking..... but maybe I'm also not the only one with this wish to facilitate local agent building ;-)

  • @SamBeera
    @SamBeera 6 วันที่ผ่านมา

    Hi Tim, thank you very much for the great video showcasing open source llms, and tools like anythingllm to create agents. I followed your video and successfully was able to do everything in your video. Are there other agentic videos for other usecases you made, look forward to see them. Cheers

  • @rockon-wbfqlkjqhsydic72683
    @rockon-wbfqlkjqhsydic72683 8 วันที่ผ่านมา

    Great job! This is wonderful! I will be responding after using to let you know my thoughts if you care to see them :)

  • @GoranMarkovic85
    @GoranMarkovic85 15 วันที่ผ่านมา

    Amazing work 👏

  • @mehmetnaciakkk3983
    @mehmetnaciakkk3983 9 วันที่ผ่านมา

    A fantastic beginning! When do you think we willbe able to create our own agents?

  • @FlynnTheRedhead
    @FlynnTheRedhead 28 วันที่ผ่านมา +4

    So training/finetuning is coming up as well? Loving the progress and process updates, keep up the great work Tim!

    • @TimCarambat
      @TimCarambat  28 วันที่ผ่านมา +8

      how'd you know!?
      We will likely make some kind of external supplemental process for fine-tuning, but at least make the tuning process easy to integrate with AnythingLLM.
      RAG + Fine-tune + agents = very powerful without question

    • @FlynnTheRedhead
      @FlynnTheRedhead 27 วันที่ผ่านมา

      @@TimCarambat That's awesome to hear!! I created an agent to get insider info, that's how I know of course!

    • @TimCarambat
      @TimCarambat  27 วันที่ผ่านมา +1

      @@FlynnTheRedhead !!!!! I thought i was hearing clicks during my phone calls!!!

  • @star95
    @star95 23 วันที่ผ่านมา

    Great video! I also want to know how well the RAG function of AnythingLLM performs. It's important that text, images, and papers are handled properly and meaningful chunking are achieved

  • @themax2go
    @themax2go 21 วันที่ผ่านมา

    very cool!!! subbed!

  • @Augmented_AI
    @Augmented_AI 4 วันที่ผ่านมา

    What agents do you have planned for future?

  • @AGI2030
    @AGI2030 วันที่ผ่านมา

    Great work Tim! If using 'AnythingLLM' in the 'LLM Provider' section, can I load other LLMs that are not listed? Like the '8b-instruct-q8_0' you mention? So I don't have to rum Ollama separately to load a model?

  • @red_onex--x808
    @red_onex--x808 25 วันที่ผ่านมา

    Awesome info……thx

  • @UrbanCha0s
    @UrbanCha0s 27 วันที่ผ่านมา

    Looks really good and simple. I tried PrivateGPT using conda/Poetry and could never get it to work, so jumped into WSL for Windows connecting to Ubuntu running ollama, via WEBUI. Works great, but this just looks so much easier. Will have to give it a try. What I do like with the WEBUI I have is I can select different model, and even use multiple models at the same time.

    • @TimCarambat
      @TimCarambat  27 วันที่ผ่านมา

      Yeah, we didnt want to "rebuild" what is already built and amazing like text-web-gen. No reason why we cant wrap around your existing efforts on those tools and just elevate that experience with additional tools like RAG, agents, etc

  • @sharankumar31
    @sharankumar31 17 วันที่ผ่านมา

    this is seriously very neat tool👏👏👏 Pls add some feature to custom develop agents with function calls. It will be helpful for our local automations.

    • @TimCarambat
      @TimCarambat  17 วันที่ผ่านมา

      This is shown in the UI that we will be supporting custom agents soon!

  • @JacquesvanWyk
    @JacquesvanWyk 16 วันที่ผ่านมา

    Really awesome demonstration. I am excited about agents. Would be nice to be able to build custom tools in python for agents to use.

  • @carloscms23
    @carloscms23 26 วันที่ผ่านมา

    Great Work :)

  • @EddieAdolf
    @EddieAdolf 21 วันที่ผ่านมา

    I've been using it for months. Love it! Will you enable voice to voice soon?

    • @TimCarambat
      @TimCarambat  19 วันที่ผ่านมา

      We just did in our most recent update. TTS is live for all, STT is only live for the docker version. There are some restrictions and limitations we need to work around to get STT to fully function cross-platform. It will be solved soon

  • @aimademerich
    @aimademerich 28 วันที่ผ่านมา +2

    Would love to see this run stable diffusion and comfy ui workflows

  • @mrgyani
    @mrgyani 20 วันที่ผ่านมา

    This is incredible..

  • @madhudson1
    @madhudson1 27 วันที่ผ่านมา +1

    Been struggling to get custom agents to integrate reliably with external tooling, using frameworks like crewui with local LLMs. Would love a video guide explaining best practices for this

  • @elu1
    @elu1 21 วันที่ผ่านมา

    really nice!

  • @DaveEtchells
    @DaveEtchells 19 วันที่ผ่านมา

    Wow, this looks *_amazing!_*
    I’m just starting to experiment with local LLMs and wanting to play with agents; this looks SO easy! I’m going to download and set it up right away.
    I’m also interested in Open Interpreter for having an AI assistant do things on my local machine. Can this interface with that, or is it really meant as a substitute/enhancement to it?
    (Also, how can I support your project? I gather your biz model is selling the cloud service, but my usage will be purely local. Anywhere I could send a token few bucks?)

  • @aimademerich
    @aimademerich 28 วันที่ผ่านมา

    Phenomenal

  • @zirize
    @zirize 27 วันที่ผ่านมา +3

    I think it's a very good application, easy to use, and after testing it for a day or so, I have some wishes.
    1. direct commands Bypass Agent LLM in Agent mode. It takes time for the agent to understand the sentence and convert it into internal command, and url parsing sometimes fails depending on the agent. For example, a command that scrapes the specified URL and shows the result, or a command that lists the currently registered documents with numbering. And a command that summarizes the document by this number instead of its full name.
    2. I wish there was a way to pre-test the settings in the options window to make sure they are correct, such as specifying LLM or search engine.
    I hope this application is widely known and loved by many people.

  • @ImSlo7yHD
    @ImSlo7yHD 16 วันที่ผ่านมา

    This is perfect it just needs more tools and agent customization like crew ai and it is going to be an absolute killer for the ai industry.

    • @TimCarambat
      @TimCarambat  16 วันที่ผ่านมา

      Will be coming soon! Just carving out how agents should work within the context of AnythingLLM and should be good.
      Also, it would be nice to be able to just import your current CrewAI and use it in AnythingLLM - save you the work you have done so far

  • @HarpaAI
    @HarpaAI 15 วันที่ผ่านมา

    🎯 Key Takeaways for quick navigation:
    00:00 *🤖 Introduction to Ollama & AnythingLLM and AMA*
    - Introduction to Ollama and AnythingLLM
    - Explanation of AMA application for running LLMs on local devices
    - Overview of quantization process and agent capabilities in LLMs
    02:30 *🧠 Understanding Model Quantization and Selection*
    - Importance of selecting the right quantization level for LLMs
    - Differences between various quantization levels like Q1 and Q8
    - How quantization impacts model performance and reliability
    06:07 *🛠 Setting up AnythingLM with Q8 Model and AMA*
    - Instructions for setting up AMA with Q8 LLW model
    - Steps to download and run AnythingLM on local devices
    - Connecting to AMA server and configuring privacy settings
    08:27 *💬 Enhancing Model Knowledge Using RAG and Workspace*
    - Uploading documents for model referencing in workspace
    - Improving model responses by utilizing documents in the workspace
    - Configuring workspace settings for better model performance
    11:41 *🌐 Using Agents for Advanced Functionality in AnythingLLM*
    - Utilizing agents to enhance LLMs capabilities beyond basic text responses
    - Enabling web scraping, file generation, summarization, and memory functions
    - Integrating external services like Google for web browsing functionalities
    Made with HARPA AI

  • @SebastianMuller-pz9xl
    @SebastianMuller-pz9xl 27 วันที่ผ่านมา

    Amazing ⭐⭐⭐⭐⭐

  • @jimmysrandomness
    @jimmysrandomness 6 วันที่ผ่านมา

    Can it also use dalle but unrestricted?

  • @Great_Muzik
    @Great_Muzik 19 วันที่ผ่านมา

    Awesome tutorial Tim! Can this extract specific data from PDF files and save it to an Excel file?

  • @TokyoNeko8
    @TokyoNeko8 25 วันที่ผ่านมา +4

    Debug mode would be ideal. Agent to scrape the web just exits without any error even though I do have search engine api defined

  • @marinetradeapp
    @marinetradeapp 16 วันที่ผ่านมา

    Great work - thanks for sharing - Question - how can we send data to the agent via webhooks - is this a possibility?

  • @johnbrewer1430
    @johnbrewer1430 17 วันที่ผ่านมา

    @sergiofigueiredo1987, @TimCarambat, I agree with Sergio. Wow! I have Ollama installed locally on a Windows machine in WSL. (I was leery of the Windows preview, but I may switch because NATing the Docker container is a pain.) I also pondered how to build a vector DB on my machine and integrate agents. You guys have already done it!

  • @finessejones3109
    @finessejones3109 16 วันที่ผ่านมา

    I'm so happy I came across your video. Thank you. I am having trouble on where you to get the base link that you pasted in @6:36 mark to install the ollama3

    • @finessejones3109
      @finessejones3109 16 วันที่ผ่านมา +1

      I was able to follow along from your other video to install it. Thank you I'm now a new sub.

  • @mouradlaraba
    @mouradlaraba 13 วันที่ผ่านมา

    thanks a lot for your video, this the first video that i see and it's really simple to understand, i have a question, if for example the model that i use know that the capital of france is paris, how can i change that information and make the answer different from paris? best regards

  • @Alex29196
    @Alex29196 24 วันที่ผ่านมา

    Hi Tim, thank you for your dedication and effort in teaching us about local LLMs. I have a medium-spec computer with 4GB VRAM and 16GB RAM. The last time I installed ALLM, the inference speed was a bit slower compared to other alternatives. How does it perform with the new version? Thanks again.

    • @TimCarambat
      @TimCarambat  19 วันที่ผ่านมา

      Unfortunately, i doubt much would change on the inference side. When you say alternatives, what were you using? You might get slower responses in AnythingLLM vs just chatting via CLI in ollama, but that is because we are adding that valuable context to the prompt. More tokens = more work on the LLM to respond!

  • @TheShawn2880
    @TheShawn2880 25 วันที่ผ่านมา

    Your the best

  • @marius2591
    @marius2591 14 วันที่ผ่านมา

    Hi,
    How does quantization type affects the system resources needed to properly run that model?
    Great video by the way!

    • @TimCarambat
      @TimCarambat  14 วันที่ผ่านมา

      It mostly impacts the RAM and overall storage side of the GGUF modelfile. It's tricky to determine the exact requirement decrease because it has to do with the specific model parameters and other factors. Im not aware of a simple equation or expression that is a direct calculation for all models.
      In general, lower quant -> Smaller file size and memory footprint when loaded, but much worse output performance

  • @Oliver-zy8sq
    @Oliver-zy8sq 10 วันที่ผ่านมา

    Hey, thank you for putting out anytingllm. I have two questions: 1. When I ask the llm to remember something, is that long term memory stored on my pc on a server? 2. is the summary part of the long term memory necessary? And I have a feature request for an automatic long term memory. Meaning that I don't have to say specifically what to remember but that the llm will be able to recall the entire chat history - eveything i have ever said in that thread. Is that in the picture?

  • @redbaron3555
    @redbaron3555 25 วันที่ผ่านมา

    Amazing software!! Congratulations and thank you! Very similar to MemGPT server but seems easier to set up and use. I wonder whether you can save a whole company database (i.e. ERP data: products, materials etc.) in it and being g able to ask questions about it? Also can you instigate more than one agent simultaneously?

    • @TimCarambat
      @TimCarambat  25 วันที่ผ่านมา +1

      In theory, this would be better delegated by some purpose-built agent that can traverse the data. Currently, we only have one-agent conversations but the code _does_ support multi-agent. We just find it to be really messy and cumbersome when many agents are once are trying to do something and your Ollama instance is already at max use generating tokens!

  • @rogerunderhill4267
    @rogerunderhill4267 12 วันที่ผ่านมา

    Brilliant! Could it use my own computer as a data source for the agents? Can I scrape my mac?

  • @SagarRana
    @SagarRana 18 ชั่วโมงที่ผ่านมา

    Thank you so much the only problem i have is i cant seem to find anything llm github pdf file. Where do i download it from?

  • @leninmariyajoseph352
    @leninmariyajoseph352 23 วันที่ผ่านมา

    Great!!!...

  • @CotisoHanganu
    @CotisoHanganu 19 วันที่ผ่านมา +1

    Great things shown.
    Tx for all the work and commitment.
    🎉 Here is a kind of dedicated use case I am interested to get acces:
    I am a mind mapping addict. I use Mind Manager, that stores the mm in .mmap format.
    I would like to ask ANYTHINGLLM to help me scan all folders for mind maps on different subjects and Rag & summarize on them, without having to export all mmap files in another format. Is this doable at this stage? What else should have or have created?

  • @emil8367
    @emil8367 18 วันที่ผ่านมา

    Many thanks for nice introduction !
    Is there a way to configure this LanceDB ? Is there a doc how it's integrated with the AnythingLLM ?

    • @TimCarambat
      @TimCarambat  18 วันที่ผ่านมา

      There is nothing to configure, it is preinstalled and saves to the same location as the application's main storage folder!

  • @LakerTriangle
    @LakerTriangle 28 วันที่ผ่านมา

    Literally sitting here wondering this when you dropped the video

  • @shannonbreaux8442
    @shannonbreaux8442 3 วันที่ผ่านมา

    @Tim do you know anything about home assistant, home automation application. Reason i ask is they already have some intergration with LLM but not with agents and not specialized for home assistant auto automations. When you have time check it out and see if its possible to integrate this with home assistant that would be great. Great job with the video!

  • @jackiekerouac2090
    @jackiekerouac2090 7 วันที่ผ่านมา

    @Tim: I am a professional translator (English to French), and I've just discovered AnythingLLM. Sometimes I have to translate confidential documents that cannot be shared on the cloud. They need to remain locally on my own computer. Once the translation is done, they have to be encrypted to be sent to clients.
    Could I use AnythingLLM to help me with the translation process?
    Could I use it with my actual Lexicum, glossaries and personal dictionaries? Most are PDF or DOCX files.
    How would I do that? What are the first steps?
    Many thanks if you can give me some hints on how to proceed.
    I'm now a new subscriber! 😊

  • @amulbhatia-te9jl
    @amulbhatia-te9jl 15 วันที่ผ่านมา

    Would it be possible to see a vide of setting up your Ollama models on Anything LLM, I followed these instructions but my ollama models never load.

  • @SiliconSouthShow
    @SiliconSouthShow 28 วันที่ผ่านมา +1

    @TimCarambat
    Hey Tim it wont let me select anything under Workspace Agent LLM Provider even though everything is setup and working, obviously ollama is running and everything else in anything is using ollama fine in the app, but this selection option doesn't show like yours does.

  • @nagisupercell
    @nagisupercell 18 วันที่ผ่านมา

    Can I edit my question and regenerate the result in AnythingLLM? I use OpenAI GPT-4o api, but I don't find the edit button in AnythingLLM UI.

  • @DanRegalia
    @DanRegalia 17 วันที่ผ่านมา

    Hey, just found you on a random youtube video suggestion. Love this concept.. A few questions, how deep into a website can this scrape? Can it read a sitemap or robots.txt and download all the data, summarize, etc? Can I hook it into different LLMs? For instance, assign agents to different LLMs? Most importantly, if we're using a vector database, can I feed it rows and rows of data to remember forever?

    • @TimCarambat
      @TimCarambat  16 วันที่ผ่านมา +1

      The one in the document uploader is a single site, but we have a deep website scraper as you mentioned.
      You can use a different LLM per workspace and also per workspace-agent. So yes.
      The vector database we use runs locally and is built in. It works like any other and yes does persist information - so yes to the last point as well

  • @user-mz2ei2nx2p
    @user-mz2ei2nx2p 10 วันที่ผ่านมา +1

    Great video! however, i followed every step you described in every detail, but i could not make the agents communicate with outside world. in any ''search'' or 'webscrape'' request, the model is hallusinating, and presents data that are already to its knowledge insted of real time data (i.e. current gold price ). i used llama3 Q8, i inserted google api and id code, i also tried the other search engine.. nothing. the logs show that it really creates json commands, but nothing comes in from the internet.... any help ?

  • @tonyppe
    @tonyppe 7 วันที่ผ่านมา

    i tried anything LLM and RAG sort of works but I can never pull anything factually from my uploaded text files which are configuration files.
    Is this a model issue? I was using Llama3 Q8 via ollama and llm studio.

  • @4AlexeyR
    @4AlexeyR 14 วันที่ผ่านมา

    Hi, Tim. Great work. I'm trying to use Google. But... it is free for 100 queries per day. How I can control it or limit it? Other options are payable :)

  • @SiliconSouthShow
    @SiliconSouthShow 28 วันที่ผ่านมา +1

    wOOHOO I GOT IT NOW! ID LOVE A UPDATE BUTTON LOL!

    • @TimCarambat
      @TimCarambat  28 วันที่ผ่านมา +1

      It probably just was not refreshed yet. I think we have it on a 1 hour expiration to check so it may have been in between checks

  • @vishalchouhan07
    @vishalchouhan07 20 วันที่ผ่านมา +2

    Hi Tim.. I am absolutely impressed with the capabilities of AnythingLLM. Just a small query..how can I deploy it on a cloud machine and serve it as a chat agent on my website?
    I actually want to add few learning resources as pdf for the rag document of this llm so that my users can chat with the content of those pdfs on my website.
    I also want to understand how many such parallel instances of similar scenario but with different set of pdf is possible? For instance, if I am selling ebooks as digital product to my users, can I have unique instances autogenerated for each user based on their purchase?

    • @TimCarambat
      @TimCarambat  19 วันที่ผ่านมา +1

      We offer a standalone docker image that is a multi-user version of the desktop app. It has a public chat embed that is basically a publicly accessible workspace chat window. You can deploy a lot of places depending on what you want to accomplish: github.com/Mintplex-Labs/anything-llm?tab=readme-ov-file#-self-hosting
      For this, you could do one AnythingLLM instance, multiple workspace where each has its own set of documents, and then a chat widget for each. This would give you the end result you are looking for

  • @biorig
    @biorig 12 วันที่ผ่านมา

    WOW! Mindblown!
    The RAG is fantastic! I uploaded a 'Davidson's textbook of medicine' and was able to ask questions and what not out of it! Thank you for the AnythingLLM Desktop. Thank you! Thank you! I have no more words!

    • @agentred8732
      @agentred8732 11 วันที่ผ่านมา

      Did you ask it questions that validated that the agent/bot was not straying from its training data, and into realms of general knowledge - or hallucination? I have a gigabytes-large proprietary data set that I need to train on, without straying. Open to ideas from anyone reading this comment. Thanks!

    • @biorig
      @biorig 11 วันที่ผ่านมา

      @@agentred8732 The answers were pretty much on the point. I am trying to upload a much larger 'Harrison's Textbook of Medicine', but seems there are limits to the size of the book that can be uploaded.
      So far, no hallucination, but I may not be pushing it to the edge.
      I was told that if we train it on too much material, the output gets generalised and loses depth - I am not yet able to test it with more material.

    • @alpha007org
      @alpha007org 10 วันที่ผ่านมา

      Which model are you running? I tried llama 8B q8, and when I asked question about "Release Notes History.pdf", the results were ... bad.

    • @biorig
      @biorig 10 วันที่ผ่านมา

      @@alpha007org OpenAI API.

  • @betterlifeexe4378
    @betterlifeexe4378 14 วันที่ผ่านมา +1

    I know it's a huge ask, but it would be great if it could listen to a inputs and active windows. it could be really cool if it could capture and describe my workflow, i could analyze what i am doing, and than generate macros for me.

  • @foxnyoki5727
    @foxnyoki5727 25 วันที่ผ่านมา +2

    Does Internet Search Work for You ?
    I configured the agent to use Google Custom Search Engine but search does not return any results.

    • @TimCarambat
      @TimCarambat  25 วันที่ผ่านมา +1

      With some models you _might_ have to word a prompt more directly. Like even explicitly asking it to call `web-browsing` and run this search. Which i know breaks the "fluidity" of conversation, but this is just a facet of the non-determinisic non-steerable nature of LLMs and trying to get them to listen.
      Mostly, its the model that needs to be better so it can follow prompts more closely, but its also not always that simple!

  • @pradeepjain2872
    @pradeepjain2872 24 วันที่ผ่านมา

    Hello. I was just playing with RAG. It seams that the acuracy and results are very poor. I tried with laama 3, wizardlm etc. LLM is unclear of my questions. Is the context windows too short? LLM is giving answeres in a hindsight

  • @Nicola-cc2di
    @Nicola-cc2di 20 วันที่ผ่านมา

    @TimCarambat can you please let me know wich model is anythingLLM using to generate embedding and if it is possible to choose another one? thanks

    • @TimCarambat
      @TimCarambat  19 วันที่ผ่านมา

      We use the huggingface.co/sentence-transformers/all-MiniLM-L6-v2 by default, 384 dimension

  • @morganblais5046
    @morganblais5046 วันที่ผ่านมา

    guessing things have changed but I cannot seem to find where my programmatic access api key would be

  • @deylightmedia3266
    @deylightmedia3266 19 วันที่ผ่านมา +1

    @TimCarambat sir kindly have a for loop so that multiple agents can talk to each other in a chatroom style conversation

  • @caleb.miller
    @caleb.miller 22 วันที่ผ่านมา

    Thanks for the tutorial Tim. For some reason I am not able to get web search working. I am using the same setting you showed in the video. Can you do another video with more detail on setting up the google search engine for this purpose?

    • @ChristianIsai
      @ChristianIsai 20 วันที่ผ่านมา

      I have the same issues, the agent will answer that it doesn't need to use any function and will answer its alucinarions, if I give it the direct order to scrape it will trow a lack of openid key, I think is a work in progress still

    • @TimCarambat
      @TimCarambat  19 วันที่ผ่านมา +1

      Is the model just refusing to call the tool at all or when it does call the tool it says it failed?

    • @ChristianIsai
      @ChristianIsai 19 วันที่ผ่านมา

      @@TimCarambat the model will tell me no need for using any tool I got this and then hallucinate

  • @user-ld8sy9xu2v
    @user-ld8sy9xu2v 23 วันที่ผ่านมา +1

    Hey Tim,what is actual folder that Anything LLM use to store models?
    I have all models downloaded using it on other apps so i would rather just put the model in the right folder then download it again.
    Thanks in Advance!

    • @TimCarambat
      @TimCarambat  19 วันที่ผ่านมา +1

      on Mac: /Library/Application Support/anythingllm-desktop/storage/models
      On window: /Users/user/AppData/Roaming/anythingllm-desktop/storage/models

    • @user-ld8sy9xu2v
      @user-ld8sy9xu2v 19 วันที่ผ่านมา

      @@TimCarambat thanks!

  • @AndyBerman
    @AndyBerman 28 วันที่ผ่านมา

    @TimCarambat Can this run on an old slow server and connect to ollama on a fast server, or does AnythingLLM use a lot of local CPU when invoked?

    • @TimCarambat
      @TimCarambat  28 วันที่ผ่านมา +3

      Actually, this is a perfect combination. AnythingLLM using an external LLM and embedder is no more overhead than just running an HTML page - seriously.
      The only demanding process is if you use the built-in embedder, and that is really only when you are embedding documents. Depending on the size of your documents you could crash the server with the built-in embedder. For reference, our hosted starter tier is 2vCPU and 2GB RAM and we squeak by.
      If it's more than that, you are golden.
      The vector database is so lightweight and fast it is legitimately a non-issue.

  • @user-tz1hj8em7e
    @user-tz1hj8em7e 18 วันที่ผ่านมา

    can you upload a video showing how to embed a chat widget onto a website using the llm ran locally on ollama?

  • @UK-Expat-in-USA
    @UK-Expat-in-USA 11 วันที่ผ่านมา

    it would be nice if the desktop version could have a web server integrated into it so you do not have to mess around with Docker which I found to be slower

  • @davidgalea430
    @davidgalea430 27 วันที่ผ่านมา

    Will not load models in the linux version when I select local Ollama

  • @RhythmRiftsDataDreams
    @RhythmRiftsDataDreams 26 วันที่ผ่านมา +1

    What is the chunking method you use to create the vectors?
    Is there a way that the user can control the method of chunking?
    Say : Short, Token Size, Semantic, Long etc...

    • @TimCarambat
      @TimCarambat  25 วันที่ผ่านมา +2

      We currently use a static recursive chunk splitter. So basically just character counts. You can modify those chunking settings in the settings when you go to "embedder preference". So you can define max length and overlap

  • @SonGoku-pc7jl
    @SonGoku-pc7jl 28 วันที่ผ่านมา

    more more! ;)