OLLAMA | Want To Run UNCENSORED AI Models on Mac (M1/M2/M3)

แชร์
ฝัง
  • เผยแพร่เมื่อ 5 มิ.ย. 2024
  • OLLAMA | How To Run UNCENSORED AI Models on Mac (M1/M2/M3)
    One sentence video overview: How to use ollama on a Mac running Apple Silicon.
    🚀 What You'll Learn:
    * Installing Ollama on your Mac M1, M2, or M3 (Apple Silicon) - ollama.com
    * Downloading Ollama models directly to your computer for offline access
    * How to use ollama
    * How to harness the power of open-source models like llama2, llama2-uncensored, and codellama locally with Ollama.
    Chapters
    00:00:00 - Intro
    00:00:15 - Downloading Ollama
    00:01:43 - Reviewing Ollama Commands
    00:02:29 - Finding Open-Source Uncensored Models
    00:05:39 - Running the llama2-uncensored model
    00:07:25 - Listing installed ollama models
    00:09:18 - Removing installed ollama models
    🦙 Ollama Commands:
    View Ollama Commands: ollama help
    List Ollama Models: ollama list
    Pull Ollama Models: ollama pull model_name
    Run Ollama Models: ollama run model_name
    Delete Ollama Models: ollama rm model_name
    📺 Other Videos you might like:
    🖼️ Ollama & LLava | Build a FREE Image Analyzer Chatbot Using Ollama, LLava & Streamlit! • Mastering AI Vision Ch...
    🤖 Streamlit & OLLAMA - I Build an UNCENSORED AI Chatbot in 1 Hour!: • Build an UNCENSORED AI...
    🚀 Build Your Own AI 🤖 Chatbot with Streamlit and OpenAI: A Step-by-Step Tutorial: • Build AI Chatbot with ...
    🔗 Links
    Ollama - ollama.com
    Ollama Models - ollama.com/models
    🧑‍💻 My MacBook Pro Specs:
    Apple MacBook Pro M3 Max
    14-Core CPU
    30-Core GPU
    36GB Unified Memory
    1TB SSD Storage
    ℹ️ Other info you may find helpful👇
    Can you run LLM tool on your computer: huggingface.co/spaces/Vokturz...
    Remember that you will need a GPU with sufficient memory (VRAM) to run models with Ollama. If you are unsure how much GPU memory you need you can check out a calculator HuggingFace created called "Model Memory Calculator" here huggingface.co/docs/accelerat...
    Also, here is an article that runs you through the exact mathematical calculation for "Calculating GPU memory for serving LLMs" - www.substratus.ai/blog/calcul....
    _____________________________________
    🔔 / @aidevbytes Subscribe to our channel for more tutorials and coding tips
    👍 Like this video if you found it helpful!
    💬 Share your thoughts and questions in the comments section below!
    GitHub: github.com/AIDevBytes
    🏆 My Goals for the Channel 🏆
    _____________________________________
    My goal for this channel is to share the knowledge I have gained over 20+ years in the field of technology in an easy-to-consume way. My focus will be on offering tutorials related to cloud technology, development, generative AI, and security-related topics.
    I'm also considering expanding my content to include short videos focused on tech career advice, particularly aimed at individuals aspiring to enter "Big Tech." Drawing from my experiences as both an individual contributor and a manager at Amazon Web Services, where I currently work, I aim to share insights and guidance to help others navigate their career paths in the tech industry.
    _____________________________________
    #ollama #mac #apple #llama2 #aichatbot #ai
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 5

  • @AIDevBytes
    @AIDevBytes  หลายเดือนก่อน +1

    🧑‍💻 My MacBook Pro Specs:
    Apple MacBook Pro M3 Max
    14-Core CPU
    30-Core GPU
    36GB Unified Memory
    1TB SSD Storage
    ℹ Other info you may find helpful👇
    Can you run LLM tool on your computer: huggingface.co/spaces/Vokturz/can-it-run-llm
    Remember that you will need a GPU with sufficient memory (VRAM) to run models with Ollama. If you are unsure how much GPU memory you need you can check out a calculator HuggingFace created called "Model Memory Calculator" here huggingface.co/docs/accelerate/main/en/usage_guides/model_size_estimator
    Also, here is an article that runs you through the exact mathematical calculation for "Calculating GPU memory for serving LLMs" - www.substratus.ai/blog/calculating-gpu-memory-for-llm

  • @JoshFKDigital
    @JoshFKDigital หลายเดือนก่อน +1

    Should post the commands in the description 😁

    • @AIDevBytes
      @AIDevBytes  หลายเดือนก่อน +1

      👍 Thanks for the feedback! Commands now the description.

  • @everry3357
    @everry3357 หลายเดือนก่อน

    How's the response time with your macbook pro specs does it go anywhere near chatgpt 4?

    • @AIDevBytes
      @AIDevBytes  หลายเดือนก่อน +1

      Once the models loads into the GPU memory for the first time the follow up responses seem to be slightly slower than GPT4. It's honestly not too noticeable if you are running on similar or better hardware specs that I listed in the description.