Jetson AI Labs - Generative AI on the Edge

แชร์
ฝัง
  • เผยแพร่เมื่อ 30 ต.ค. 2023
  • NVIDIA recently launched the NVIDIA Jetson Generative AI Lab. This Jetson AI Lab is a gathering place for tutorials and resources for running Generative AI applications on the Edge.
    If you have a Jetson Xavier or Jetson Orin, this is a great place to explore the development environments and capabilities of Generative AI on the Jetson.
    In the video, we use a NVIDIA Jetson AGX Orin 64 GB Developer Kit:
    amzn.to/45VEaTp
    The tutorial in the video covers LLaVA, a multimodal model which combines vision transformers and LLMs.
    NVIDIA Jetson Generative AI Lab: www.jetson-ai-lab.com/index.html
    As an Amazon Associate I earn from qualifying purchases.
    Visit the JetsonHacks storefront on Amazon: www.amazon.com/shop/jetsonhacks
    Visit the website at jetsonhacks.com
    Sign up for the newsletter! newsletter.jetsonhacks.com
    Github accounts: github.com/jetsonhacks
    github.com/jetsonhacksnano
    Twitter: / jetsonhacks
    Some of these links here are affiliate links. As an Amazon Associate I earn from qualifying purchases at no extra cost to you.
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 23

  • @bertbrecht7540
    @bertbrecht7540 8 หลายเดือนก่อน +3

    It's the early days of this technology. It will become part of our daily lives (self driving cars, assistant for the blind, weed pulling machine, garbage sorter, etc.... times a million)
    Thanks for presenting this. I will be putting my Jetson to work.

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน +2

      You're welcome. I like your analysis! Even though it's early on, there are glimmers of what could be. Thanks for watching!

  • @suryanshu3724
    @suryanshu3724 8 หลายเดือนก่อน +2

    will this work on the 4gb jetson nano? or only the orin line?

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน +4

      The generative AI is centered on the Jetson Orin and Xavier machines. The memory requirements of these models rule out using smaller machines. Thanks for watching!

  • @allvisualmedia7575
    @allvisualmedia7575 8 หลายเดือนก่อน

    Can you please make a video on LLama 2, memgpt, autogen running on this machine?

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน

      Thank you for the suggestions. Here's a video on running Llama 2: th-cam.com/video/HlH3QkS1F5Y/w-d-xo.htmlsi=149llDVY3LH_6Sl0&t=436
      The other two are still a little bit too young to be able to work with. Is there something in particular you want to know?

  • @robrever
    @robrever 6 หลายเดือนก่อน +1

    This might be a stupid question, but it is performing these analysis with or without internet access?

    • @JetsonHacks
      @JetsonHacks  6 หลายเดือนก่อน +1

      Once you download the docker images and models, everything runs on the Jetson with no further need of an Internet connection. Thanks for watching!

    • @robrever
      @robrever 6 หลายเดือนก่อน +1

      @@JetsonHacks Great thanks a lot! I have some big plans in the next year utilizing the Jetson computer vision/image recognition. I look forward to your videos as this is all very new to me. Thank you for doing what you do.

    • @JetsonHacks
      @JetsonHacks  6 หลายเดือนก่อน

      @@robrever You are welcome. Hopefully you can share some of your work.

  • @leibaleibovich5806
    @leibaleibovich5806 8 หลายเดือนก่อน +1

    Greetings, Jim!
    I would love to hear your opinion on the following question:
    Often, when I read or watch videos about hands-on deep learning (i.e. neural networks, NNs), people discuss the hardware requirements because most of deep learning is done on GPUs. Someone told me that one has to have a graphics card with min. of 16 Gb of memory. Hi-end graphics cards are expensive!
    Recently I was browsing through tech specs of different Jetson Orin products. For example, reComputer J4012 (based on Jetson Orin NX 16GB) is capable of up to 100 TOPS. If memory serves, Jetson Orin Dev Kit is capable of up 40 TOPS.
    My question is: how does one compare graphics cards vs. Jetson Orin in terms of performance for deep learning? In case of Orin, you can get a mini-PC, a complete system. On the other hand, you need a decent rig to fit it with a capable graphics card. What are trade-offs?
    I have not found much info on this subject, so your take will be much appreciated.

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน +2

      Big question! Here's the way I usually look at it. If you compare a NVIDIA RTX 30 series (ie 3050, 3090 or in between) they use the same GPU architecture (Ampere) as the Orins. The RTX 40 (e.g 4090) series are a generation newer. Depending on which graphics card you pick, they use 350 Watts+. The Jetson run on 1/10 of that or so.
      The number of GPU cores on the graphics cards run from ~3000 to 10,000. On the Jetson, the AGX Orin is 2,000. The Orin NX is half that. The memory in the graphics card is GDDR6 which is much faster than the Orin LPDDR4. Remember, the LP stands for Low Power meaning that the memory is tuned for power management rather than performance. The G in the GDDR6 means Graphics, which optimizes for speed.
      The memory bus width is wider on the graphics card. The Orin NX has 16GB of memory, but it is unified meaning the CPU and the GPU share it. Also, the clocks on the graphics card for the GPU are faster.
      As you note, there's the uncomfortable bit about everything fitting into graphics memory, and swapping things in and out. There are various strategies around this, but depends on the machine learning models that you are running.
      What that tells you is that it is an apples vs oranges comparison. One would expect the graphics cards to be *much* faster than the mobile processors. But the tradeoff is that it uses more than 10x the power. If low power consumption isn't your main goal (like in embedded systems like the Jetson), then there's no reason for that constraint. And vice versa, if you're looking at low power it doesn't make sense to run the big iron.
      I'm sure there are people who know way more than me about this that can address it more eloquently. Thanks for watching!

    • @leibaleibovich5806
      @leibaleibovich5806 8 หลายเดือนก่อน +2

      @@JetsonHacksThank you very much for this comprehensive answer! I appreciate time and effort you put into it, Jim!
      That's really valuable information for someone not so tech savvy, like myself! Thank you, Jim!

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน +1

      @@leibaleibovich5806 You are welcome.

    •  8 หลายเดือนก่อน +1

      @@JetsonHacks Apples are for apple pie and oranges are for orange juice. Application specific devices ;-)

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน

      @ This is a much more elegant way of saying the same thing. Thanks for watching!

  • @ArtificialDNA
    @ArtificialDNA 8 หลายเดือนก่อน +1

    Hi, big fan of your i know it is not related to this directly but some what ... because when try to flash AGX to NVME it only use 32GB of the disc and many have this problems that could not find solution I try your hack by flash to EMMC and move to NvME and other methoer all seem same stuck in 32GB .. which is nothing when try to do this generative AI because it keep run out of space ..

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน

      Thank you for the kind words. Unfortunately I don't have a good answer for you on this. There apparently is an issue in the flashing scripts that confuse the partitions on the NVMe drive. People have work arounds, but I can't say they're easy to implement.
      A workable solution might be to flash to eMMC and set everything to store to the SSD. Here's some instructions from the Jetson AI Lab: www.jetson-ai-lab.com/tips_ssd-docker.html
      Thanks for watching!

    • @ArtificialDNA
      @ArtificialDNA 8 หลายเดือนก่อน

      @@JetsonHacksthank you every much i guess I will try this work around. yes before the latest one it work fine .

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน

      @@ArtificialDNA Another workaround which will get you to where you want to be is to downgrade JetPack to 5.1.1 and flash the AGX. Once the Jetson boots and you set it up, install JetPack. Then do sudo apt update && sudo apt upgrade which will update everything on the system. You should be at the same place that a correct 5.1.2 install should have done.

    • @ArtificialDNA
      @ArtificialDNA 8 หลายเดือนก่อน +1

      @@JetsonHacks Jet it is work by install 5.1.1 however to upgrade to 5.1.2 i need to change source list from R35.3 to R35.4 thank a lot for help ..

  • @dggcreations
    @dggcreations 8 หลายเดือนก่อน

    So where is this even remotely useful?

    • @JetsonHacks
      @JetsonHacks  8 หลายเดือนก่อน +6

      Unfortunately you have to use future goggles to predict where and how this technology will be applied. The demo shown here is one of many on the website. That demo is a little esoteric and requires a little imagination on how you might apply it. "Look in the refrigerator and tell me what I could fix for lunch" may not be useful to you. However, people may find it useful to find assistance when analyzing images of natural phenomenom. Or analyzing medical images, or asking for advice in an industrial setting.
      You're probably familiar with a web-ui to LLM chat, such as GPT-4. The Jetson AI website has an open source version (LLaMA) which runs everything locally. There are advantages to being able to run these types of models in a local device.
      There's the low hanging fruit of automatic speech recognition and text to speech (ASR and TTS). Combining those with a GPT provides a conversational AI in a more natural way than a web browser.
      One application which already has a lot of traction is object segmentation which has application in many areas such as robotics and security. There are several tutorials on the Jetson AI website about how to leverage these features.
      It seems more clear that much of computer vision processing will be done with machine learning such as shown on the website.
      Thanks for watching!