Depth Estimation on Single Camera with new Depth Anything State-of-the-art Model

แชร์
ฝัง
  • เผยแพร่เมื่อ 2 ก.ค. 2024
  • Inside my school and program, I teach you my system to become an AI engineer or freelancer. Life-time access, personal help by me and I will show you exactly how I went from below average student to making $250/hr. Join the High Earner AI Career Program here 👉 www.nicolai-nielsen.com/aicareer (PRICES WILL INCREASE SOON)
    You will also get access to all the technical courses inside the program, also the ones I plan to make in the future! Check out the technical courses below 👇
    In this video 📝 we are going to take a look at the new state-of-the-art depth anything model on monocular images. We are going to go over the benchmarks, GitHub repository and I'll also show you how to set it up and run it in your own custom python scripts and environments.
    Code: github.com/niconielsen32/dept...
    If you enjoyed this video, be sure to press the 👍 button so that I know what content you guys like to see.
    _____________________________________________________________
    🛠️ Freelance Work: www.nicolai-nielsen.com/nncode
    _____________________________________________________________
    💻💰🛠️ High Earner AI Career Program: www.nicolai-nielsen.com/aicareer
    ⚙️ Real-world AI Technical Courses: (www.nicos-school.com)
    📗 OpenCV GPU in Python: www.nicos-school.com/p/opencv...
    📕 YOLOv7 Object Detection: www.nicos-school.com/p/yolov7...
    📒 Transformer & Segmentation: www.nicos-school.com/p/transf...
    📙 YOLOv8 Object Tracking: www.nicos-school.com/p/yolov8...
    📘 Research Paper Implementation: www.nicos-school.com/p/resear...
    📔 CustomGPT: www.nicos-school.com/p/custom...
    _____________________________________________________________
    📞 Connect with Me:
    🌳 linktr.ee/nicolainielsen
    🌍 My Website: www.nicolai-nielsen.com/
    🤖 GitHub: github.com/niconielsen32
    👉 LinkedIn: / nicolaiai
    🐦 X/Twitter: / nielsencv_ai
    🌆 Instagram: / nicolaihoeirup
    _____________________________________________________________
    🎮 My Gear (Affiliate links):
    💻 Laptop: amzn.to/49LJkTW
    🖥️ Desktop PC:
    NVIDIA RTX 4090 24GB: amzn.to/3Uc7yAM
    Intel I9-14900K: amzn.to/3W4Z5Cb
    Motherboard: amzn.to/4aR6wBC
    32GB RAM: amzn.to/3Jt2XVR
    🖥️ Monitor: amzn.to/4aLP8hh
    🖱️ Mouse: amzn.to/3W501GH
    ⌨️ Keyboard: amzn.to/3xUGz5b
    🎙️ Microphone: amzn.to/3w1F1WK
    📷 Camera: amzn.to/4b4Ryr9
    _______________________________________________________________
    Timestamps:
    0:00 Intro
    0:36 Project Page
    3:01 Github
    4:32 Code Walkthrough
    10:24 Results on video
    10:48 Results on webcam
    13:45 Outro
    tags:
    #Depthmap #DepthEstimation #DepthAnything #Zero-Shot
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 57

  • @NicolaiAI
    @NicolaiAI  4 หลายเดือนก่อน

    Join My AI Career Program
    👉 www.nicolai-nielsen.com/aicareer
    Enroll in My School and Technical Courses
    👉 www.nicos-school.com

  • @a.........1._.2..__..._.....__
    @a.........1._.2..__..._.....__ 23 วันที่ผ่านมา +1

    Everything im learning about my current project keeps bringing me to your videos. Thanks for posting these indepth videos. Has helped a lot during my learning process.

    • @NicolaiAI
      @NicolaiAI  23 วันที่ผ่านมา +1

      Thanks a ton man! Happy u find them helpful

    • @a.........1._.2..__..._.....__
      @a.........1._.2..__..._.....__ 23 วันที่ผ่านมา

      For those who might be curious. I'm attempting to combine yolo and depth estimation to identify products on a shelf and how many are required to fill vacant locations. Got annoyed having to manually recall, so the engineering brain kicked in :)

  • @Rahi404
    @Rahi404 4 หลายเดือนก่อน +3

    Hey there! Thanks for introducing me to the Marigold model. After testing it, I found it wasn't ideal for real-time applications, and even Midas fell short for my thesis project. Luckily, your videos led me to the perfect solution - "Depth Anything". I'm excited to fine-tune it for my project. Your content has been a huge inspiration for my work in Computer Vision. Keep up the great work! 👏👁‍🗨

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +1

      Thanks a lot for the Awesome words mate! Yeah this depth anything can now run in real time and with great performance compared to Midas. Huge leap and hope you can use it for some cool stuff!

  • @Chalce-dony
    @Chalce-dony 4 หลายเดือนก่อน

    I have setup both midas and Depth-Anything with your instructions and i am using it to do inference on rtsp stream. But the Depth-Anything model is being much slower than Midas model in my setup. What could be the reason of this?

  • @abdullahal7512
    @abdullahal7512 4 หลายเดือนก่อน

    any idea how i can use/implement this algorithm for certain case for my banchelor thesis ?

  • @entrepreneerit4490
    @entrepreneerit4490 4 หลายเดือนก่อน

    Can these be converted to actual distances or do we need to use the metric depth model for that?

  • @aleven27
    @aleven27 หลายเดือนก่อน

    is there anyway you can create a code where you can import a video and export it out as a depth map, please let me know 🙏

  • @eugenegamez3112
    @eugenegamez3112 21 วันที่ผ่านมา

    So, let's say I have a single normal camera and I take a photo of a certain object. Is it possible to get real-world coordinates of that object from the photo using this model?

  • @rajmehta8170
    @rajmehta8170 2 หลายเดือนก่อน

    How can I get depth information from heat map created so that I know if particular object is how much far or near relative to camera ?

  • @unveil7762
    @unveil7762 3 หลายเดือนก่อน +1

    Awesome! How can i get the cuda memory address before to go the cpu? Is a waste of process to the image processing in cpu. Anyway awesome video!!

  • @teetanrobotics5363
    @teetanrobotics5363 4 หลายเดือนก่อน +1

    One of the best channels for ai, computer visiion and deep learning

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน

      Wow, thanks a lot for the nice words! Appreciate all of u

  • @nourhanahmed9757
    @nourhanahmed9757 4 หลายเดือนก่อน

    can we print out the depth map as we did in the MiDaS model ?

  • @lyafpv9056
    @lyafpv9056 4 หลายเดือนก่อน +1

    I wanted to ask, is there already a version for stereo vision? And how can one calculate distances with an AI detection overlay to obtain object distances?
    thanks for this super video

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +2

      Thanks a lot, then you will have to use metric depth from the model, check out their GitHub repo! Might do more videos about that as well

    • @entrepreneerit4490
      @entrepreneerit4490 4 หลายเดือนก่อน

      @@NicolaiAI please do! I'd love to understand how metric depth works.

  • @steffen3536
    @steffen3536 4 หลายเดือนก่อน

    Thank you! Are these depth estimators fast enough to run on edge devices like iphones?

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน

      They are the fastest by far but will need to be optimized and exported to a format that’s supported by apps. But in that case it can run real time but not easy to do

    • @steffen3536
      @steffen3536 4 หลายเดือนก่อน +1

      Ah I see thanks ! :)

  • @AAA-xc4zm
    @AAA-xc4zm 4 หลายเดือนก่อน

    Hi Nicolai. Is there a model that can output the depth information of a point if we provide the pixel coordinates of that point?

    • @vipulpardeshi2868
      @vipulpardeshi2868 4 หลายเดือนก่อน

      I also want that using Single camera

  • @RasitEvduzen
    @RasitEvduzen 4 หลายเดือนก่อน +2

    How can we get scaler distance value for specific object from depth video if it's possible?

    • @ed6280
      @ed6280 4 หลายเดือนก่อน +1

      same question

    • @ajarivas72
      @ajarivas72 4 หลายเดือนก่อน

      ​@@ed6280
      AI will get it with no problem.

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +1

      You will need to extract the positions. Normally you would come it with a segmentation model to do that

    • @ashutoshshukla4680
      @ashutoshshukla4680 4 หลายเดือนก่อน

      So does it gives default depth in meters can we get that without training​@@NicolaiAI

  • @malek3764
    @malek3764 4 หลายเดือนก่อน

    Thanks for this video!
    I just need to ask on what hardware did you run this for real-time performance? And what FPS (or inference time in milliseconds) did you get for each model?

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +1

      Thanks for watching! Just from raw model I get around 50 fps or so. Can definitely be bumped up with optimization and deleting the visualizations etc. This is only a 25mb model so very small vision model! Can run pretty fast. I’m running this on a 3070

    • @malek3764
      @malek3764 4 หลายเดือนก่อน +1

      @@NicolaiAI wow that's actually impressive! thanks alots

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +1

      @malek3764 yeah that’s their small model they also have way larger models which can’t run real time. But now we have amazing results in real time

    • @malek3764
      @malek3764 4 หลายเดือนก่อน

      @@NicolaiAI based on ur experiments with the models, how big is the gap in quality between the small one and the larger versions?

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +1

      @@malek3764 not much! Definitely go with the smaller models unless you do some 2D to 3D stuff and those things where you want the highest detail

  • @vipulpardeshi2868
    @vipulpardeshi2868 4 หลายเดือนก่อน +1

    Hey i just wanna know what will be this models performance on Embedded Devices like Jetson or Raspberry Pi

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +1

      Can prob run a few frames on a jetson

    • @vipulpardeshi2868
      @vipulpardeshi2868 4 หลายเดือนก่อน +1

      Thanks for the reply😊. Just one more question Can we get the depth data in like centimetres using this model

  • @keshav2136
    @keshav2136 4 หลายเดือนก่อน

    The video is nice. Why don't you provide required links in the captions?

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน +1

      Thanks! Will upload to my GitHub today! Had some urgent stuff that came up right at release

    • @keshav2136
      @keshav2136 4 หลายเดือนก่อน

      @@NicolaiAI Alright, fair enough, and I've seen your other videos as well for Stereo Cam Calibration and now this Mono Camera Depth Estimation, it would be SOO helpful if you provide links to the targetted GitHub repos or links in the description or Comment section.
      We literally are the ones who watch your videos full from the beginning to the end. So please, take care of that.
      Thanks 😃

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน

      I appreciate all of you! I promise ill do my very best and do that going forward
      @@keshav2136

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน

      github.com/niconielsen32/depth-anything

  • @cg.man_aka_kevin
    @cg.man_aka_kevin 3 หลายเดือนก่อน +2

    Can this code run in Python?

    • @NicolaiAI
      @NicolaiAI  3 หลายเดือนก่อน +1

      Yup this is running in python

    • @cg.man_aka_kevin
      @cg.man_aka_kevin 3 หลายเดือนก่อน +1

      ​@@NicolaiAI Owh, because I saw the icon software doesn't look like Python... 😅😅😅

  • @kirkeby7875
    @kirkeby7875 4 หลายเดือนก่อน

    Siger ikke hele ideen til min bachelor lige røg i vasken, men jeg bruger stereo-vision med IR-projektør... Fedt man nu kan bare med et enkelt kamera... yay! 🤣🤣

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน

      Haha sorry 😂

    • @kirkeby7875
      @kirkeby7875 4 หลายเดือนก่อน +1

      @@NicolaiAI Tak for altid at poste den nyeste viden, det holder os på dupperne 💪

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน

      @@kirkeby7875 mange takker for at følge med!

  • @ruanjiayang
    @ruanjiayang 3 หลายเดือนก่อน

    What is "relative depth"?

    • @donrikk8546
      @donrikk8546 2 หลายเดือนก่อน

      depth relative to the camera/viewer i would assume. since depth can be captured in different ways i think the term relative in this case is just a designation for the camera it is being captured on mimmicking depth youde produce using both of your eyes naturally, thats why this depth map is also used in 2d to 3D video conversions cuz it helps mimmick stereo view like the way studios do it

  • @LukasSmith827
    @LukasSmith827 4 หลายเดือนก่อน +1

    Nicolai Nielsen our beloved

    • @NicolaiAI
      @NicolaiAI  4 หลายเดือนก่อน

      Thanks a ton mate!

  • @NurbNerb
    @NurbNerb หลายเดือนก่อน +1

    Not anymore!

    • @NicolaiAI
      @NicolaiAI  หลายเดือนก่อน

      What’s new?

  • @Nourhan-jw8iv
    @Nourhan-jw8iv 4 หลายเดือนก่อน

    can we print out the depth map as we did in the MiDaS model ?