Build a Custom Gymnasium Reinforcement Learning Environment & Train w Q-Learning & Stable Baselines3

แชร์
ฝัง
  • เผยแพร่เมื่อ 2 ธ.ค. 2024

ความคิดเห็น • 23

  • @johnnycode
    @johnnycode  15 วันที่ผ่านมา

    Ready to get started with Stable Baselines3? th-cam.com/video/OqvXHi_QtT0/w-d-xo.html

  • @graycomet
    @graycomet 16 วันที่ผ่านมา

    Thanks for the great video. It makes me easier to build my own training environments.

  • @a_samad
    @a_samad 7 หลายเดือนก่อน +3

    Thank You!🎉
    We need more videos (course) like that: create custom env using open ai gym.

  • @Ayhamtechtips
    @Ayhamtechtips 8 หลายเดือนก่อน +2

    Very helpful and nice explanation.
    Thanks!

  • @hrsharma02
    @hrsharma02 7 หลายเดือนก่อน +2

    Thanks, dear 🎉, Please upload more videos on multi-agent RL for robotics and path planning for multi-robot with custom environment in Gymnasium.

  • @navaneethbuilds
    @navaneethbuilds 4 หลายเดือนก่อน +1

    thanks for the amazing video

  • @ashutoshmishra5901
    @ashutoshmishra5901 7 หลายเดือนก่อน +4

    Your Explanation is too good. I have a humble request. Can you make RL training using MuJoCo Ant but registering it as custom environment. The GAIT parameter generation is quite treaky. If possible please make a tutorial on it.

    • @towerboi-zg3it
      @towerboi-zg3it 7 หลายเดือนก่อน

      I also want this tutorial, so please Johnny

  • @buzzbuzz1691
    @buzzbuzz1691 8 หลายเดือนก่อน +1

    Thank you

  • @rickyolal
    @rickyolal 5 หลายเดือนก่อน +1

    Hey Johnny, I was wondering if you knew how to make the algorithm learn some already known states? I have a challenge related to make a DQN learn and start with already known states stored in a csv file, and I am struggling because I have no idea how to do that. Is it possible?

    • @johnnycode
      @johnnycode  5 หลายเดือนก่อน

      I'm guessing if you know those states, then you would know what action to take or not take in relation to those states. For example, a pawn on a chess board can't go backwards, since you know that state is impossible. If my interpretation of your question is correct, then you might want to look into "action masking", which prevents the agent from taking illegal actions. You can start with this SB3 reference, but the concept is not limited to PPO: sb3-contrib.readthedocs.io/en/master/modules/ppo_mask.html

  • @BoxingBytes-y9i
    @BoxingBytes-y9i 6 หลายเดือนก่อน

    Thanks for your video. Which ressources would you adivse to learn practical applications of reinforcement learning? I've been trying to implement a bot for a specific game and have to create my own environment and DQN. I'm familiar with neural nets, but all the rest is so hard to find good information on

    • @johnnycode
      @johnnycode  6 หลายเดือนก่อน

      Sorry, I'm not an expert. I suggest inquiring at the r/reinforcementlearning subreddit. There are some very knowledgeable people there.

    • @BoxingBytes-y9i
      @BoxingBytes-y9i 6 หลายเดือนก่อน

      @@johnnycode Thank you for the answer, will do!

  • @tieqi5623
    @tieqi5623 6 หลายเดือนก่อน

    Thanks, good video. Dose Gymnasium can support NeoGeo (SNK) roms? How to make it to support?

    • @johnnycode
      @johnnycode  6 หลายเดือนก่อน

      It doesn’t support Neo Geo roms. I think it would be extremely hard to bridge that support.

  • @arnavmodanwal6295
    @arnavmodanwal6295 5 หลายเดือนก่อน

    Hi, your videos are great and helped me a lot since you were using the latest version of stable baseline3...But I am facing an issue that the verbose values are not getting printed in output I have put verbose = 1 and even tried to use verbose = 2 but not getting the desired outputs (like rewards, loss, iterations, ep_len_mean etc.) as it was getting printed in your videos. Can you please help me? Is this due to the custom environment I am using or something else?
    Also, tensorboard logs are also not working...

    • @johnnycode
      @johnnycode  5 หลายเดือนก่อน

      You should try creating a new conda environment and then install SB3 again. In my SB3 introduction video, I just ran pip install stable-baselines3[extra] and didn't do anything else special: th-cam.com/video/OqvXHi_QtT0/w-d-xo.html

    • @arnavmodanwal6295
      @arnavmodanwal6295 5 หลายเดือนก่อน

      @@johnnycode Hi, I will try this one again...Thanks a lot for the reply and your time! Might need your help again...

    • @arnavmodanwal6295
      @arnavmodanwal6295 5 หลายเดือนก่อน

      Hi, @johnnycode, I tried reinstalling the stable-baselines3[extras] but I am not getting the monitor data also the tensorboard logs are also not getting displayed...Is there some issue with the new version of stable-baselines3[extra] can you please give me the version you installed when making the video?

    • @johnnycode
      @johnnycode  5 หลายเดือนก่อน

      stable-baselines3 2.0.0
      tensorboard 2.13.0

  • @sergiogirona2988
    @sergiogirona2988 7 หลายเดือนก่อน

    Could i use it for own game made it with Godot Engine?? Thanks!!

    • @johnnycode
      @johnnycode  7 หลายเดือนก่อน

      Yes, of course!