Reinforcement Learning Course - Full Machine Learning Tutorial

แชร์
ฝัง
  • เผยแพร่เมื่อ 22 ธ.ค. 2024

ความคิดเห็น • 238

  • @MachineLearningwithPhil
    @MachineLearningwithPhil 5 ปีที่แล้ว +235

    Here are some time stamps folks!
    Intro 00:00:00
    Intro to Deep Q Learning 00:01:30
    How to Code Deep Q Learning in Tensorflow 00:08:56
    Deep Q Learning with Pytorch Part 1: The Q Network 00:52:03
    Deep Q Learning with Pytorch part 2: Coding the Agent 01:06:21
    Deep Q Learning with Pytorch part 3: Coding the main loop 01:28:54
    Intro to Policy Gradients 01:46:39
    How to Beat Lunar Lander with Policy Gradients 01:55:01
    How to Beat Space Invaders with Policy Gradients 02:21:32
    How to Create Your Own Reinforcement Learning Environment Part 1 02:34:41
    How to Create Your Own Reinforcement Learning Environment Part 2 02:55:39
    Fundamentals of Reinforcement Learning 03:08:20
    Markov Decision Processes 03:17:09
    The Explore Exploit Dilemma 03:23:02
    Reinforcement Learning in the Open AI Gym: SARSA 03:29:19
    Reinforcement Learning in the Open AI Gym: Double Q Learning 03:39:56
    Conclusion 03:54:07

  • @real-chipmunk
    @real-chipmunk 7 หลายเดือนก่อน +364

    Anytime I fall asleep to anything I watch these videos haunt my youtube. I never intentionally watch this channel. What the flip guys?

    • @dawidchlebek3251
      @dawidchlebek3251 7 หลายเดือนก่อน +32

      broooooo, the last couple of nights when I fell asleep and my laptop was goin, Ive had these videos playing when i wake up in the morning.

    • @Thony91
      @Thony91 7 หลายเดือนก่อน +10

      Same!

    • @jochemvdberg8898
      @jochemvdberg8898 7 หลายเดือนก่อน +9

      Yes i have the same thing. Im 2 hours in at the moment 😂

    • @Mothaf4ckajones
      @Mothaf4ckajones 7 หลายเดือนก่อน +1

      Same

    • @shans2707
      @shans2707 7 หลายเดือนก่อน +7

      Same here! something is wrong with TH-cam algo

  • @kiransilwal
    @kiransilwal 3 หลายเดือนก่อน +20

    Deep Reinforced Sleeping shall
    be the title. Please rename it. Woke up to this. 3.5 hours in!

  • @A.h0808
    @A.h0808 9 หลายเดือนก่อน +127

    Bro I fell asleep watching a different video and woke up this morning to this video playing and I was an hour deep in it 😭

    • @kaylor87
      @kaylor87 6 หลายเดือนก่อน +8

      Lol same except I was 3.5hrs through when I woke up 😁

    • @__j_o_s__
      @__j_o_s__ 6 หลายเดือนก่อน +3

      Me too 🤣🤣🤣🤣

    • @DariusDareDevil
      @DariusDareDevil 5 หลายเดือนก่อน +2

      Me too, just now I woke up to this whaaatt 😂

    • @theodosiushartmann9448
      @theodosiushartmann9448 5 หลายเดือนก่อน

      lol same xd

    • @riflehitta-4157
      @riflehitta-4157 3 หลายเดือนก่อน +2

      same. what is bru even talkin about?🤣

  • @TBPony
    @TBPony 4 หลายเดือนก่อน +2

    The nerd talk and keyboard typing is very ASMR and it helps me sleep. Put a mic at the keyboard itself and edit it in, this is wonderful and i might actually wake up smarter in the morning

  • @amitbuch
    @amitbuch 2 ปีที่แล้ว +12

    Extremely well explained. Kudos to the tutor. Simple explanation to workign code in less than an hour is amazing and yet very clearly laid out. Thanks for this upload.

  • @SmartassEyebrows
    @SmartassEyebrows 5 ปีที่แล้ว +17

    One minor correction for those watching at 1:19:12 and trying to follow along (like myself): on line 77 after the "else", the "memStart = int(np.random.choice(range(self.memCntr - batch_size - 1)))" should actually be "memStart = int(np.random.choice(range(self.memSize - batch_size - 1)))".
    The self.memSize is needed here instead of self.memCntr because at this point the self.memory list is now full (the "else" branch), but the self.memCntr value is continuing to grow and is now larger than the max self.memory size. That leads to line 78 giving miniBatch an empty list, [ ], leading to memory being an empty array, because memStart will be a larger value than the self.memory list length, while then being used as the index for grabbing the miniBatch from that same self.memory list -- no good. Ultimately that leads to an exception: "too many indices for array" on line 81 (since we are trying to forward an empty 1D numpy array and call 2D indices that don't exist). With self.memSize for line 77, that no longer happens, and memStart stays within the bounds of the self.memory length/size. With that, everything works, and you can watch the agent play :)

  • @InturnetHaetMachine
    @InturnetHaetMachine 5 ปีที่แล้ว +83

    This is a great video if you already understand the topic, understand the code and just want a guy saying what he's typing out aloud, kinda explaining bits and pieces here and there.

    • @henrypowell3496
      @henrypowell3496 3 ปีที่แล้ว +10

      yah and people in the comment section be like: thank you, what a great tutorial for free, great explanation, while they get nothing and just being smart in typing a comment

    • @ramtinnazeryan
      @ramtinnazeryan 2 ปีที่แล้ว +5

      He said at the beginning no need to know about this and that. 14 minutes into the video he is typing the line 123. Honestly why didn't he copy and paste it? :))))

    • @surajbherwani5488
      @surajbherwani5488 ปีที่แล้ว

      Worst video

    • @elasdebastos235
      @elasdebastos235 ปีที่แล้ว

      yeah, i hae the same feeling, i didn t undesrtand a crap
      @@ramtinnazeryan

    • @Jesus_was_God
      @Jesus_was_God 8 หลายเดือนก่อน

      Its a "course" - so it is what it is :D . Take it as an overview to the topic, watch him code to see whats up, but not code urself here.

  • @MrMarkyr1000
    @MrMarkyr1000 6 หลายเดือนก่อน +1

    Thanks!

  • @vsp540
    @vsp540 5 ปีที่แล้ว +5

    ⭐️ Course Contents ⭐️ ⌨️ (00:00:00) Intro ⌨️ (00:01:30) Intro to Deep Q Learning ⌨️ (00:08:56) How to Code Deep Q Learning in Tensorflow ⌨️ (00:52:03) Deep Q Learning with Pytorch Part 1: The Q Network ⌨️ (01:06:21) Deep Q Learning with Pytorch part 2: Coding the Agent ⌨️ (01:28:54) Deep Q Learning with Pytorch part ⌨️ (01:46:39) Intro to Policy Gradients 3: Coding the main loop ⌨️ (01:55:01) How to Beat Lunar Lander with Policy Gradients ⌨️ (02:21:32) How to Beat Space Invaders with Policy Gradients ⌨️ (02:34:41) How to Create Your Own Reinforcement Learning Environment Part 1 ⌨️ (02:55:39) How to Create Your Own Reinforcement Learning Environment Part 2 ⌨️ (03:08:20) Fundamentals of Reinforcement Learning ⌨️ (03:17:09) Markov Decision Processes ⌨️ (03:23:02) The Explore Exploit Dilemma ⌨️ (03:29:19) Reinforcement Learning in the Open AI Gym: SARSA ⌨️ (03:39:56) Reinforcement Learning in the Open AI Gym: Double Q
    Learning ⌨️ (03:54:07) Conclusion

  • @claude.detchambila
    @claude.detchambila 2 ปีที่แล้ว +4

    This is one of the best free RL videos available. Please make some more.

  • @aaryasankhe5973
    @aaryasankhe5973 3 ปีที่แล้ว +23

    Anyone interested in learning the terminologies of what he is talking about should go check out the video lectures Stanford did on MDPs(Markov decisions processes and RL), they're about each an hour long and do go in depth behind the math for a lot of this stuff. Cheers!!!

    • @AnthonyAnalog
      @AnthonyAnalog 3 ปีที่แล้ว +1

      Underrated. Thank you.

  • @evanstark4823
    @evanstark4823 3 ปีที่แล้ว +16

    man i just fell asleep on youtube and now i’ve been watching this for 2 hours 19 minutes

    • @danilkos7411
      @danilkos7411 2 ปีที่แล้ว +2

      I sleep to this all night. TH-cam knows I’m using it as white noise at bedtime and offers sleepy stuff like this 😬

  • @petermcfarland9236
    @petermcfarland9236 2 ปีที่แล้ว +2

    This video has helped me find clues that ultimately helped me to understand machine learning. Thanks blue Steve.

  • @Bill0102
    @Bill0102 10 หลายเดือนก่อน +2

    I'm immersed in this. I read a book with a similar theme, and I was completely immersed. "The Art of Saying No: Mastering Boundaries for a Fulfilling Life" by Samuel Dawn

  • @barbellbender4232
    @barbellbender4232 5 ปีที่แล้ว +7

    The length of the flatten outputlayer can actually be calculated from first conv layer tracing the data through the network. Just use the function:
    ((dimension length - kernal size for the dimension + 2*padding)/stride)+1 = output length for the dimension
    do this for each dimension for each conv layer and multiply by number of outputs in the end to find the length of the flat dimension as such:
    1st conv layer: ((185 - 8 + 2*1)/4) + 1 = 44 (acutally 44.75 but you always round down, since there are no 0.75 pixels)
    ((95 - 8 + 2*1)/ 4) + 1 = 22 (rounded down from 22.25)
    2nd conv: ((44 - 4 + 2*0)/2) + 1 = 21
    ((22 - 4 + 2*0)/2) + 1 = 10
    3rd conv: ((21 - 3 + 2*0)/1) + 1 = 19
    ((10 - 3 + 2*0)/1) + 1 = 8
    this means the 3rd layer outputs 128 frames with each having dimensions 19*8 and therefore if you wanted to flatten them into one you would get one dimension with 128*19*8 vectors.
    Just neat little trick for those who want it

  • @bzknorthstar
    @bzknorthstar 2 หลายเดือนก่อน +1

    8 minutes in right now I’m cracking uppppp what is this?!

  • @mehershrishtinigam5449
    @mehershrishtinigam5449 ปีที่แล้ว

    thanks for taking time out your busy day to teach us RL fellow martian!

  • @fabrizioantonazzo3113
    @fabrizioantonazzo3113 3 ปีที่แล้ว +1

    grazie mille ho iniziato da poco a seguire il tuo corso ben fatto

  • @dummypg6129
    @dummypg6129 4 ปีที่แล้ว +41

    This is the only issue that i often see on any "basic tutorial" videos. There's no explaination on the terminologies during the intros.

    • @constantinetikh
      @constantinetikh 3 ปีที่แล้ว +6

      The terminology and concepts are explained in the two blocks starting 03:08:20

  • @lsagar
    @lsagar 5 ปีที่แล้ว +13

    I'm a beginner and the Background loop seems more interesting than what he's talking. I hope I understand what he's saying someday

  • @zoerycroft4300
    @zoerycroft4300 2 ปีที่แล้ว +1

    for anyone watching, inheriting from object is implied, you haven't needed to type that since even the oldest versions of python 3, save yourself some time ;) `class foo(object):` is exactly the same as `class foo:`
    the reason he types it here is probably for intercompatability between py2 and py3, but not even a year after this was uploaded py2 went end of life, so you shouldn't need to worry about that anymore :))

  • @natashasamuel9346
    @natashasamuel9346 3 ปีที่แล้ว +1

    Great class.
    Keep up the good work.
    Thank You,
    Natasha Samuel

    • @henrypowell3496
      @henrypowell3496 3 ปีที่แล้ว

      so you understood everything in the tutorial?

  • @sunmustbedestroyed
    @sunmustbedestroyed 5 ปีที่แล้ว +107

    Heads up: this one isn't for beginners.

    • @opkp
      @opkp 4 ปีที่แล้ว +1

      @Black Hole lel , got him good

  • @joafus
    @joafus 7 หลายเดือนก่อน +3

    Woke up to this

  • @amelielikethemovie
    @amelielikethemovie 29 วันที่ผ่านมา

    thanks I woke up nicely to this

  • @DynesLair-kb6qs
    @DynesLair-kb6qs 2 หลายเดือนก่อน +6

    I fell asleep watching a documentary about government corruption... I woke up 3.5 hours into this video instead somehow...

  • @Johniakson
    @Johniakson 5 ปีที่แล้ว +4

    This is AWESOME! 👍👍👍 Thank you for this!

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 5 ปีที่แล้ว +9

    what can you recommend to watch of your other youtube videos before watching this one?

    • @matthimuller
      @matthimuller 5 ปีที่แล้ว +1

      python guides and some linear algebra to understand what is happening

    • @kae4881
      @kae4881 4 ปีที่แล้ว

      Check out his channel
      his original one.

  • @that_endo
    @that_endo 2 ปีที่แล้ว +1

    2nd time waking up to this guy

  • @johnnychapman6322
    @johnnychapman6322 3 หลายเดือนก่อน +2

    I just woke up 3 hours in . Not a normal listen for me🤔 but the dream I was having 🤯

  • @charlespeck5493
    @charlespeck5493 2 ปีที่แล้ว

    easy to understand, much pleasure

  • @remoncomputer
    @remoncomputer 4 ปีที่แล้ว

    Good awsome view and awsome channel, good work

  • @CaribouDataScience
    @CaribouDataScience 2 ปีที่แล้ว +3

    Reinforced learning, is like when you have to write 1000 times "I will not talk in class"?

  • @멋진남자-q1t
    @멋진남자-q1t 3 ปีที่แล้ว

    I saw this video
    I'll see this video again
    Thank you

  • @0GRANATE0
    @0GRANATE0 2 ปีที่แล้ว

    When I wanted to implement a multi agent reinforcement learning envirenment for a soccer game (multiple agents, separately trained, with separate models) what algorithms would I use for a continious envirenment (not a grid world) where the players can walk/run/shot everywhere? - DQN Reinforcement Learning?

  • @billchenxi
    @billchenxi 5 ปีที่แล้ว +2

    Do you know why his kernel size is even number? Normally we use odd number for easy calculation.

    • @underlecht
      @underlecht 3 ปีที่แล้ว

      why is that a difference?

    • @billchenxi
      @billchenxi 3 ปีที่แล้ว

      ​@@underlecht Because a non-symmetric kernel (even number) yields a non-symmetric filter response. In the example above, this non-symmetry leads to a shift of the blurred image by half a pixel.

  • @ShapeShifter369
    @ShapeShifter369 8 หลายเดือนก่อน

    Thank You a lot !!!!!!!!!!!!!!!!!!!!

  • @imtiazahmad7826
    @imtiazahmad7826 5 ปีที่แล้ว +2

    thanks, so helpful video

  • @kerryd2060
    @kerryd2060 5 ปีที่แล้ว +2

    Thank you. Great job in explaining the content.

  • @sebtaciak
    @sebtaciak ปีที่แล้ว

    keep going with Sutton but please move to DLR too!

  • @firelordzaki1600
    @firelordzaki1600 5 ปีที่แล้ว +2

    Wow fcc really stepping up

  • @freetellurian
    @freetellurian 10 หลายเดือนก่อน

    ✍️👍 Больше спасибо

  • @nicklansbury3166
    @nicklansbury3166 5 ปีที่แล้ว +7

    You had me at "Atari"!!!

  • @carlosalbertoleyva6460
    @carlosalbertoleyva6460 5 ปีที่แล้ว +5

    How exactly do I run these python programs?
    I'm using Atom IDE

    • @kae4881
      @kae4881 4 ปีที่แล้ว

      You can run via the terminal(linux and macOS)/command prompt(windows) or you can download the Run Script addon in Atom(that's what i use)

  • @finarwa3602
    @finarwa3602 3 ปีที่แล้ว +2

    I really appreciate the work you are doing . Could you mention which the development tool you are
    using for the whole series?

    • @Beny123
      @Beny123 2 ปีที่แล้ว +1

      Vscode

  • @emanuelepapucci59
    @emanuelepapucci59 2 ปีที่แล้ว

    my utils package do not have any plotLearning func...

  • @butter_pp
    @butter_pp 4 หลายเดือนก่อน +1

    日本人だけど眠る度にこの動画にたどり着く

  • @YashPrabhu-o7n
    @YashPrabhu-o7n 21 วันที่ผ่านมา +1

    Fell asleep woke up to 2:12:00

  • @imrielseraphis
    @imrielseraphis 5 ปีที่แล้ว +3

    I am currently creating an agent-based model that will generate x numver of agents. Each agent has a step function. I would LOVE to incorporate this reinforced learning method into the model. How would you adjust it from taking a visual frame like from a game to using only the global environment variables? Is it simply as easy as swapping out one for the other?

  • @សេងហ៊ាង-ប6ឋ
    @សេងហ៊ាង-ប6ឋ 2 ปีที่แล้ว

    អរគុណ

  • @YeaBroer
    @YeaBroer 25 วันที่ผ่านมา

    Everytime i fall asleep i wake up to this, is it a sign that I should learn coding?

  • @mafiamustafa
    @mafiamustafa 5 ปีที่แล้ว

    The stacking frame reshaping code is wrong, it messes up the entire array. I am trying to debug it.

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 5 ปีที่แล้ว +1

    are you using tensor flow 2.0?

    • @kae4881
      @kae4881 4 ปีที่แล้ว

      No, it's tensorflow 1.4.10. If you want to look at tensorflow 2.0 or keras, go check out Phil's TH-cam channel

  • @timraiser4811
    @timraiser4811 2 ปีที่แล้ว

    I get errors when i run the pip install commands for box2d-py, tensorflow-gpu and torch. Is there something i'm missing?

  • @CharlieBahr
    @CharlieBahr 7 หลายเดือนก่อน +1

    Keyboard sound is disturbing.

  • @mafiamustafa
    @mafiamustafa 5 ปีที่แล้ว +1

    Hello Phil, I think there is another mistake in the code, in the learn function it should be reward_batch + gamma*np.max(Q_next, axis=1)*(1-terminal_batch) instead of just terminal_batch. Since we are passing int(done) as a stored observation. Therefore for done=False, int(done)=0 and vice versa. And if episode does not end that is done equals False then we need to add the next Q_value otherwise we only add reward. What do you think? Am I correct?

    • @kae4881
      @kae4881 4 ปีที่แล้ว

      yeah i do think so. but i encourage you to try it both ways. sometime it trains the agent to finish the episode as fast as possible, that's not what we want.

    • @krishnendubose2620
      @krishnendubose2620 3 ปีที่แล้ว

      Could you post the link to the lectures here?

  • @YMARihab
    @YMARihab 5 ปีที่แล้ว

    Can I really get a job after learning stuff like machine learning and etc online? I did BSc in EEE but as I was not interested I did no do well and did not learn much, only thing I somewhat enjoyed was the C and C++ courses and digital logic design and verilog/vhdl courses . Later I did python course from MITx and enjoyed solving all the exercises. Pls give me hope.

    • @YMARihab
      @YMARihab 5 ปีที่แล้ว

      wild doggo appears. wild doggo regrets wasting his time.

  • @FundamentalistLohar
    @FundamentalistLohar 3 ปีที่แล้ว

    raise error.UnregisteredEnv("No registered env with id: {}".format(id))
    gym.error.UnregisteredEnv: No registered env with id: Breakout-v0

  • @0GRANATE0
    @0GRANATE0 2 ปีที่แล้ว

    How comes that I am not capable to UNDERSTAND this stuff? Is it OK to accept that I am dumb? Serious question... (I am a software engineer, did computer science, but had always problem with math, but did always the best mark when I had to deliver a project..)

  • @BharatwaajShankar
    @BharatwaajShankar 5 ปีที่แล้ว

    Why are we onehot encoding the actions?

  • @KushChoudhary
    @KushChoudhary 5 ปีที่แล้ว +7

    Let me tell you
    It is so good :)

  • @ShahFahad-hj1ps
    @ShahFahad-hj1ps 2 ปีที่แล้ว

    Is there any course (Reinforcement Learning code based) for beginners ?

    • @innerdragonfruit
      @innerdragonfruit ปีที่แล้ว

      Coursera has a really cool 4-course series "Reinforcement Learning Specialization"

  • @devsinharoy8811
    @devsinharoy8811 4 ปีที่แล้ว

    when you do def(build network) wouldnt it be easier just to use keras

  • @foundev
    @foundev 4 ปีที่แล้ว +2

    Clearly the video is only for people who have already researched about RL. Not for beginners at all!

  • @IveReignSupreme
    @IveReignSupreme หลายเดือนก่อน

    I fell asleep landing on this video too 😂

  • @jordanolson11
    @jordanolson11 5 ปีที่แล้ว

    so it seems I can't even install this or get started, I'm on windows 10 and I have python 3.7 working, I installed pip, gym, but when I got to tensorflow it's telling me I have "no matching distribution found for tensorflow-gpu", some people suggested that it's because I got the latest version of py, others suggest to use anaconda, what should I do?

    • @jordanolson11
      @jordanolson11 5 ปีที่แล้ว

      so after a few hours I was able to figure out that anaconda and miniconda are the same thing and going through the github repo for box2d-py it just says to use anaconda, but I'd rather follow how you're doing this vs using a stupid VM with py on it, seems really ridiculous that installing a few things you show in your command line running so fluidly, like what am I missing? I reinstalled python 4 times and made sure my env variables were working correctly, each with a different python version each time and they all gave the same error with pip installing box2d-py and tensorflow-gpu, including 3.6 which is the version you're on. What shell are you using? I'm just using command prompt and I wonder if that's the problem.

    • @jordanolson11
      @jordanolson11 5 ปีที่แล้ว

      lol after buckling and trying anaconda tensorflow fails, but everything else runs properly. Really fun tutorial.

    • @MachineLearningwithPhil
      @MachineLearningwithPhil 5 ปีที่แล้ว

      @@jordanolson11 Sorry, just seeing this now. To the best of my knowledge, Tensorflow only works with python 3.6. You can just install 3.6 in parallel with other versions, without nuking the whole install, I believe.

    • @1ycx
      @1ycx 5 ปีที่แล้ว

      Use Google colaboratory

    • @1ycx
      @1ycx 5 ปีที่แล้ว

      Use Google Colaboratory

  • @vasanthrs1140
    @vasanthrs1140 5 ปีที่แล้ว

    Can you please make a full tutorial in flutter?
    Thanks, I've been watching your tutorial for a long time.
    Excellent work.

    • @freecodecamp
      @freecodecamp  5 ปีที่แล้ว +2

      We're posting a full flutter tutorial next week. :)

    • @rbalachandran007
      @rbalachandran007 5 ปีที่แล้ว

      Thanks

    • @vasanthrs1140
      @vasanthrs1140 5 ปีที่แล้ว

      @@freecodecamp Thank you so much,
      I'm looking forward for that tutorial.

  • @tariqshah2767
    @tariqshah2767 3 ปีที่แล้ว

    Is there anyone who knows which version of TensorFlow is used in this video?

  • @renaton3992
    @renaton3992 3 ปีที่แล้ว +7

    Yep... He is theaching for who already knows.....

  • @mafiamustafa
    @mafiamustafa 5 ปีที่แล้ว

    self.q = tf.reduce_sum(tf.multiply(self.Q_values, self.actions))
    Why are you doing this? I fail to understand the meaning of this line?? Thank you in advance :)

    • @mt345678
      @mt345678 5 ปีที่แล้ว +1

      he made a mistake.... the github code is corrected.

    • @mafiamustafa
      @mafiamustafa 5 ปีที่แล้ว

      @@mt345678 I see so many other tutorials which do the same thing, and I fail to understand why.

  • @IND_development
    @IND_development 4 หลายเดือนก่อน

    This course is reinforcement Machine Learning or Deep learning????

  • @KleevesDoesStuff
    @KleevesDoesStuff หลายเดือนก่อน

    Bro I fell asleep with my phone on and this is what I wake up to

  • @EdgarBalderas-f6c
    @EdgarBalderas-f6c 11 หลายเดือนก่อน

    새해 첫날 듣는 노래가 그 해 운이 달렸다는거 몰라서 2023년 첫곡을 니소식으로 들었다가 이틀만에 사귀던 남친이랑 헤어졌었는데 2024년은 내가제일잘나가 듣고 완전 끝내주게 살아보려고! 1년내내 연애도 끝났고 직장도 안좋게되서 속상했는데 겨우 [빠]칭코닷[컴]하면서 잊고지내고 우울한거도 잊으면서 지냈는데 2024년에는 더 잘될 수 있을것같아! 이제 돈도 잘 벌고 내가 더 행복한 사람되야지!

  • @abhishekshah11
    @abhishekshah11 5 ปีที่แล้ว

    How much linear algebra and statistics should I know for this track?

    • @MachineLearningwithPhil
      @MachineLearningwithPhil 5 ปีที่แล้ว +3

      Not much is needed. I explain it all as we go along.

    • @kae4881
      @kae4881 4 ปีที่แล้ว

      not much alegbra and statistics i can say, but more about the basic reinforcement learning terms

  • @joebarco5674
    @joebarco5674 5 ปีที่แล้ว

    Is this something someone can learn without previous experience?

    • @MachineLearningwithPhil
      @MachineLearningwithPhil 5 ปีที่แล้ว +1

      You sure can, I'll cover the basics as we go along.

    • @kae4881
      @kae4881 4 ปีที่แล้ว +1

      Not really, i suggest you read about reinforcement learning basics before you watch this.

  • @haneulkim4902
    @haneulkim4902 3 ปีที่แล้ว

    Amazing course, thanks alot Phil! One question, you were comparing policy gradient methods with reinforcement learning however after few searches it seems like policy gradient method is an algorithm within RL. Could you clarify?

  • @mindflayer99
    @mindflayer99 4 ปีที่แล้ว

    ok guys where is the code available

  • @cactus_duke1639
    @cactus_duke1639 5 ปีที่แล้ว

    What program do you use to run your code?

    • @kae4881
      @kae4881 4 ปีที่แล้ว

      It's the linux distribution native terminal, not sure about the exact distribution. Maybe ubuntu

  • @luci0rullz
    @luci0rullz 4 ปีที่แล้ว

    Phil, you are a fucking boss :>

  • @honkmangoose6567
    @honkmangoose6567 2 ปีที่แล้ว

    Steve from blue's clues wasn't joking when he said he was going to college

  • @jansprlak4157
    @jansprlak4157 2 ปีที่แล้ว

    Beta tester od roku 2016 oceňujem,,💖💖💝

  • @boxsewatiodoxsewstio1355
    @boxsewatiodoxsewstio1355 2 ปีที่แล้ว

    nice

  • @MikePianoMan85
    @MikePianoMan85 4 วันที่ผ่านมา

    Why tf does this keep playing everytime I just keep videos playing?

  • @lalomega7966
    @lalomega7966 5 ปีที่แล้ว

    This is awesome..

  • @gorimor
    @gorimor 2 ปีที่แล้ว

    12 minutes in and scratching my head. Hi yes I'm lost

  • @umairnehri8459
    @umairnehri8459 5 ปีที่แล้ว +1

    💯💯💯

  • @ankitkpandey369
    @ankitkpandey369 2 ปีที่แล้ว +2

    Not useful. Creator should at least run a code once before wrapping up a section. Small syntax/logical errors are fine but when there is error in code itself (eg: new_state_batch in first section) and there isn't much explanation about it, whole hour spent in coding goes for nothing.

  • @AbhishekJain-zu1uf
    @AbhishekJain-zu1uf 5 ปีที่แล้ว

    i can hardly hear anything

  • @niklasdamm6900
    @niklasdamm6900 2 ปีที่แล้ว

    25.10.22 22:30

  • @isalihkilic
    @isalihkilic 4 ปีที่แล้ว

    Information is so good but the background is grabbing the attention away from the information.

  • @hninnwe6015
    @hninnwe6015 5 ปีที่แล้ว

    Please use some illustrations so that can understand more easily

  • @SphereofTime
    @SphereofTime 5 หลายเดือนก่อน

    12:34

  • @motbus3
    @motbus3 4 ปีที่แล้ว +1

    Is this tf1 or tf2?

    • @kae4881
      @kae4881 4 ปีที่แล้ว

      tf1

  • @joeljoseph26
    @joeljoseph26 3 ปีที่แล้ว

    When you mentioned a course on Full Machine Learning Tutorial - Reinforcement Learning and there is no proper order to it. I don't recommend watching this thing. There are tons of materials that are a lot simpler than this.

  • @sihabportal8623
    @sihabportal8623 5 ปีที่แล้ว

    machine learning is really trending matter

  • @MuhammadAhmed-vi2xt
    @MuhammadAhmed-vi2xt 5 ปีที่แล้ว +1

    please make complete course on WordPress 5.2

  • @mmanuel6874
    @mmanuel6874 5 ปีที่แล้ว +1

    YEET!

  • @robihamdani4017
    @robihamdani4017 5 ปีที่แล้ว +2

    i feel uncomfortable with his hands

  • @saeedalafifi0
    @saeedalafifi0 5 ปีที่แล้ว

    Thanks for your information
    Can you help me to set an environment to trade forex and how to set it with metatrader 4
    Please help me