Batch Normalization in Deep Learning | Batch Learning in Keras

แชร์
ฝัง
  • เผยแพร่เมื่อ 25 พ.ย. 2024

ความคิดเห็น • 97

  • @balrajprajesh6473
    @balrajprajesh6473 2 ปีที่แล้ว +96

    I can't even imagine the amount of effort you put behind making these videos. When I have to explain something worth 10 minutes to my friends, I have to prepare myself for hours. You make these long well researched videos. I really can't imagine the effort you have to put. You are really a gem. Thank you very much for doing this.

    • @tejasburadkar189
      @tejasburadkar189 2 ปีที่แล้ว +5

      Exactly true bro...personally I want say ….I never seen such informative channel on TH-cam that give detail information on machine learning topics. After seeing the 100 days of machine learning series my ml knowledge is totally boost up and Thank you sir for Everything. hope sir you are going to make video playlist on image processing... take one image apply filter and process the image

    • @alokdixit5835
      @alokdixit5835 2 ปีที่แล้ว

      Yes bro

    • @sanjaisrao484
      @sanjaisrao484 11 หลายเดือนก่อน

      Yes

  • @MohadisTanoli
    @MohadisTanoli 8 หลายเดือนก่อน +6

    looks like 48 days of information in 43 minutes salute to you man

  • @SuperRia33
    @SuperRia33 23 วันที่ผ่านมา +1

    So many blogs and other videos are contradictory and confusing...until Nitesh sir explained it visually with simplification...thanks a ton😊

  • @anirudhthatipelli8765
    @anirudhthatipelli8765 ปีที่แล้ว +12

    Thanks a ton! This is a criminally underrated channel!

  • @cadc-pn1ir
    @cadc-pn1ir หลายเดือนก่อน +1

    O my God! Sir salute to you 🔥 you're like leaving no questions in my mind, your videos are in so much detail .... Hat's off to you sir

  • @shamshersingh9680
    @shamshersingh9680 11 หลายเดือนก่อน +5

    By far the best explanation of BN I have ever seen. Thanks a lot for your such a tremendous effort. Now I will surely go through all your videos about NNs.

  • @DataScienceSchools
    @DataScienceSchools 2 ปีที่แล้ว +4

    Just can't believe it , that someone can explain this complex topic in such a smooth way.

  • @gopikishan1028
    @gopikishan1028 2 ปีที่แล้ว +8

    I am learning data science since a year now and your channel is the last place where i get my every query , errors or concepts are cleared . your channel is institute itself it has every thing to become a data scientist. you are doing a great job and plz keep it up Sir . Thank you very much for putting these much efforts to teach in-depth . and one request plz make video on time series analysis 🙏

  • @bbagchi3771
    @bbagchi3771 4 วันที่ผ่านมา

    Amazing clarity and explanation .. by far the best channel on the subject. Take a bow Nitish !! thanks a ton.

  • @shamshadhussainsaifi6476
    @shamshadhussainsaifi6476 8 วันที่ผ่านมา

    Amazing explanation ,No where I found such amazing explanation. Thank you very much

  • @shivamchaurasia8140
    @shivamchaurasia8140 หลายเดือนก่อน +1

    Dhanya ho guru , bahut clearly samjha diya

  • @illumi.........7864
    @illumi.........7864 8 หลายเดือนก่อน

    Thank you so much sir, You are a true gem! I'll never forget your contribution in my journey of ML/DL.

  • @nomannosher8928
    @nomannosher8928 หลายเดือนก่อน

    Most underrated channel.

  • @motivation_with_harsh
    @motivation_with_harsh 10 หลายเดือนก่อน

    sir you are best teacher ever in this world of data science

  • @Shweta-ll3mk
    @Shweta-ll3mk หลายเดือนก่อน

    You are amazing sir.. 😊 thank you for giving such amazing and hard content in a very precise manner. You have ma saraswati s blessing for grasping and spreading knowledge🙏 .

  • @sidindian1982
    @sidindian1982 2 ปีที่แล้ว +1

    code is simple thx to Keras lib ,,but the batch normalization concept my God - So deeply involved & interesting to ... sir 🙂

  • @HimanshuSharma-we5li
    @HimanshuSharma-we5li 2 ปีที่แล้ว +1

    I'm so happy he is gaining momentum in subscribers 👌👌👌

  • @csescholar8890
    @csescholar8890 2 ปีที่แล้ว +1

    it was amazing video .............everything explained so nicely .....great ....thank you

  • @sidshukla2286
    @sidshukla2286 2 ปีที่แล้ว +3

    sir please optimizers videos its very hard to understand them and you're doing great work so thanks

  • @akshatsamdani
    @akshatsamdani ปีที่แล้ว

    Your effort really shows up in your teaching and video. Keep it up. You have gained a new subscriber.⭐

  • @AbhijeetKumar-cj4fe
    @AbhijeetKumar-cj4fe 9 หลายเดือนก่อน

    you are doing best job sir , at least for me.

  • @avishinde2929
    @avishinde2929 2 ปีที่แล้ว +1

    your teaching way is very good sir ji thank you so much
    please upload lecture on CNN as soon as possible sir ji ,please please

  • @chambyalatul
    @chambyalatul 11 หลายเดือนก่อน

    Thanks sir. Very informative and detailed discussion was there.

  • @pavantripathi1890
    @pavantripathi1890 5 หลายเดือนก่อน

    Thanks a lot for the wonderful explanation!

  • @afifaniks
    @afifaniks ปีที่แล้ว

    All I can say is, I wholeheartedly appreciate your effort! Just one thing though, if your videos were in English, people accross the globe could be benifitted. Thank you!

  • @MuhammadSaadullah-t3f
    @MuhammadSaadullah-t3f 3 หลายเดือนก่อน

    Awesome, yaar!
    You are very good at making anyone understand anything. But bad luck 💔, you are underrated.
    Keep up your work, and Allah will help you.

  • @saptarshisanyal6738
    @saptarshisanyal6738 3 หลายเดือนก่อน

    This is amazing stuff, but the challenge is to keep all the concepts in the brain. Another point, you are telling to normalizing the activation Z11, but Z11 is output for a particular node, so I think you meant to say firstly to normalize the output, then secondly, multiply with gamma and addd beta and then thirdly, pass this resultant through the activation function. got little confused here.

  • @rafibasha4145
    @rafibasha4145 2 ปีที่แล้ว +7

    Hi Bro,please finish interview ,NLP playlist as well topic modelling NER CRF encoders transformers chatbots etc ..also start deployment mlops

  • @shubhamhundet8351
    @shubhamhundet8351 2 ปีที่แล้ว +1

    Great sir keep uploading 🙏🙏🙏 after finishing DL plz complete NLP

  • @Ankit-hs9nb
    @Ankit-hs9nb ปีที่แล้ว

    This channel is so under rated.
    Sir if you want to start your paid courses then I will be the first one to buy it

  • @mimansamaheshwari4664
    @mimansamaheshwari4664 ปีที่แล้ว

    I love you man, what a lecture!!

  • @skaffan4667
    @skaffan4667 ปีที่แล้ว

    great explanation of a critical topic.

  • @dragnar4743
    @dragnar4743 ปีที่แล้ว

    Great video. Not only u explained batch normalization with implementation, but also cleared y nomalization is useful in general.

  • @ali75988
    @ali75988 ปีที่แล้ว +3

    Its a great lecture, probably the best one. just some minor things which bothered me in maths, so writing it down for any other person who is watching.
    20:02, i think the bias term would be 2x1 matrix (2 rows and 1 columns), instead of 1,2 stated above?
    update: something wrong with the maths logic here, matrix multiplication is not commutative or wt.x not equal to x.wt. In video (4,2)*(2,2) = x.wt
    another flaw, can't add two terms with different matrix dimensions. in video (4,2) + (1,2)
    My interpretation is, it takes single row and calculates a1, then another point and calculates a2, in the end we would have four numbers at node1 and then we can calculate meu, std at node1 of these four points and process repeats like a loop for given batch size (4 here).

  • @Btajicrew
    @Btajicrew 4 หลายเดือนก่อน

    amazing vid on BN sir

  • @ahsan7163
    @ahsan7163 7 หลายเดือนก่อน

    Great explanation

  • @rohitpotluri4378
    @rohitpotluri4378 6 หลายเดือนก่อน +1

    Goated Tutorial

  • @ParthivShah
    @ParthivShah 7 หลายเดือนก่อน +1

    Thank You Sir.

  • @afaqueumer7968
    @afaqueumer7968 ปีที่แล้ว +3

    Isn't the mean is 0 and SD 1 after Standardization? Normalization is when the values are made between 0 & 1? I am little confused with the terms

    • @pratikneupane951
      @pratikneupane951 4 หลายเดือนก่อน

      I am trying to make sense of the same.

  • @pratiksedhain5782
    @pratiksedhain5782 หลายเดือนก่อน

    great job bro

  • @ganeshy574
    @ganeshy574 4 หลายเดือนก่อน

    Thank you

  • @sajalhsn13
    @sajalhsn13 2 ปีที่แล้ว +2

    could you explain "Disharmony between Dropout and Batch Normalization" and suggest a good solution?

  • @mr.deep.
    @mr.deep. 2 ปีที่แล้ว +2

    Thanks sir

  • @ali75988
    @ali75988 ปีที่แล้ว

    34:05 had lot of problem in understanding how ewma is working, (came back after taking lecture on ewma and messing with chatgpt) this is what my tiny brain could grasp, correct me kindly, whereever wrong - as still confused:
    Q1: how ewma would be calculated after each epoch?
    Soln:
    alpha = 1 - beta
    EWMA_now = (1-alpha) EWMA_last_batch + alpha * MEAN_for_this_batch
    you get a new EWMA . after each epoch, you would use the same previous calculated ewma and move forward.
    For test,
    standardized_value = (new_data_point - mean or EWMA) / σ (if you have it) -------- formula copied from chatgpt
    Q2: I was confused kae std kon sa use karain gae testing main, as each node per std nikal rha.
    so here is chatgpt answer:
    during testing, use the per-node, per-epoch standard deviations calculated during training for the nodes in your network to ensure consistent and accurate normalization.
    Note: it also said, if you don't have the standard deviation explicitly calculated during training, you would use only the mean or EWMA values (if you have calculated them during training) for standardization during testing.

  • @MANALI-sn9li
    @MANALI-sn9li 10 หลายเดือนก่อน

    Hello sir, Firstly thankyou for teaching such great content highly appreciate your efforts. I had one doubt (34:05) why are we using the last batch values of mean and SD (EWMA) during testing?

  • @vaibhav1468
    @vaibhav1468 2 ปีที่แล้ว +3

    Sir, we request you to please complete the machine learning interview questions playlist as soon as possible.
    Campus placements are starting from 25 July.

  • @PavanUSAVlogs
    @PavanUSAVlogs 6 วันที่ผ่านมา

    Nitesh - Does the weight initialization concepts apply to Ganma and Beta also? And are there concepts like sgd to optimize Gamma and Beta values also ?

  • @tejasburadkar189
    @tejasburadkar189 2 ปีที่แล้ว +1

    Hi sir Since I following you on TH-cam from couple of year...there is a chance that you are going to make 100 days of playlist in Image processing video because lot of company are looking that the person should have expertise in Image processing.

  • @bibhutibaibhavbora8770
    @bibhutibaibhavbora8770 ปีที่แล้ว

    WOW What an explanation

  • @aritraray2501
    @aritraray2501 2 ปีที่แล้ว +2

    csv file is not available in the directory

  • @atharvajoshi4831
    @atharvajoshi4831 4 หลายเดือนก่อน

    I went through the paper for He initialisation. There they proved that once we initialize the weights according to the given formula, the input and output functions (Z) have the same variance. And this is preserved during the backpropagation step as well. So if we just normalise the inputs, then there would be no need to use batch normalisation right? Kaiming He initialisation should be sufficient.

  • @ShubhamSharma-gs9pt
    @ShubhamSharma-gs9pt 2 ปีที่แล้ว +1

    thanks for the great video:)

  • @asheeshmathur
    @asheeshmathur ปีที่แล้ว

    Good explanation. Keep it up

  • @Nil_money
    @Nil_money 7 หลายเดือนก่อน

    31:44 Why is there a need, if I have trained my network for batch normalization then I will use same weights and biase to test.

  • @RAHUDAS
    @RAHUDAS 2 ปีที่แล้ว

    Really Great

  • @satyamgupta4808
    @satyamgupta4808 ปีที่แล้ว

    your teaching is damn good sir

  • @aryanverma642
    @aryanverma642 หลายเดือนก่อน

    How come it's normalization? could anyone please clarify this for means 0 and standard deviations 1 means standardization.
    17:20,

  • @slingshot7602
    @slingshot7602 8 หลายเดือนก่อน

    In 5:13 can you please explain what are the contour plots of.

    • @MohadisTanoli
      @MohadisTanoli 8 หลายเดือนก่อน

      loss functions

  • @AISynthetic
    @AISynthetic 2 ปีที่แล้ว +1

    can you share the onennote you have used it would be very helpful

  • @nadmaanfazeel2691
    @nadmaanfazeel2691 2 ปีที่แล้ว +1

    Please finish NLP playlist

  • @DEEPAKYADAV-vb2ee
    @DEEPAKYADAV-vb2ee 2 ปีที่แล้ว +2

    if we have an activation function which will also provide output in normalize form then why do we need to add batch normalization on top of that?

    • @0Fallen0
      @0Fallen0 2 ปีที่แล้ว

      leakyrelu, elu, selu do not have fixedrange, i.e, it is unsaturated so this will help.

  • @jyothiyadav2595
    @jyothiyadav2595 ปีที่แล้ว

    Amazinggggg

  • @sandipansarkar9211
    @sandipansarkar9211 2 ปีที่แล้ว

    finished watching

  • @vatsalshingala3225
    @vatsalshingala3225 ปีที่แล้ว

    ❤❤❤❤❤

  • @yashjain6372
    @yashjain6372 ปีที่แล้ว

    best

  • @Chaatpakodi
    @Chaatpakodi 2 ปีที่แล้ว +1

    sir ji thumbnail pe thoda dhyaan do view count mein kaafi difference hai if you are adding good thumbnail ,even in large video 🧐

  • @SumitKumar-h9d3e
    @SumitKumar-h9d3e 10 หลายเดือนก่อน

    sir please share concentriccircle dataset

  • @shashanknakka8926
    @shashanknakka8926 11 หลายเดือนก่อน

    ⁉⁉⁉⁉⁉⁉🤔🤔🤔🤔Sir I have a doubt in batch normalization , as the the trainable paramaters increases for the batch normalization i.e., gamma, and beta and also it calculates its mean and standard deviation for all values in the batch. So my question is:
    Doesn't it Increases it's Computation time during training?

  • @ajitkumarpatel2048
    @ajitkumarpatel2048 2 ปีที่แล้ว

    🙏

  • @ssalma9647
    @ssalma9647 2 ปีที่แล้ว

    SITA RAMAM
    1 second ago
    Hi Sir can you please clarify my doubt here.
    Dense Hidden layer parameters like : {512*500)+500
    I also got like BN_1 : 500*4 in this what is this 4 SIr. according to me its no.of learning parameters.
    Can you please confirm me

  • @santoshhonnungar5543
    @santoshhonnungar5543 2 ปีที่แล้ว +1

    How many days it will take still to complete deep learning

  • @ssalma9647
    @ssalma9647 2 ปีที่แล้ว

    Hi SIr can you please clarify my doubt here.

  • @DarkShadow00972
    @DarkShadow00972 10 หลายเดือนก่อน

    Pls share dl onenote

  • @AntonySamuel
    @AntonySamuel 2 ปีที่แล้ว

    This is basically standardization right ? then why it is called batch normalization..?

  • @adityamishra6954
    @adityamishra6954 2 ปีที่แล้ว +3

    Sir deep learning ki videos boht late ho jati hai

    • @mr.deep.
      @mr.deep. 2 ปีที่แล้ว +1

      Ha bhai sir ko mai ne live me pucha tha sir ne bola ke research lag ta hai is liye late hoti hai video

    • @adityamishra6954
      @adityamishra6954 2 ปีที่แล้ว +2

      Then that's even more Great

    • @campusx-official
      @campusx-official  2 ปีที่แล้ว +13

      Bhai log sach me bahot mehnat lag rahi. Mazak nai kar raha 😪

    • @adityamishra6954
      @adityamishra6954 2 ปีที่แล้ว +2

      @@campusx-official thank you so much sir for your efforts you are giving in it for us
      That means a lot
      And sry if that's offending

    • @mr.deep.
      @mr.deep. 2 ปีที่แล้ว +2

      @@campusx-official ha sir mai khali bata raha tha
      Aditya Mishra ji ko sorry agar galat laga

  • @Ankit-hs9nb
    @Ankit-hs9nb ปีที่แล้ว

    3:55 mean =0 means normalization or standardization?
    Thanks in advance

  • @KumR
    @KumR 7 หลายเดือนก่อน

    22

  • @HimanshuSharma-we5li
    @HimanshuSharma-we5li 2 ปีที่แล้ว

    💎💎💎💎💎

  • @rb4754
    @rb4754 5 หลายเดือนก่อน

    You are damn good...