Principal component analysis step by step | PCA explained step by step | PCA in statistics

แชร์
ฝัง
  • เผยแพร่เมื่อ 3 ม.ค. 2025

ความคิดเห็น •

  • @UnfoldDataScience
    @UnfoldDataScience  2 ปีที่แล้ว

    Access English, Hindi Course here - www.unfolddatascience.com/store
    Don't forget to register on the website, it's free🙂

  • @tanmaychakraborty7818
    @tanmaychakraborty7818 2 ปีที่แล้ว +4

    Underrated channel for machine learning god bless you Aman

  • @victorbelmarlandaeta4763
    @victorbelmarlandaeta4763 7 หลายเดือนก่อน +4

    Dude, you are a really good teacher, awesome methodology!!!

  • @MirnaliniGunaraj
    @MirnaliniGunaraj 6 หลายเดือนก่อน +1

    One of the best videos that I have come across for PCA . Thanks @UnfoldDataScience

  • @souravbiswas6892
    @souravbiswas6892 2 ปีที่แล้ว +5

    Weight is vector, mass is scaler. However explained in detail. Great work.

  • @milliesadie486
    @milliesadie486 2 ปีที่แล้ว

    sir thank yoU to clear this concept coz i have been in youtube since 2 hour understanding pca and after watching this video i am clear my doubt

  • @giniyag8606
    @giniyag8606 2 ปีที่แล้ว

    Thunbs up with 2 hands . Was never able to understand this concepot before.Big Thank you :)

  • @krishnab6444
    @krishnab6444 2 ปีที่แล้ว +1

    perfectly explained aman thank you!

  • @abdullajubair9326
    @abdullajubair9326 25 วันที่ผ่านมา

    really good explanation

  • @URCristianoCR7z
    @URCristianoCR7z ปีที่แล้ว

    Awesome lecture.
    Better than so called professors

  • @ashwanibalyan9047
    @ashwanibalyan9047 ปีที่แล้ว

    Best video on PCA....keep it up

  • @shaileshpokharel586
    @shaileshpokharel586 ปีที่แล้ว

    Absolutely underrated tutor.

  • @JosyEmma
    @JosyEmma ปีที่แล้ว

    That's a detailed course thanks.

  • @rinkygupta8267
    @rinkygupta8267 9 หลายเดือนก่อน

    Today i become a big fan of your lectures... Hi i am following your lectures since last few months and i like them from the beginning, the way how you explain in very simple manner, the technique how to relate all the theory with real world examples, etc... You really doing a fantastic job... You really know how to explain maths in a very common language so it doesn't only fit in my mind but also touches my heart... Thank you for sharing your knowledge with us... I really want to learn more and more with you in near future... You seriously provide a training to the current teachers how to teach the things and how to generate the intrest of learners in any topic... 🥰 Best wishes

  • @malavikadutta1011
    @malavikadutta1011 2 ปีที่แล้ว

    Thanks Aman for such an awesome explanation for a confusing topic like PCA.

  • @rinkygupta8267
    @rinkygupta8267 9 หลายเดือนก่อน

    The way you explained the vectors mathematically correlated with flight example was wonderful.... 🥰 🎉

    • @UnfoldDataScience
      @UnfoldDataScience  9 หลายเดือนก่อน

      Your comment mean a lot to me. Welcome onboard to UFDS

  • @sajalhsn13
    @sajalhsn13 2 ปีที่แล้ว

    Unbelievable explanation. Wow!!!!

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว +1

      Thanks alot for your positive feedback. Please share with others as well so that everyone gets the knowledge.

  • @Krishna-pm8ty
    @Krishna-pm8ty 2 ปีที่แล้ว

    Wow . That was just superb.👏👏👏👏

  • @kunalthakre7164
    @kunalthakre7164 2 ปีที่แล้ว

    Thanks aman...
    It was really a helpful video.

  • @agirmaus-lh9zi
    @agirmaus-lh9zi 2 ปีที่แล้ว

    Thanks Aman for this wonderful explanation

  • @amarnathdhinakaran9522
    @amarnathdhinakaran9522 2 ปีที่แล้ว +1

    Thanks for the amazing content Aman.

  • @naageshk1256
    @naageshk1256 7 หลายเดือนก่อน

    Great explanation.. thank you so much 🎉❤

  • @saqibjawed3001
    @saqibjawed3001 ปีที่แล้ว

    good job simple and clear understanding

  • @tradetalks101
    @tradetalks101 2 ปีที่แล้ว

    Thanks boss ... Really appreciated .. Good work

  • @Birdsneverfly
    @Birdsneverfly 2 ปีที่แล้ว

    You have an outstanding explanation for PCA. All the technical jargon out there is only to confuse people. Cheers.

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว +1

      Thanks for watching, your comments mean a lot.

    • @Birdsneverfly
      @Birdsneverfly 2 ปีที่แล้ว

      @@UnfoldDataScience Thank you actually for sharing your knowledge. I am a data scientist myself, I regularly search TH-cam for quality education. Kudos for the work 😌

  • @RamanKumar-ss2ro
    @RamanKumar-ss2ro 2 ปีที่แล้ว +1

    Thanks for the video, it's too good.

  • @GauravSingh-zt4us
    @GauravSingh-zt4us หลายเดือนก่อน

    love u sir for this support

  • @radhakrishnanananthan1585
    @radhakrishnanananthan1585 2 ปีที่แล้ว +1

    Great explanation 👌

  • @rohitgaikwad2266
    @rohitgaikwad2266 2 ปีที่แล้ว

    Thanks Aman Nicely Explained 🙂👍

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว

      Welcome Rohit. pls share with friends who may be interested.

  • @Anamika-nb3tv
    @Anamika-nb3tv 5 หลายเดือนก่อน

    Very nice explanation😊

  • @sambitmohanty1758
    @sambitmohanty1758 2 ปีที่แล้ว +1

    Great video Aman as usual expected.

  • @rahulmedcure
    @rahulmedcure 2 ปีที่แล้ว +2

    of course, it was a great effort to explain PCA in a simple way. I would say at the end of the tutorial you should show the two-way plot explaining the information we are getting from the PCA which was difficult to predict while just looking at the data. Just a suggestion.

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว

      Appreciate your suggestion Rahul. Thanks for watching

  • @kidya-moohustories4764
    @kidya-moohustories4764 2 ปีที่แล้ว

    very clear and valuable

  • @rev1nth64
    @rev1nth64 6 หลายเดือนก่อน

    than you so much,love you bro

  • @achumohan5908
    @achumohan5908 2 ปีที่แล้ว

    Thanks a lot Aman!! well explained 🙂

  • @upendram2820
    @upendram2820 2 ปีที่แล้ว

    Very well explained... Thank you very much...

  • @akashprabhakar6353
    @akashprabhakar6353 2 ปีที่แล้ว

    I first hit like on your videos and then watch...coz i know you are always awesome 🙂

  • @kalpanapatil1028
    @kalpanapatil1028 2 ปีที่แล้ว

    Thanks Aman👍🙏

  • @ManojMaddineniBCS
    @ManojMaddineniBCS 2 ปีที่แล้ว

    Thank you so much for the detailed explanation. Really loved the way you covered each individual basic topic building up to the main topic.

  • @priyankathakur1691
    @priyankathakur1691 ปีที่แล้ว

    Great video. Request you to make more videos from basics for the entire data science project lifecycle.

  • @renvigautam6458
    @renvigautam6458 7 หลายเดือนก่อน

    Thank you so much sir ....🎉

  • @shaelanderchauhan1963
    @shaelanderchauhan1963 2 ปีที่แล้ว

    Great Videos Aman

  • @salahmahmoud2119
    @salahmahmoud2119 ปีที่แล้ว

    Your explanation is incredible!!!! 👏

  • @anuragrai7662
    @anuragrai7662 2 ปีที่แล้ว

    great video sir your explanation is amazing🔥

  • @BeaverMonkey
    @BeaverMonkey 2 ปีที่แล้ว

    You do a fantastic job explaining complex topics. Definitely subbing

  • @preranatiwary7690
    @preranatiwary7690 2 ปีที่แล้ว

    Amazing video, thanks for sharing 🙂

  • @akhildevjr
    @akhildevjr ปีที่แล้ว

    You did it better, and few of them are need more clarification especially for bigger datas having 100 or more columns, and how we can rotate the axis by which terms

  • @VKRealsta
    @VKRealsta 2 ปีที่แล้ว +1

    Nice session 👌

  • @faridhusen6651
    @faridhusen6651 2 ปีที่แล้ว

    I was hanging around until I find this video. Thank you sir!

  • @jaswanthgannamaneni8461
    @jaswanthgannamaneni8461 2 ปีที่แล้ว

    Great video sir

  • @adithyaboyapati
    @adithyaboyapati 2 ปีที่แล้ว +1

    Very Nice Explanation. You will never disappoint us 😄

  • @arni5na
    @arni5na 2 ปีที่แล้ว +9

    Weight is not scalar; it's mass pointing towards the direction of gravity. Mass is scalar.

  • @shekharkumar1902
    @shekharkumar1902 2 ปีที่แล้ว +1

    Jai ho Gurudev ! Sakshat Saraswati ka vaas hi apke kanth me ! Very well explained....one questions. How it gets decided that how much data is explained by PCA1 and how much data has explained by PCA2 and so on ?

    • @xendu-d9v
      @xendu-d9v 2 ปีที่แล้ว

      check EVR (explained variance ratio)

  • @sohaibyousuf
    @sohaibyousuf ปีที่แล้ว +1

    You have explained well but beginners are not able to undersatand the coding phase

  • @VKRealsta
    @VKRealsta 2 ปีที่แล้ว

    It's very understandable

  • @kishorem4406
    @kishorem4406 2 ปีที่แล้ว +1

    Very nicely explained 👌. Will be good if a Playlist is created for all ML algo explanations

  • @VKRealsta
    @VKRealsta 2 ปีที่แล้ว +2

    Please cover data mining, regression, correction, time series

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว

      Let me check on these topics, regression and time series playlist are there. You can check in playlist section.

  • @SujanShrestha-pv3tf
    @SujanShrestha-pv3tf ปีที่แล้ว

    I wnt want to know more about entity embedding for categorical variables as like this

  • @indiannationalist07
    @indiannationalist07 2 ปีที่แล้ว +2

    Waiting for videos on LDA , MDS ,t-SNE and PcoA

  • @nehalverma8063
    @nehalverma8063 2 ปีที่แล้ว

    Thanks a lot.

  • @sushantgunjan7597
    @sushantgunjan7597 2 ปีที่แล้ว +1

    Thanks Aman !
    Well explained as always. This was my demand few days back and you created this video for all of us once again thanks for this. I have one question if we convert the data to mean centric and taking the covariance matrix what is the intuitions behind this ? Somewhere I read that eigenvector are those vector whose direction does not change when we scaling the matrix so after getting the covariance matrix we are looking that covariance vector whose direction does not change after scaling the data and all those vector are principal component of that data.
    Please clarify my doubt and correct my understanding.

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว +1

      Thansk Sushant,
      Think like this.
      Make data mean centric ( so that covariance matrix is not screwed much even if data columns are on different scale consider milegae of car and it's cost in INR as two different coulmns)
      Calculate covariance matrix ( just to understand relationship between variables)
      Find Eigen value and eigen vector( to know on which direction maximum variance is there, may be 1,2,3 any number of directions, as I showed as V1, V2 in matrix example)
      Once we know in which direction/directions, maximum variance is there, we don't care about covariance matrix anymore, we just take our original data to that direction, we can say project original data to that direction to reduce dimension)

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว +1

      Also those vectors are not principal components, once u project your original data to that vector direction then u get principal components

    • @sushantgunjan7597
      @sushantgunjan7597 2 ปีที่แล้ว

      Thanks ! Got it.

  • @AnjanBasumatary
    @AnjanBasumatary 2 ปีที่แล้ว

    Sir please make video on exploratory data analysis

  • @mathematicalwisdom1226
    @mathematicalwisdom1226 2 ปีที่แล้ว

    While explaining Eigen value you expanded the matrix like determinant without telling that you are using determinant expansion as matrix can’t be expanded like this this-

  • @Jatindersingh-wo5hf
    @Jatindersingh-wo5hf ปีที่แล้ว

    Kindly come to basics like on which type of variables PCA is applied. Why not other methods. How to deal with variables having different scales. Everything should start from basics which I found every where missing

  • @AnilN-td6fy
    @AnilN-td6fy 2 ปีที่แล้ว

    Can you logic behind how to calculate Variance Explanantion by each PCA component? Keep up the good work. Thanks

  • @sharanm5718
    @sharanm5718 2 ปีที่แล้ว +1

    Hi aman,
    Can you please explain about quantization aware training, why it is used compared to floating point model

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว

      Thanks Sharan, I will try to bring video on it.

  • @jayasimhayenumaladoddi1602
    @jayasimhayenumaladoddi1602 2 ปีที่แล้ว

    Can you please make a video on OLPP

  • @AmitPatel-cl6ou
    @AmitPatel-cl6ou ปีที่แล้ว

    Pls use presentation mode in jupyter so I can view code fonts large in mobile, thnks

  • @nerdymath6
    @nerdymath6 2 ปีที่แล้ว

    Can u help with regularised k means clustering

  • @cagataydemirbas7259
    @cagataydemirbas7259 ปีที่แล้ว

    Great explanation thanks. Also I have a question; On my dataset 2 features has 0.8 corelation if I use PCA them to decrease one column is it handle 2 features without losing information ? Or should I just drop one column ?

  • @mustafachenine7942
    @mustafachenine7942 2 ปีที่แล้ว

    thank you

  • @hirdeshkumar4069
    @hirdeshkumar4069 2 ปีที่แล้ว

    Thank you sir. It is great video. Just one thing need to know, incase of PCA also, we need to do data cleaning or directly we can proceed for PCA??

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว

      Data cleaning will help PCA create more meaningful contents.

  • @indiannationalist07
    @indiannationalist07 2 ปีที่แล้ว +2

    What happen if we don don't pass n_components argument

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว +2

      Suppose if I say n=1, n=2, it means we want those many principal components. If you don't pass this argument at all in sklearn, all component are kept which will be equal to no of feature

    • @shantanuarya3214
      @shantanuarya3214 2 ปีที่แล้ว

      @@UnfoldDataScience How do you decide what should be the optimal number for "n" ?

  • @nareshkumarpatra6006
    @nareshkumarpatra6006 ปีที่แล้ว

    I have one question, PC1 shows more percentage, which means it should strongly correlate with the original output data. If possible, please clear this doubt.

  • @bijayalaxmikar6982
    @bijayalaxmikar6982 2 ปีที่แล้ว

    Hello Aman, Nice Explanation. but one question is it necessary all data set go through PCA or when we will use PCA

  • @VivadiMusic
    @VivadiMusic 2 ปีที่แล้ว

    How do we come up with the number for "n_components"?

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว +3

      For creating components, you can create all components, you can keep it default(equal to number of features) however, for choosing how many components for the next step, we see how many "minimum" Components can explain "maximum" Variance together.
      Let's say PC1 explains 80% variance
      PC2 explains 15% variance
      And rest All PC together explain remaining 5% of variance.
      In this case, we will choose only first two components, PC1 and PC2 for the next step.
      Just like we choose optimal number of K in K means cluster using elbow method.

    • @VivadiMusic
      @VivadiMusic 2 ปีที่แล้ว

      @@UnfoldDataScience am just reading your comment while watching your Regularisation video. Thank you sooo much. ♥️

  • @beifafana6905
    @beifafana6905 ปีที่แล้ว

    Can i use PCA to identify Climate smart Agriculture practices mainly used (adopted) by Household in the study area? pls help how can it is possible. Eg. i have 1.Conservation agriculture (Reduced tillage, Crop residue management-mulching, Crop-rotation/intercropping with cereals and legumes): 2.ISFM (Compost and manure management, Efficient fertilizer application techniques) 3...

  • @pratibhasingh1843
    @pratibhasingh1843 2 ปีที่แล้ว +1

    Sir pls make videos in hindi also

  • @ketakishitut2713
    @ketakishitut2713 ปีที่แล้ว

    Thank you, but why do we do mean centered

  • @cinimenosh5013
    @cinimenosh5013 7 หลายเดือนก่อน

    What the results tell or what it denote

  • @kshitijjain1385
    @kshitijjain1385 2 ปีที่แล้ว

    how are we calculating PC1 after projecting our data to new axis

  • @dhivyaakumar
    @dhivyaakumar 2 ปีที่แล้ว

    Sir, how do i label or annotate the data point after clustering. I have used covid 19 data set for pca analysis

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว

      Good question, take reference from original data rowwise.

  • @nerdymath6
    @nerdymath6 2 ปีที่แล้ว

    How to find the value of pc1 using python code

  • @narendrakumarpatel6910
    @narendrakumarpatel6910 2 ปีที่แล้ว

    Hi, I used XLSTAT and PAST tool to calculate PCs. I need "Contribution of the variables (%)" which I could get in XLSTAT easily but in PAST, I got value of "% variance". Is "% variance" in PAST is same as "Contribution of the variables (%)" in XLSTAT? Please respond. Thanks.

  • @VKRealsta
    @VKRealsta 2 ปีที่แล้ว

    I have one little doubt in python . If interviewer ask tell me about data types in python. Then what exactly we have to told . In our answers how I start . Can I start to data structure or start with saying numeric, logical, ....

    • @UnfoldDataScience
      @UnfoldDataScience  2 ปีที่แล้ว +1

      You can say simple data type like string, number, Boolean
      Then come list, array, dict, set
      Then comes some specific data structure like namedtuple etc. Read about collection module.

    • @VKRealsta
      @VKRealsta 2 ปีที่แล้ว +1

      @@UnfoldDataScience Thank you sir

  • @tamalikasikder5066
    @tamalikasikder5066 2 ปีที่แล้ว

    Can we do PCA on the combined results of samples from two separate distributions?

  • @SatishKumar-yn8tr
    @SatishKumar-yn8tr ปีที่แล้ว

    you took and mentioned 2 by 2 matrix. but data u took for python is 3 by 2 (three students and two subjects). This cretes confusion. A is not square now. And you first showed plot of original data. Plots after PCA not shown in video. Please show these for better understanding.

  • @pforpray41
    @pforpray41 2 ปีที่แล้ว

    Can you provide the source code..

  • @Samtoosoon
    @Samtoosoon 2 หลายเดือนก่อน

    14:37

  • @AshwiniTekude
    @AshwiniTekude ปีที่แล้ว

    Well Explained.......Thank You!

  • @RR-rg5lr
    @RR-rg5lr 21 วันที่ผ่านมา

    Thanks