6. Singular Value Decomposition (SVD)

แชร์
ฝัง
  • เผยแพร่เมื่อ 19 ธ.ค. 2024

ความคิดเห็น • 159

  • @tetrabromobisphenol
    @tetrabromobisphenol 3 ปีที่แล้ว +41

    Gilbert Strang is a treasure to the world. His books are top notch, and the MIT Open Courseware has made it possible for myself and so many others to learn at our own pace. Many thanks Dr. Strang!

    • @niamcd6604
      @niamcd6604 ปีที่แล้ว

      Brown nose. MIT also fails to hire anyone with brown skin and will blow out of proportion any work conducted by white old guys. Unbelievable.

  • @melisaalkan8832
    @melisaalkan8832 4 หลายเดือนก่อน +3

    "The key is that A transpose A is a great matrix" :D simply GOAT!

  • @gzitterspiller
    @gzitterspiller 4 ปีที่แล้ว +37

    I love the way he always explains everything from the beginning, he tells you the whole process.

  • @indrajitjana
    @indrajitjana 5 ปีที่แล้ว +81

    Thumbs up to the camera angle management team; great video audio editing.

  • @nishantraj95
    @nishantraj95 2 ปีที่แล้ว +14

    One of the best lectures on TH-cam, thanks Prof. Gilbert for explaining this beautifully.

  • @saitaro
    @saitaro 4 ปีที่แล้ว +37

    His explanations have always been impeccable. What a legend.

  • @dalisabe62
    @dalisabe62 3 ปีที่แล้ว +9

    Awesome! I could watch the lectures over and over and each time, I gain a different and extra insight into the results and the process. Linear Algebra could turn into magic after you apply it’s theory--the once dull computations of rows and columns. The most fascinating aspect of linear Algebra is when the entire matrix details unfold into a single effect, one that resembles the effect of simple scalars used in simple math formulas. Thanks to the fundamental concept of the Eigenvectors and Eigenvalues. Understanding the very essence of a matrix and what it does to any vector quantity (change in magnitude and direction) is the key concept to appreciating the underlying concept of various matrix decompositions. It all boils down to how many ways we could see the workings of a matrix in terms of its decomposed forms. The SVD composition is one of those decompositions of a matrix that allows a matrix to be seen as a sequential transformation of direction and magnitude in terms of some orthonormal basis characteristic of such matrix and its Eigenspace. Such a beauty!

  • @FranekMag
    @FranekMag 5 ปีที่แล้ว +136

    Fantastic lecture, really appreciate MIT sharing the course lectures for free.

  • @cpaniaguam
    @cpaniaguam 2 ปีที่แล้ว +2

    This SVD business is truly poetry in motion.

  • @akshayshrivastava97
    @akshayshrivastava97 3 ปีที่แล้ว +4

    Thank you very much Dr. Strang for making this intuitive and natural. I had only gone through SVD in bits and pieces and never had what I could consider a good enough understanding until I saw this lecture.

  • @mauriceli6949
    @mauriceli6949 5 ปีที่แล้ว +29

    love Prof Strang's lecture, enlighten me on the depth of linear algebra

  • @youngjim7987
    @youngjim7987 4 ปีที่แล้ว +10

    He just saved my from the brim of breaking down trying to understand SVD! What he presented is so natural, so beautiful that I instantly catch them. Appreciate MIT open courseware also!

  • @Hank-ry9bz
    @Hank-ry9bz 8 หลายเดือนก่อน

    16:23 symmetry; 28:50 geometry; 42:24 "econ" SVD

  • @megh280193
    @megh280193 3 ปีที่แล้ว

    Genius of this Man is the way he is explaining such a difficult concept so easily..... Hatts Off. As per Indian Ethos I just want to touch your feet.

  • @clecarosc
    @clecarosc 4 ปีที่แล้ว +12

    This is pure gold... such a fantastic Professor!! Thank you

  • @lfalfa8460
    @lfalfa8460 ปีที่แล้ว

    God bless your existence, Prof. Strang.

  • @wxie7689
    @wxie7689 2 ปีที่แล้ว

    Prof Strang is mathematical elegant in delivering linear algebra in an understandable way

  • @Enerdzizer
    @Enerdzizer 5 ปีที่แล้ว +13

    It is better not to rotate when getting ellipse. Ellipse has to be done without rotating, because multipling by lambda makes stretching along axes. Then of course should be rotating by matrix U. I.e. when ellipse first appeared on the board it should be vertically placed, not inclined.
    It is possible to draw - you should placed vectors 1 and 2 inclined initially on the first step. Then after multipling by V they will be along axes. Then stretching, then once again rotating. Still, it is a good lecture! :)

    • @chriselgoog9744
      @chriselgoog9744 4 ปีที่แล้ว +1

      You are right. For a picture see en.wikipedia.org/wiki/Singular_value_decomposition#/media/File:Singular-Value-Decomposition.svg

    • @coreybean8280
      @coreybean8280 4 ปีที่แล้ว

      His drawing isn't necessarily wrong, if for instance the vectors in V happen to be the unit vectors

  • @allyourcode
    @allyourcode 3 ปีที่แล้ว +3

    @6:49 The fact that A^T * A and A * A^T have the same non-zero eigenvalues is blowing my mind right now. I guess this is intimately related to the fact that det(B) = det(B^T), since det is where we get the characteristic polynomial/egenvalues...

    • @allyourcode
      @allyourcode 3 ปีที่แล้ว

      I think it turns out that this is not actually used, so no need to lose sleep over it, I guess...

    • @dariuszspiewak5624
      @dariuszspiewak5624 2 ปีที่แล้ว

      @@allyourcode You're mistaken. It is not only used but it's crucial for this to work. The fact that the eigenvalues are the same is what enables to have the same Sigma matrix for both of the matrices, AA' and A'A. If that was not the case, well, the technique would lead us astray.

    • @freeeagle6074
      @freeeagle6074 2 ปีที่แล้ว

      In a previous lecture, professor mentioned that eigenvalues of AB and BA (similar matrixes) are the same. If we consider A' as B and A as A, then A'A and AA' should have the same eigenvalues.

  • @gyeongchankim5423
    @gyeongchankim5423 5 ปีที่แล้ว +8

    This lecture in specific was truly beautiful

  • @BigBen866
    @BigBen866 2 ปีที่แล้ว

    I am confident once I went through a Refresher up to this point that his explanation will be as clear as water😀 Because none of my advanced mathematics professors ever explained in such detail!! Awesome professor 👍🏾🙏

  • @vigneshbhaskaran6436
    @vigneshbhaskaran6436 3 ปีที่แล้ว +1

    This lecture singularly helped me through the concepts on use of eigen values for capturing variability

  • @xc2530
    @xc2530 ปีที่แล้ว

    50:00 the important/biggest/principal part of matrix

  • @georgesadler7830
    @georgesadler7830 3 ปีที่แล้ว

    Professor Strang thank you for breaking down and explaining Singular Value Decomposition in linear algebra.

  • @stephanienam2218
    @stephanienam2218 4 ปีที่แล้ว +3

    I cannot thank you enough for this clear, insightful lecture. It is very inspiring that you are calling the matrices "beautiful", and I can feel your love for them :) For me your lectures are beautiful.

  • @xc2530
    @xc2530 ปีที่แล้ว

    42:00 use SVD, the smallest SVD and square SVD

  • @diegooliveira7713
    @diegooliveira7713 2 ปีที่แล้ว +2

    I love prof. Strang and I'm very grateful for his lectures, but I have to admit I don't find this proof convincing. Correct me if I'm wrong, but the project he starts out with (@3:32) is to prove that every matrix A has this decomposition. But @13:26, he starts out by **assuming** that A has this decomposition, and then he deduces what U, Σ, and Vᵀ must be. But that doesn't show that every matrix A has this decomposition. Instead, it shows that *if A has this decomposition*, then *that's what U, Σ, and Vᵀ must be*. Even if we were to argue that we can obtain such U, Σ, and Vᵀ for any matrix A, prof. Strang doesn't show that by multiplying UΣVᵀ we will get back A for any A.
    I struggled with this topic for months until I finally could arrange everything into what I believe is a proof for why every matrix A can be decomposed this way.

    • @dariuszspiewak5624
      @dariuszspiewak5624 2 ปีที่แล้ว

      Can you please share your proof with us? That would be awesome. Thank you!

    • @diegooliveira7713
      @diegooliveira7713 2 ปีที่แล้ว

      ​@@dariuszspiewak5624 Hi. I'm going to paste what I've written in my notes. The non-mathematic notation is in Portuguese (my native language), but I think Google Translate will probably do a good job.
      In any case, I haven't been peer-reviewed, so I cannot guarantee the accuracy of this proof.
      [Teorema]
      Se A é mxn e Rank(A) = r -> A = UΣVᵀ, onde:
      Vᵀ nxn é ortogonal e tem em suas primeiras r linhas os autovetores unitários de AᵀA para os quais os autovalores são maiores que zero, e nas restantes n-r linhas os autovetores unitários de AᵀA para os quais os autovalores são exatamente zero. As primeiras r linhas de Vᵀ são uma base para C(Aᵀ), e as restantes n-r linhas de Vᵀ são uma base para N(A).
      Σ mxn é diagonal mxn (retangular se m≠n) e tem em suas primeiras r entradas diagonais a raiz quadrada dos autovalores de AᵀA maiores que zero (que são os mesmos autovalores maiores que zero de AAᵀ), ordenados de acordo com as linhas de Vᵀ, e todas as outras entradas são 0.
      U mxm é ortogonal e a i-ésima de suas primeiras r colunas são Avᵢ->/σᵢ, onde σᵢ = √λᵢ (λᵢ = autovalor de vᵢ->), sendo que essas r colunas formam uma base para C(A), e as restantes m-r colunas formam qualquer base ortonormal para N(Aᵀ). E também é o caso que nas primeiras r colunas de U estão os autovetores unitários de AAᵀ para os quais os autovalores são maiores que zero, e nas restantes m-r colunas de U estão os autovetores unitários de AAᵀ para os quais os autovalores são exatamente zero.
      [Prova]:
      Considere A mxn (pode ser m≠n), com Rank(A) = r (r pode ser qualquer valor).
      Recapitulação básica:
      Sabemos que o espaço colunar C(A) e o espaço nulo à esquerda N(Aᵀ) ∈ ℝᵐ, e o espaço de linha C(Aᵀ) e o espaço nulo N(A) ∈ ℝⁿ.
      Sabemos que a dimensão tanto de C(A) quanto de C(Aᵀ) é r.
      Chame {v₁->, v₂->, ..., vᵣ->} os autovetores ortonormais de AᵀA cujos autovalores são maiores que 0, e {vᵣ₊₁->, vᵣ₊₂->, ..., vₙ->} os autovetores ortonormais de AᵀA cujos autovalores são exatamente 0.
      Esses conjuntos sempre precisam existir, pois:
      AᵀA nxn é simétrica, e portanto tem decomposição espectral QoΛQo⁻¹, o que faz seus n autovetores serem ortogonais. Normalizando esses n autovetores, obtemos um conjunto ortonormal.
      AᵀA é positiva semidefinida, portanto seus autovalores são maiores ou iguais a zero.
      Como Rank(AᵀA) = Rank(A) = r, segue-se, pelo Teorema da Nulidade, que dim(N(AᵀA)) = n-r, e portanto, por Gram-Schmidt, pode-se achar uma base ortonormal de n-r vetores para N(AᵀA). E como sabemos, cada vetor linearmente independente em um espaço nulo de uma matriz é um autovetor da matriz com autovalor igual a zero. Então existem pelo menos n-r autovetores ortonormais em AᵀA com autovalor igual a zero. Na verdade, existem _exatamente_ n-r autovetores ortonormais em AᵀA com autovalor igual a zero, pois se houvesse mais um autovetor ortornormal em AᵀA com autovalor igual a zero, então a dim(N(AᵀA)) seria maior que n-r.
      Então o conjunto {vᵣ₊₁->, vᵣ₊₂->, ..., vₙ->} sempre pode ser obtido.
      O conjunto {v₁->, v₂->, ..., vᵣ->} é obtido simplesmente selecionando-se os autovetores ortogonais em AᵀA cujos autovalores são maiores que zero. Eles sempre precisam existir, pois todos os autovetores em AᵀA com autovalor igual a zero estão no outro conjunto, e os autovalores de AᵀA são ou zero ou um número maior que zero.
      Como vimos, {vᵣ₊₁->, vᵣ₊₂->, ..., vₙ->} é uma base para N(AᵀA). E como já provado, N(AᵀA) = N(A).
      Como já provado: Se V ∈ ℝⁿ e dim(V) = r -> [(V⟂ é um subespaço ortogonal a V) & (dim(V⟂) = n-r) V⟂ é o complemento ortogonal de V]
      O subespaço spannado por {v₁->, v₂->, ..., vᵣ->} ∈ ℝⁿ e tem dimensão r, e o subespaço spannado por {vᵣ₊₁->, vᵣ₊₂->, ..., vₙ->}, que é N(A), é ortogonal e ele e tem dimensão n-r.
      Portanto o subespaço spannado por {v₁->, v₂->, ..., vᵣ->} é o complemento ortogonal de N(A), que como sabemos é C(Aᵀ).
      Então veja que {v₁->, v₂->, ..., vᵣ->} é um conjunto ortonormal que é uma base para C(Aᵀ), enquanto {vᵣ₊₁->, vᵣ₊₂->, ..., vₙ->} é um conjunto ortonormal que é uma base para N(A).
      Chame {u₁->, u₂->, ..., uᵣ->} o conjunto de vetores tais que u₁-> = Av₁->/σ₁, u₂-> = Av₂->/σ₂, ..., uᵣ-> = Avᵣ->/σᵣ, onde σᵢ = √λᵢ
      Esses vetores sempre precisam existir, pois λ₁, λ₂, ..., λᵣ são maiores que zero.
      Vamos provar que {u₁->, u₂->, ..., uᵣ->} são ortogonais:
      Para i≠j:
      uᵢ->ᵀuⱼ-> = (Avᵢ->/σᵢ)ᵀ(Avⱼ->/σⱼ)
      uᵢ->ᵀuⱼ-> = (1/(σᵢσⱼ))vᵢ->ᵀAᵀAvⱼ->
      Como vⱼ-> é um autovetor de AᵀA, segue-se que AᵀAvⱼ-> = σⱼ²vⱼ->. Substituindo:
      uᵢ->ᵀuⱼ-> = (1/(σᵢσⱼ))vᵢ->ᵀσⱼ²vⱼ->
      uᵢ->ᵀuⱼ-> = (σⱼ²/(σᵢσⱼ))vᵢ->ᵀvⱼ->
      Como os vᵢ->'s são ortonormais, vᵢ->ᵀvⱼ-> = 0. Substituindo:
      uᵢ->ᵀuⱼ-> = (σⱼ²/(σᵢσⱼ))*0
      uᵢ->ᵀuⱼ-> = 0
      Vamos provar que {u₁->, u₂->, ..., uᵣ->} são unitários:
      ||uᵢ->|| = √(uᵢ->ᵀuᵢ->)
      ||uᵢ->|| = √((Avᵢ->/σᵢ)ᵀ(Avᵢ->/σᵢ))
      ||uᵢ->|| = √((1/(σᵢσᵢ))vᵢ->ᵀAᵀAvᵢ->)
      Como vᵢ-> é um autovetor de AᵀA, segue-se que AᵀAvᵢ-> = σᵢ²vᵢ->. Substituindo:
      ||uᵢ->|| = √((1/(σᵢσᵢ))vᵢ->ᵀσᵢ²vᵢ->)
      ||uᵢ->|| = √((σᵢ²/(σᵢσᵢ))vᵢ->ᵀvᵢ->)
      Como os vᵢ->'s são ortonormais, vᵢ->ᵀvᵢ-> = 1. Substituindo:
      ||uᵢ->|| = √((σᵢ²/(σᵢσᵢ))*1)
      Simplificando:
      ||uᵢ->|| = √1
      ||uᵢ->|| = 1
      Portanto o conjunto {u₁->, u₂->, ..., uᵣ->} é ortonormal.
      Porque eles foram definidos como uᵢ-> = (Avᵢ->/σᵢ), segue-se que todos eles pertencem a C(A), pois são simplesmente combinações lineares das colunas de A.
      Como Rank(A) = r, então esses r vetores {u₁->, u₂->, ..., uᵣ->}, que são linearmente independentes (pois são ortogonais entre si), precisam ser uma base para C(A).
      Chame {uᵣ₊₁->, uᵣ₊₂->, ..., uₘ->} uma base ortonormal para N(Aᵀ). Sempre é possível achar essa base ortonormal por Gram-Schmidt.
      Chame V = [v₁-> v₂-> ... vₙ->]. Veja que V é ortogonal.
      Chame U = [u₁-> u₂-> ... uₘ->]. Veja que U é ortogonal.
      Chame Σ = matriz mxn cujas primeiras r entradas diagonais serão {σ₁, σ₂, ..., σᵣ}, e todas as outras entradas (diagonais e não-diagonais) serão 0. Note que Σ é diagonal (retangular se m≠n).
      Vamos explicar por que a seguinte igualdade é verdadeira:
      AV = UΣ
      Primeiro, vamos analisar AV.
      Sabemos que a i-ésima coluna de AV é simplesmente Avᵢ->. Para as primeiras r colunas de AV, a i-ésima coluna será, como definimos acima, σᵢuᵢ->, pois uᵢ-> = Avᵢ->/σᵢ. Todas as restantes n-r colunas serão 0->, uma vez que Avᵣ₊₁-> = 0->, ..., Avₙ-> = 0->, pois esses vetores formam uma base para N(A).
      Logo, AV = [σ₁u₁-> σ₂u₂-> ... σᵣuᵣ-> 0-> 0-> ... 0->], onde a quantidade de 0->'s é n-r. Note que AV é mxn, pois A é mxn e V é nxn.
      Agora, vamos analisar UΣ.
      Esta também será uma matriz mxn, pois U é mxm e Σ é mxn.
      Porque Σ é diagonal com r entradas diagonais não-zero e n-r entradas diagonais zero, temos o seguinte:
      O primeiro vetor colunar de Σ é , o segundo vetor colunar é , o r-ésimo vetor colunar é (na r-ésima posição), e todos os outros n-r vetores colunares são 0->.
      Portanto segue-se que UΣ será [σ₁u₁-> σ₂u₂-> ... σᵣuᵣ-> 0-> 0-> ... 0->]
      Logo, AV = UΣ.
      Como V é ortogonal, seu inverso é Vᵀ. Pós-multiplicando ambos os lados por Vᵀ:
      A = UΣVᵀ
      Logo, toda matriz tem decomposição por valores singulares.
      Vamos ver o que é AAᵀ:
      AAᵀ = UΣVᵀVΣᵀUᵀ
      Porque V é ortogonal, VᵀV = I. Substituindo:
      AAᵀ = UΣΣᵀUᵀ
      Como já provado, porque Σ é uma matriz diagonal mxn, ΣΣᵀ precisa ser mxm diagonal, e como Σ estava estruturado de forma que suas primeiras r entradas diagonais eram maiores que zero e as restantes são exatamente zero, segue-se que as primeiras r entradas diagonais de ΣΣᵀ são maiores que zero e as restantes são exatamente zero. Chame D := ΣΣᵀ. Substituindo:
      AAᵀ = UDUᵀ
      Porque U é ortogonal, Uᵀ = U⁻¹. Substituindo:
      AAᵀ = UDU⁻¹
      Veja que o LDE é a diagonalização de AAᵀ.
      Como já provado, precisa ser o caso que os vetores em U são autovetores de AAᵀ. Mais especificamente, nas primeiras r colunas de U estão os autovetores unitários de AAᵀ para os quais os autovalores são maiores que zero, e nas restantes m-r colunas de U estão os autovetores unitários de AAᵀ para os quais os autovalores são exatamente zero.

  • @hyperduality2838
    @hyperduality2838 9 หลายเดือนก่อน

    Poles (eigenvalues) are dual to zeros -- optimized control theory.
    The standard basis is dual to the eigen vector basis -- symmetric matrices.
    Stretch is dual to squeeze -- eigenvalues, forces are dual, ellipsoids are dual to circles or hyperspheres.
    "Always two there are" -- Yoda.
    Syntax is dual to semantics -- languages or communication.
    If mathematics is a language then it is dual.
    The teacher (Gilbert Strang, Yoda) is dual to the pupil (You, Luke Skywalker) -- the Hegelian dialectic.

  • @causalinference4176
    @causalinference4176 4 ปีที่แล้ว +14

    Good lecture, but I think the geometric intepretation of SVD is its motivation. We should use SVD as a tool to understand general linear transformation

  • @chiahungmou8010
    @chiahungmou8010 2 ปีที่แล้ว

    Love the cool idea to light up the U and V!

  • @renanwillianprado9187
    @renanwillianprado9187 3 ปีที่แล้ว

    If A^T A V = V S is the spectral decomposition of A^T A with S diagonal and V orthogonal, we have that (A V)^T (A V)= S. Hence, the nonzero columns of AV, which is associated with the nonzero columns of A, is orthogonal with all other nonzero columns of AV. Hence, is possible to find an orthogonal matrix U such that A V = U sqrt(S) -- note that the diagonal of S give us the squared norm of each column of AV, which means this is completely fine. Now, this means A=U sqrt(S) V^T. Completing U, if necessary, we have the SVD decomposition. In practice, after the calculation of the spectral decomposition of A^T A, we would calculate the Kern of (AV)^T -- orthogonal to Im(AV) -- and find an orthogonal basis to it, and of course, all the singular values associated with this process would be zero.

    • @renanwillianprado9187
      @renanwillianprado9187 3 ปีที่แล้ว

      P.S.: In the picture at 32:35, the u_1 and u_2 should be a_1 and a_2, after applying the rotation U.

  • @Unstable_Diffusion89
    @Unstable_Diffusion89 3 ปีที่แล้ว +1

    @25:00 Professor Strang manipulates V_1^t * (A^T * A) *V_2 into V_1^t * (∂_2 ^2) *V_2......
    How does A^T * A = ∂_2 ^2 exactly?

    • @MaicaZhott
      @MaicaZhott 3 ปีที่แล้ว

      Those v's are just eigenvectors of A^TA. Then we have A^TA v = ∂^2 v.

    • @johnnash8515
      @johnnash8515 3 ปีที่แล้ว

      Just come from AX=§X.

    • @intuitivecontrolsystem4436
      @intuitivecontrolsystem4436 3 ปีที่แล้ว +2

      1. Start with Ax = ∂ x and keep the transpose x^T A^T= x^T ∂ ready
      2. Multiply A^T both sides thus, A^T Ax = A^T ∂ x
      3. Multiply x^T both sides thus, x^T A^T Ax = x^T A^T ∂ x
      4. Replace x^T A^T with x^T ∂
      5. Thus, x^T(A^T A)x =x^T (∂ ^2 ) x
      So you believe (∂ ^2 ) are the eigenvalues of A^T A, so are the sigma^2

  • @fierydino9402
    @fierydino9402 ปีที่แล้ว

    Thank you so much for this beautiful and brilliant lecture.

  • @richardcasey4439
    @richardcasey4439 4 ปีที่แล้ว +3

    Awesome simplicity. Wish I had him in college

  • @manoranjansahu7161
    @manoranjansahu7161 2 ปีที่แล้ว +1

    Beautiful exposition of SVD

  • @xuantong1634
    @xuantong1634 3 ปีที่แล้ว

    Thank you professor Strang, thank you MIT OCW

  • @janferkl7649
    @janferkl7649 4 ปีที่แล้ว +1

    32:12 I do not understand why are the resulting vectors called u1 and u2. I expect Ax1 and Ax2 to be there as a result of A multiplication. If U multiplication is rotation, than u1 and u2 are just rotated unit vectors (exactly as v's, different rotation)... Can somebody prove me wrong please?

    • @HKwak
      @HKwak 3 ปีที่แล้ว

      I think it is correct to write σ_i*u_i not just u_i. U is just the rotation, so σ_i should remain there.

  • @Bridgelessalex
    @Bridgelessalex 5 ปีที่แล้ว +12

    I thought A^{T} A is positive semidefinite isn't it?

    • @zkhandwala
      @zkhandwala 5 ปีที่แล้ว +4

      I was going to respond the same thing. I believe that's correct. The general argument, however, still holds.

    • @gzitterspiller
      @gzitterspiller 4 ปีที่แล้ว

      that is why he writes positive or equal to 0.

    • @xaviert7224
      @xaviert7224 3 ปีที่แล้ว

      Yes I think you are right.

  • @jhatzi99
    @jhatzi99 ปีที่แล้ว +1

    "Six angles in 4D rotations. We'll leave it there..."

  • @Random-sm5gi
    @Random-sm5gi 2 ปีที่แล้ว

    Splendidly explained. Thank you so much

  • @neoneo1503
    @neoneo1503 3 ปีที่แล้ว

    The proof for why v1,v2,... ,vr are orthogonal means u1,u2,... ,ur are orthogonal by A*v=sigma*u at 23:00.

  • @mariap.9768
    @mariap.9768 ปีที่แล้ว

    4:49 There is an error in the subtitles. Should say 'It's symmetric', not 'It's a metric'

  • @jxchtajxbt53
    @jxchtajxbt53 3 ปีที่แล้ว

    Truly awesome instructor. Fantastic!

  • @xc2530
    @xc2530 ปีที่แล้ว

    45:30 second approach to find SVD

    • @xc2530
      @xc2530 ปีที่แล้ว

      Polar decomosition

  • @abhinandanmohanty3833
    @abhinandanmohanty3833 หลายเดือนก่อน

    At 19:37 why does Prof Strang encourages to use only the first formula

  • @karimsaid2248
    @karimsaid2248 5 ปีที่แล้ว +4

    @37 Doesn't rotation in 3d space require two angles only, theta(zenith) and phi(azimuth)?

    • @DeepakSingh-xt5io
      @DeepakSingh-xt5io 5 ปีที่แล้ว +3

      for 3d rotation we need roll,pitch, and yaw; you can specify zenith with pitch, azimuth with yaw, so we need roll parameter too.

    • @gustavstreicher4867
      @gustavstreicher4867 4 ปีที่แล้ว +1

      You can think of it as 3 combination 2, which gives 3. Two directions combine to form a single rotation from the one vector to the other. So, the number of angles is D combination 2, where D is the dimension of your space. Four dimensions will thus require 6 angles, etc.

    • @codywan3816
      @codywan3816 4 ปีที่แล้ว +1

      I think 2 angles are enough to specify the direction. Another parameter to specify the magnitude. I'm not a native speaker of English and didn't understand what they meant by roll, pitch and yaw physics.stackexchange.com/questions/175354/why-are-three-parameters-required-to-express-rotation-in-3-dimension

  • @k10shetty
    @k10shetty 3 ปีที่แล้ว +1

    5:06 did he mean positive semi definite ?

  • @asikakhatun4885
    @asikakhatun4885 3 ปีที่แล้ว

    খুবই সুন্দর লাগলো...

  • @elvykamunyokomanunebo1441
    @elvykamunyokomanunebo1441 2 ปีที่แล้ว

    Has any body else spotted the error @25:21
    A^T A = sigma_2 ^2 /( sigma_1 * sigma_2 ) is that an error there?
    sigma_2 ^2 relates to the vector v_2
    while A^T A relates sigma^2 , a diagonal matrix
    i.e. A^T A = V sigma^2 V^T
    or am I wrong, if so would appreciate an explanation
    :)

    • @vivianhuang5647
      @vivianhuang5647 2 ปีที่แล้ว

      We know that A^T*A = V(E^T*E)V^T in which V2 is an eigenvector to A^T*A, and sigma2^2 is its eigenvalue (from E^T*E). Then (A^T*A)*V2= (sigma2^2)*V2.

    • @elvykamunyokomanunebo1441
      @elvykamunyokomanunebo1441 2 ปีที่แล้ว

      @@vivianhuang5647
      Thanks
      :)

  • @henryzhu7309
    @henryzhu7309 4 ปีที่แล้ว +6

    38:35 pilot in class

  • @hanshans387
    @hanshans387 7 หลายเดือนก่อน

    really useful lecture, thank you!

  • @RamakrishnanAngaraiGanesan
    @RamakrishnanAngaraiGanesan 3 ปีที่แล้ว +1

    How can any mXn matrix A be factorized into SQ, where both are square matrices (symmetric and orthogonal)? The dimensions are not matching. Or, is this only for a square matrix A?

  • @wasabibleach
    @wasabibleach 3 ปีที่แล้ว

    Actually better than his textbook :)

  • @k10shetty
    @k10shetty 3 ปีที่แล้ว

    26:06 What did the professor mean by-> orthogonal v's in the "row space" ?

    • @thebigmonstaandy6644
      @thebigmonstaandy6644 3 ปีที่แล้ว +1

      i think he sayed it,becaurce v's belong to V^t(V ist transposed) ,so v^t belongs to V and v^t are rows

  • @evanparshall1323
    @evanparshall1323 2 ปีที่แล้ว

    One thing that I find confusing is that because the sigma matrix is left multiplied by V transposed, it is not pure scaling. This is because the sigma values are multiplied by the rows of V transposed and not the columns. This will cause additional rotation and changes in dimensionality.

    • @tdchayes
      @tdchayes 2 ปีที่แล้ว

      You need to do the steps in sequence. Yes, if you consider the operations from left to right you get mixed scaling and rotation. But considered from right to left, it's a (pure) rotation, a (pure) scaling) and a final rotation.

  • @ParthPaTeL-wm3kt
    @ParthPaTeL-wm3kt 11 หลายเดือนก่อน

    Very fruitful content

  • @omerblk1773
    @omerblk1773 3 ปีที่แล้ว +1

    Amazing lesson ♥♥

  • @vigneshrb2529
    @vigneshrb2529 9 หลายเดือนก่อน

    Amazing explanation by Gilbert sir as always but I have a small doubt. We find the U vectors from the V vectors because of the reason that there might be a collection of Vectors in U to choose from which are orthogonal but what happens when even the V vectors are a collection of vectors from where we have to choose the orthogonal one's?

  • @thomassun3046
    @thomassun3046 6 หลายเดือนก่อน

    I have a question, could anyone answer it? many thanks! how could v1 v2 in row spaces?, u1,u2 in column space which is true obvious

  • @learningstatistics1290
    @learningstatistics1290 4 ปีที่แล้ว

    26:09 I doubt the statement that v must be in the row space.

    • @jayhene830
      @jayhene830 3 ปีที่แล้ว

      (A^T*A)v = cv (A^t)Av = cv --> v must be in the column space of A^T since it is a linear combination of the columns of A^T. Therefore it is in the rowspace of A

  • @suzum0978
    @suzum0978 3 หลายเดือนก่อน

    He's an elite prof

  • @JulieIsMe824
    @JulieIsMe824 4 ปีที่แล้ว

    Love Prof.Strang so much!!!really excellent lecture!!!!

  • @piotrekm7943
    @piotrekm7943 ปีที่แล้ว

    OK, but If AA^T = U \Lambda U^T and A^TA=V \Lambda V^T what is derivation of formulae AV=\Sigma U?

  • @flyflyflyfly7757
    @flyflyflyfly7757 ปีที่แล้ว

    How to prove A times A transpose has the same non-trivial eigenvalue as A transpose times A?

  • @MohakKant
    @MohakKant 3 ปีที่แล้ว

    One question I have here is with the rotation and stretching - if A is a rectangular matrix does it not project x into a lower or higher dimensional space?

    • @yuwuxiong1165
      @yuwuxiong1165 3 ปีที่แล้ว

      projections are square matrices at the first place I guess.

  • @matiassandacz9145
    @matiassandacz9145 5 ปีที่แล้ว +3

    I think he made an error. He said the determinant of an orthogonal matrix is 1, but actually it is 1 or -1. Then Det(A) = + or - the product of the singular values.

    • @papaonn
      @papaonn 5 ปีที่แล้ว

      actually it refers to the symmetry matrix, so he simply ignore the negative value because symmetry matrix has eigen of positive semi definiteness, and the orthogonal matrix here refers to the eigen matrix.

    • @imrematajz1624
      @imrematajz1624 4 ปีที่แล้ว

      Matías Sandacz Hi, I think he was only talking about (semi-) definite matrices in which case Prof Strang was right.

    • @lunablind1456
      @lunablind1456 3 ปีที่แล้ว

      ​@@imrematajz1624 I think @Matías Sandacz refers to the U and V matrices. Those are orthogonal which means the determinant is either 1 or -1. If it is 1 it would be a rotation. If it -1 it would be a reflection.
      At 29:46 he even says that it could be a reflection.
      Can someone please confirm this?

  • @StarryNightSky587
    @StarryNightSky587 11 หลายเดือนก่อน

    „6 angles in 4 dimensions, we leave it there“ 😂

  • @kevinshao9148
    @kevinshao9148 3 ปีที่แล้ว

    One question Prof. and guys, how comes sigma in the middle is diagonal (square)? shouldn't its dimension (m, n) ? Thanks!

    • @kevinshao9148
      @kevinshao9148 2 ปีที่แล้ว

      @What could go wrong? - Niagara edition Ah, that makes sense. Thanks for your explain!

  • @mohamedabdallah7354
    @mohamedabdallah7354 4 ปีที่แล้ว

    How could I plug the UtU after sigma although they don't satisfy multiplication conditions concerning the size of matrices?

    • @EternusVia
      @EternusVia 4 ปีที่แล้ว

      I think he is referring to a square matrix A in that case

  • @NolanZewariligon
    @NolanZewariligon 5 ปีที่แล้ว

    Where can we find the notes professor Strang mentions in the videos? ( 45:17 )

    • @mitocw
      @mitocw  5 ปีที่แล้ว +5

      The materials for the course are available on MIT OpenCourseWare at: ocw.mit.edu/18-065S18. Best wishes on your studies!

    • @mitocw
      @mitocw  5 ปีที่แล้ว +1

      Prof Strang's notes were turned into his most recent book -- visit the OCW course site for links and the Reading section for guidance on what chapters go with the lectures: ocw.mit.edu/18-065S18.

  • @jorgeblanco9797
    @jorgeblanco9797 5 ปีที่แล้ว +8

    I wish I had gotten into MIT.

    • @dectorey7233
      @dectorey7233 5 ปีที่แล้ว +2

      To be fair the stress MIT students go through their undergrad/grad careers is inconceivable. Though it's incredible how much you learn even in one lecture at this school, let alone an entire 4 years

    • @jorgeblanco9797
      @jorgeblanco9797 5 ปีที่แล้ว

      @@dectorey7233 lol I know, I have friends that are going there. The summer programs I did there were tough too.

  • @myanimalist
    @myanimalist 3 ปีที่แล้ว

    Professor Strang mentions that U and V form bases for the four fundamental subspaces of A, but it's not clear to me how C(A) = C(Ur) and C(A') = C (Vr'). I know that U and V were determined by the eigenvectors of AA' and A'A, respectively, but how are these related to the column and row spaces of A?

    • @yuwuxiong1165
      @yuwuxiong1165 3 ปีที่แล้ว

      i think the structure of Sigma tells that.

    • @dariuszspiewak5624
      @dariuszspiewak5624 2 ปีที่แล้ว

      Here's how I understand it. They are connected to the column and row spaces by the fact that they constitute bases for those spaces, which in turn means there are r of them (rank of the matrix) and they are orthogonal (within each basis, that is to say, because the row space and column space generally belong to two different spaces if the matrix is not square). They are not columns or rows from the matrix itself. They can be constructed from the matrix (especially the basis for the row space), though, by the Gram-Schmidt process. I still need to confirm how the ortho-basis for the column space comes into life... By the way, it just so happens (miraculously) that the U and V matrices that store orthonormal vectors contain the eigenvectors of A'A and AA'. It's just a strange and beautiful fact. Additionally, the sigma matrix is built from the square roots of the eigenvalues which all happen to be real non-negative for the symmetric matrices A'A and AA'. Obviously, since the entries in U and V are eigenvectors belonging to the row/column-spaces, they definitely are linear combinations of the rows/columns (and there are r of them). Is that clearer?

  • @gemacabero6482
    @gemacabero6482 4 ปีที่แล้ว

    When U and V transposed are the same, do these matrices represent a change of basis matrix? Because for this to happen, A has to be positive symmetric, so then the singular values of A would equal the eigenvalues of A right? So if A represents a transformation , then the diagonal matrix sigma could be understood as that same transformation matrix but in the basis of eigenvalues, so then U = V transposed, should be change of basis matrices ????
    But when are change of basis matrices equal to each other????
    Thank you!!

    • @dariuszspiewak5624
      @dariuszspiewak5624 2 ปีที่แล้ว

      When transposed they are not the same. They are orthonormal which means when multiplication happens - UU' or VV' - the outcome is an identity matrix in the correct space (m-by-m or n-by-n).

  • @nickfruneaux5232
    @nickfruneaux5232 ปีที่แล้ว

    Love this!!

  • @Tyokok
    @Tyokok 3 ปีที่แล้ว

    33:16, 40:50, I think it's incorrect to claim capital sigma is EigenValues. I think for square symmetric matrix it still has singular decomposition, U equals V, but capital sigma here is actually Eigenvalue square. Because SVD is based A'A and AA', not A itself. Can anyone help clarify? Thanks!

    • @Tyokok
      @Tyokok 2 ปีที่แล้ว

      @What could go wrong? - Niagara edition just diff opinions. I think fundamental concepts are important to clarify to whoever study it or how they study it, especially those ones easily get confused. I actually took 18.065 from Prof. Strang. Since you mentioned order, do you mind remind which lesson(episode) focuses on this specific topic? otherwise nvmd

    • @Tyokok
      @Tyokok 2 ปีที่แล้ว

      @What could go wrong? - Niagara edition Thank you! Let me revisit and pay more attention.

    • @Tyokok
      @Tyokok 2 ปีที่แล้ว

      @What could go wrong? - Niagara edition well noted! much appreciation!

  • @김경환-p2z3c
    @김경환-p2z3c 2 ปีที่แล้ว

    Why do U and V mean rotation? Why not reflection?

  • @ajiteshbhan
    @ajiteshbhan 4 ปีที่แล้ว

    I have a query guys at 36:00 professor speaks about the no. of components in U, Sigma, V just looking at A for 2x2, 3x3 and 4x4, how do we look at rectangular matrices for such cases? Any help would be appreciated.

  • @mrudhulraj2824
    @mrudhulraj2824 3 ปีที่แล้ว +1

    Feels like we can do math on our own when we intuite it better

  • @gustavstreicher4867
    @gustavstreicher4867 4 ปีที่แล้ว

    Great video thanks!

  • @mosesgu0415
    @mosesgu0415 ปีที่แล้ว

    Since we can get free lecture from MIT, is it really necessary to go to college? I bet the wherever it is, MIT has better in quality.

  • @armandfavrot3210
    @armandfavrot3210 3 ปีที่แล้ว

    Awesome, thanks!

  • @naterojas9272
    @naterojas9272 5 ปีที่แล้ว

    How wonderful!

  • @vivaviiv
    @vivaviiv 5 ปีที่แล้ว

    Amazing stuff

  • @SphereofTime
    @SphereofTime 5 หลายเดือนก่อน

    0:53

  • @zeroheisenburg3480
    @zeroheisenburg3480 2 ปีที่แล้ว

    Did he miss the explanation why u is eigenvector of A^t A after proving the u's are orthogonal?

    • @coderhd8554
      @coderhd8554 2 ปีที่แล้ว

      I dont think any proof is needed because he rearranged A=UEV^T to get the u's he tests for being orthogonal. That means any u he considers is by definition an eigenvalue of AA^t. But dont quote me on that

  • @StrangeS
    @StrangeS 3 ปีที่แล้ว

    Can anyone please tell me how always sigma1 > sigma2 > ... sigma_r ?

    • @sampadmohanty8573
      @sampadmohanty8573 3 ปีที่แล้ว

      that is by convention. That is what we as humans have decided to put the sigmas into the sigma matrix.

  • @ZhaoXiliang
    @ZhaoXiliang 3 ปีที่แล้ว

    A=SQ, where A must be square matrix?

  • @DrRandyDavila
    @DrRandyDavila 5 ปีที่แล้ว +2

    Amazing

  • @BleachWizz
    @BleachWizz 3 ปีที่แล้ว

    You should really rerecord the previous 5 classes with functions on the place of numbers and call it tensor calculus - introduction

  • @nicolaaseltrut8040
    @nicolaaseltrut8040 3 ปีที่แล้ว

    致敬!

  • @goutamchatterjee1196
    @goutamchatterjee1196 4 ปีที่แล้ว

    Can someone please explain why is AV1 =sigma1 U1

  • @abhishekshukla6466
    @abhishekshukla6466 2 ปีที่แล้ว

    why singular values are always positive?

  • @shahlaalikhani6281
    @shahlaalikhani6281 3 ปีที่แล้ว

    Thanks!

  • @cl3826
    @cl3826 4 ปีที่แล้ว

    Why is UtU = I?

    • @erikdzul
      @erikdzul 4 ปีที่แล้ว

      U is an orthonormal matrix, so UtU = UUt = I

  • @jimmygan801
    @jimmygan801 4 ปีที่แล้ว +1

    wait... this is the dude that wrote the textbook!!!???

  • @menolmunes5198
    @menolmunes5198 ปีที่แล้ว

    26:35😅

  • @peterpark2981
    @peterpark2981 9 หลายเดือนก่อน

    Respect

  • @rezab314
    @rezab314 4 ปีที่แล้ว

    26:35 ☺️

  • @SophySongTan
    @SophySongTan 2 ปีที่แล้ว

    |det|,absolute value

  • @brendanlydon5272
    @brendanlydon5272 3 ปีที่แล้ว

    I am not entirely sure if this is correct ( plz say if I am not), but I’m also pretty sure U, sigma, and v-transpose can also be called LDL-transpose or also LDU for L is lower D diagonal and U for upper