Programming for AI (AI504, Fall 2020), Practice 14: Neural Ordinary Differential Equations

แชร์
ฝัง
  • เผยแพร่เมื่อ 14 พ.ย. 2024

ความคิดเห็น • 10

  • @sebastiangutierrez6424
    @sebastiangutierrez6424 ปีที่แล้ว +2

    Thank you so much for this tutorial!! It is so helpful. For anyone interested out there. I experimented changing the width and length of the NN architecture. I obtained the best results by increasing the width of the NN. Changing the length did not help to obtain better results.

  • @qianbozang-fb1bs
    @qianbozang-fb1bs ปีที่แล้ว

    Thanks for your video 🙏

  • @krishnabhogaonker3368
    @krishnabhogaonker3368 3 ปีที่แล้ว +3

    Wow, this is a really wonderful resource and tutorial. Thanks so much posting it. Is it possible to provide a link to the Colab notebook, so that others can work through the code? Thanks again.

  • @metaseq16
    @metaseq16 9 หลายเดือนก่อน

    is "t[:batch_time]" in get_batch correct? shouldn't we get the corresponding time points from `t` instead of a constant first `batch_time` elements from t?

  • @chadwinters4285
    @chadwinters4285 2 ปีที่แล้ว +2

    Why in the forward method of ODEFunc do you have self.net(y**3)? Why not just self.net(y) since y is the input to the network?

    • @themightyquinn100
      @themightyquinn100 ปีที่แล้ว

      Here is the reason. The NN is approximating the dynamic system which has an analytical form of y'=A*y^3. Where A in this eqaution is equal to the transpose of the A she is using. By scalar cubing the input y --> y^3 the NN's parameters are only learning the approximation for A. The reason I can only assume is to remove some of the non-linearity that the NN has to approximate. This seems a bit inpractical in a general sense because you don't always know the analytic form of the model. If we just had the data points from the solution of y'=A*y^3, we wouldn't have that intuition to cube the input before passing it to the NN. Hope this helps.

    • @sebastiangutierrez6424
      @sebastiangutierrez6424 ปีที่แล้ว

      @@themightyquinn100 I agree with that explanation. I tried removing the y^3 and the results were bad. I even considered a deeper and wider architecture for the network, but did not help.

  • @leninrafaelrierasegura644
    @leninrafaelrierasegura644 7 หลายเดือนก่อน +1

    Can we have access to the notebook, please?

    • @mp2893
      @mp2893  7 หลายเดือนก่อน +2

      You can find all slides and codes at mp2893.com/course.html

    • @leninrafaelrierasegura644
      @leninrafaelrierasegura644 7 หลายเดือนก่อน +2

      @@mp2893 Thanks a lot