Mashaan Alshammari
Mashaan Alshammari
  • 57
  • 20 692
3D Gaussian Splatting
I've looked into Gaussian splatting computation graph presented in the paper "gsplat: An Open-Source Library for Gaussian Splatting". Towards the end of the video, I used camenduru's "gaussian splatting colab" to train Gaussian splatting. I've also used antimatter15's splat viewer to view the output of Gaussian splatting.
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_10_14_3DGS.ipynb
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
- website mashaan14.github.io/mashaan/
- github github.com/mashaan14
- X mashaan_14
- linkedin linkedin.com/in/mashaan
- sponsor me on GitHub Sponsors github.com/sponsors/mashaan14
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
📹 Video edit: DaVinci Resolve
🎧 Audio enhancement: Adobe Podcast
🖼️ Thumbnails: GIMP
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
Acknowledgment:
Thanks to Dr. Adel F. Ahmed for introducing me to this problem. Adel is a great graphics researcher, I enjoyed working with him on multiple projects including my masters thesis.
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
Chapters:
00:00 start
01:27 Gaussians in 1D, 2D, and 3D
02:38 input to Gaussian splatting
05:02 the loss function
05:54 optimizing the color and opacity
07:53 optimizing the mean
09:15 why we don't optimize the covariance matrix?
13:05 optimizing scaling and quaternions
17:59 gsplat code
20:18 training Gaussian splatting
21:40 viewing the output of Gaussian splatting
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
#GaussianSplatting #Gaussian #Splatting #NeRF #RadianceField #rendering #graphics #NeuralRadianceFields #ai #deeplearning #machinelearning #python #neuralnetworks #artificialintelligence #pytorch #computervision #tutorial
มุมมอง: 112

วีดีโอ

Swin Transformer Attention
มุมมอง 301หลายเดือนก่อน
In this tutorial, I plot the attention maps of a Swin transformer. Attention maps are great tools to look inside and invistigate the vision transformers. The Swin Transformer was presented in the paper titled 'Swin Transformer: Hierarchical Vision Transformer using Shifted Windows.' ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/...
Swin Transformer Code
มุมมอง 6582 หลายเดือนก่อน
In this tutorial, I skimmed through the paper ‘Swin Transformer: Hierarchical Vision Transformer using Shifted Windows.’ I also run the code on a Jupyter notebook with some printouts that show how the tensor changes inside the transformer. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 swin transformer annotated: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_08_19_swin_transformer_annotated.ipy...
Cluster GCN in JAX
มุมมอง 1353 หลายเดือนก่อน
In this JAX tutorial, I implemented a cluster GCN sampler similar to the one introduced in "Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks." I also trained a GCN network using the jraph library, both with full batches and mini-batches. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/ma...
GNN Node Sampler in JAX
มุมมอง 1354 หลายเดือนก่อน
In this JAX tutorial, I implemented a neighbor sampler similar to the one introduced in the GraphSAGE paper titled ‘Inductive Representation Learning on Large Graphs.’ I also trained a GCN network using the jraph library, both with full batches and mini-batches. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_0...
Mini Batch Sampling for Graph Neural Networks
มุมมอง 2555 หลายเดือนก่อน
I skimmed through the main papers in Graph Neural Network (GNN) sampling with a bit of history on graph sampling. At the end, I went through a visual guide on how to do GNN sampling in pytorch geometric. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_05_27_GNN_sampling.ipynb ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ - web...
Graph Attention Networks in DGL
มุมมอง 1955 หลายเดือนก่อน
I looked into the implementation of graph attention networks in DGL. DGL is a library for deep learning on graphs. I also talked about how data is processed inside graph attention layers. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_05_13_DGL_GAT.ipynb ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ - website mashaan14.github...
Cluster-GCN Mini Batches in PyTorch Geometric
มุมมอง 3416 หลายเดือนก่อน
In this pytorch tutorial, I looked into performing mini-batch training in graph convolutional networks (GCN). I also saw how to use ClusterData and ClusterLoader from pytorch geometric to do mini-batch training in graph neural networks (GNN). ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_04_29_GNN_mini_batch....
JAX Just In Time (JIT)
มุมมอง 1066 หลายเดือนก่อน
In this JAX tutorial, I looked into Just In Time (JIT) compilation. I also saw how JAX translates python code to jaxpr programming language for efficient execution on GPU and TPU. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_04_14_jax_speed_test.ipynb ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ - website mashaan14.github....
Convolutional Neural Net (CNN) in JAX
มุมมอง 1837 หลายเดือนก่อน
In this JAX tutorial, I looked into the architecture of convolutional neural net (CNN) in Flax library, which is a deep learning library released by google research. I also used a CNN to classify images in CIFAR10 dataset. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this video: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_04_02_jax_CNN.ipynb ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬...
Vision Transformer (ViT) in JAX
มุมมอง 2607 หลายเดือนก่อน
In this JAX tutorial, I looked into the architecture of vision transformers (ViT) in Scenic library, which was recently released by google research. ViT was introduced in the paper "An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale" by Dosovitskiy et al. I also used ViT to classify images in CIFAR10 dataset. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in th...
Graph Attention Networks in JAX
มุมมอง 1827 หลายเดือนก่อน
In this jax tutorial, I looked into the implementation of a graph attention layer in jax. A graph attention network was introduced by Velickovic et al. in their paper "Graph Attention Networks". In this video, the focus is on (1) how to package graph attention layers inside a flax module (2) solving the problem of masking inside jax.jit (3) training and testing graph attention on graph two data...
JAX Conv Layer
มุมมอง 1778 หลายเดือนก่อน
In this jax tutorial, I compared the convolutional output of a single layer using pytorch and jax. Along the way, I learned how to convert variables back and forth between these two libraries and what each library expects as an input. I also plot the output of the convolutional filters which looked the same for both libraries. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗 jupyter notebook I built in this vid...
Simple Neural Net in JAX
มุมมอง 7338 หลายเดือนก่อน
In this jax tutorial, I implemented a neural network with one hidden layer to classify a two-dimensional dataset. The network was coded using jax and flax libraries from google research. In this video, the focus is on (1) generating a dataset using Dataset class from pytorch utilities (2) building a three-layer neural network using flax (3) training the network using jax capabilities. ▬▬▬▬▬▬▬▬▬...
GCN code for PyTorch Geometric and Jraph
มุมมอง 3168 หลายเดือนก่อน
I coded a Graph Convolutional Network (GCN) using two GNN libraries: pytorch geometric and jraph. In this video, the focus is on (1) understanding how pytorch geometric and jraph package graph data (2) the different data types used in both libraries (3) training a GCN on both libraries and compare the results. ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬ 🔗GCN in JAX/Flax: github.com/mashaan14/TH-cam-channel/b...
Sparse Subspace Clustering (SSC)
มุมมอง 1998 หลายเดือนก่อน
Sparse Subspace Clustering (SSC)
Graph Attention in PyTorch Geometric
มุมมอง 7789 หลายเดือนก่อน
Graph Attention in PyTorch Geometric
GCN Variants: SGC and ASGC
มุมมอง 889 หลายเดือนก่อน
GCN Variants: SGC and ASGC
PyTorch Conv2d Explained
มุมมอง 5259 หลายเดือนก่อน
PyTorch Conv2d Explained
DETR Object Detection
มุมมอง 1979 หลายเดือนก่อน
DETR Object Detection
Spectral Clustering Code
มุมมอง 7219 หลายเดือนก่อน
Spectral Clustering Code
CNN vs ViT: PyTorch Training
มุมมอง 2659 หลายเดือนก่อน
CNN vs ViT: PyTorch Training
Simple Neural Net in PyTorch
มุมมอง 16310 หลายเดือนก่อน
Simple Neural Net in PyTorch
PyTorch code for GCN and SGC
มุมมอง 18910 หลายเดือนก่อน
PyTorch code for GCN and SGC
Perceptron Algorithm Code
มุมมอง 8410 หลายเดือนก่อน
Perceptron Algorithm Code
Graph Convolutional Networks (GCNs) in PyTorch
มุมมอง 1.8K11 หลายเดือนก่อน
Graph Convolutional Networks (GCNs) in PyTorch
Vision Transformers (ViT) pytorch code
มุมมอง 2.7K11 หลายเดือนก่อน
Vision Transformers (ViT) pytorch code
Violin plots using pandas and seaborn
มุมมอง 6111 หลายเดือนก่อน
Violin plots using pandas and seaborn
pytorch softmax function in manim
มุมมอง 177ปีที่แล้ว
pytorch softmax function in manim
Visualize numpy broadcasting using manim
มุมมอง 249ปีที่แล้ว
Visualize numpy broadcasting using manim

ความคิดเห็น

  • @kevinkatsuradanisitanggang2243
    @kevinkatsuradanisitanggang2243 22 วันที่ผ่านมา

    Hi Mashaan, should be noted that using torch.max with dim=-1 will give you the max of the row data since the end data represent the sequence of row data. It's quite different with the order when indexing in tensor e.g. Tensor([[0,1],[2,3]])[0,1] -> means row 0, col 1 that seems more intuitive. The same with torch.mean. * referring to the early explanation of plot attention function segment

    • @mashaan14
      @mashaan14 21 วันที่ผ่านมา

      you're right, in 30:30 I said "the maximum along the column dimension", actually it takes the maximum along the row dimension.

  • @honglu679
    @honglu679 หลายเดือนก่อน

    Man, you did a great job digging into the code details and also put in your own thoughts. I usually dont leave a comment, but your video is way way better than those ones that claims to teach something complicated in 10 or 15 mins with random visualization. One suggestion, maybe you could do a video on the code analysis of metaAI omnivore and omniMAE, they are extensions of swintransformer but support both video and image.

    • @mashaan14
      @mashaan14 หลายเดือนก่อน

      I'm so glad that you liked the video. Thanks for suggesting these two papers. I'll definitely look into those. The thing is I'm recording two videos on an entirely different topic. It might take me a while before getting back to vision transformers.

  • @inquisitiverakib5844
    @inquisitiverakib5844 หลายเดือนก่อน

    great! Can you make a separate video on depth-wise and point-wise convolution ?

    • @mashaan14
      @mashaan14 หลายเดือนก่อน

      Thanks for your suggestion. I'll add it to my to do list. The thing is I'm recording two videos on an entirely different topic.

  • @0兒-y4c
    @0兒-y4c หลายเดือนก่อน

    Hi sir Im a student who is studying on it i would like to use swin transformer on object detection from my project how can i accomplish thank you sir

    • @mashaan14
      @mashaan14 หลายเดือนก่อน

      Usually, an image classification model is used at the beginning of object detection pipeline, and it’s called backbone. Most object detection pipelines use ResNet as backbone. I assume that you want to replace ResNet with Swin, just like what they did in the paper (section 4.2). If that’s the case, your best option is to use MMDetection library. They already included Swin as a backbone on their github: github.com/open-mmlab/mmdetection/blob/cfd5d3a985b0249de009b67d04f37263e11cdf3d/mmdet/models/backbones/swin.py

  • @pradyumagarwal3978
    @pradyumagarwal3978 หลายเดือนก่อน

    you said the 200 epochs test u ran is not a proper wxperiment to judge the quality of this transformer architecture. So other than increasing the c vaue back to 96, what other things should I look into to experiment and get the best performance out of this architecture

    • @mashaan14
      @mashaan14 หลายเดือนก่อน

      The settings I used in the video were simple just to have a taste of this transformer. In my opinion, a proper experiment would be replicating the results in the paper on ImageNet-1K dataset (the ones in table 1). This way we can judge the model, and then look for improvement.

    • @pradyumagarwal3978
      @pradyumagarwal3978 หลายเดือนก่อน

      @@mashaan14 im sorry. which table? (Also, big thanks your videos and replies have been a big help. however, any chance I can ask somewhere more convenient than yt comments?)

    • @mashaan14
      @mashaan14 หลายเดือนก่อน

      Table 1 on page 6 of swin transformer paper. You can text me on twitter or linkedin, whichever is convenient to you. twitter.com/mashaan_14 linkedin.com/in/mashaan

    • @pradyumagarwal3978
      @pradyumagarwal3978 หลายเดือนก่อน

      @@mashaan14 Okay thanks

  • @pradyumagarwal3978
    @pradyumagarwal3978 หลายเดือนก่อน

    you had set the patch size to 4 and widow size to 6. c went from 96 to 48. what were default values for patch size and window size? also, this model does not have any positional encoding right? the regular sin-cos positional encoding or better, adding relative positional encoding is possible for improvement right?

    • @mashaan14
      @mashaan14 หลายเดือนก่อน

      The defaults were patch size=4 and window size=7. You can find all default settings in: github.com/microsoft/Swin-Transformer/blob/main/config.py You’re right, there are no positional encoding in this model. The model uses (Relative position bias) and they said, “We observe significant improvements over counterparts without this bias term or that use absolute position embedding”. They added the (Relative position bias) inside WindowAttention class: github.com/microsoft/Swin-Transformer/blob/f82860bfb5225915aca09c3227159ee9e1df874d/models/swin_transformer.py#L101

  • @pradyumagarwal3978
    @pradyumagarwal3978 2 หลายเดือนก่อน

    is the notebook where you test the model with C = 48 and 200 epochs available somewhere, I would really like to check it out

    • @mashaan14
      @mashaan14 2 หลายเดือนก่อน

      here you go: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_08_19_swin_transformer.ipynb

    • @pradyumagarwal3978
      @pradyumagarwal3978 หลายเดือนก่อน

      @@mashaan14 thankssss

  • @mashaan14
    @mashaan14 2 หลายเดือนก่อน

    Hi everyone 👋 It’s been a while since I post this video, and it’s time to reflect back. First, there are multiple ways to visualize attention in vision transformers. This paper (“Transformer Interpretability Beyond Attention Visualization”, arxiv.org/abs/2012.09838v1 ) compared different visualization methods. What I did in this video is just testing the attention at the first layer by feeding a test image and pulling out the response, which are the query, key, and value matrices. Multiplying the query matrix with the transposed key matrix gives us a squared matrix showing how each patch is “paying attention” to every other patch. If we order the patches back to their positions in the original image, we’ll see which patches have the highest attention values. I updated the code by adding more comments and printouts, just to make it more readable.

  • @tylervandermate
    @tylervandermate 2 หลายเดือนก่อน

    This is exactly what I've been trying to find for model visualization. Thank you! It's difficult finding any in-depth info on transformers involving the query, key, and value matrices.

    • @mashaan14
      @mashaan14 2 หลายเดือนก่อน

      Thank you, that's great to hear. I'm currently working on a Swin transformer visualization video. Hopefully, I'll post it in a week or so.

  • @petermerrill9803
    @petermerrill9803 2 หลายเดือนก่อน

    Excellent explanations. Thank you very much.

    • @mashaan14
      @mashaan14 2 หลายเดือนก่อน

      Glad it was helpful!

  • @akashprajapathi6056
    @akashprajapathi6056 3 หลายเดือนก่อน

    Of course

  • @yasir_rashid7780
    @yasir_rashid7780 4 หลายเดือนก่อน

    Great video sir with lot of information ... Sir can we use GCN for idenitfying influential nodes in social networks

    • @mashaan14
      @mashaan14 4 หลายเดือนก่อน

      Thanks for your question. It made me go and dig a little bit. I read a couple of papers and I found out that GCN can’t be used alone to find influential nodes. However, it can be used as an introductory step to find influential nodes. For example, in a paper named: “Finding Critical Users in Social Communities via Graph Convolutions” the authors used GCN before graph attention to learn “criticalness” over nodes. By criticalness, they mean influential nodes. Another paper called: “SocialGCN: An Efficient Graph Convolutional Network based Model for Social Recommendation” arxiv.org/pdf/1811.02815 They built a recommender system based on GCN embeddings not the feature matrix. They stated that “the proposed SocialGCN model is flexible when the user and item attributes are not available”, which makes sense! because if a node does not have features, GCN will assign embeddings similar to its neighbors.

  • @mashaan14
    @mashaan14 4 หลายเดือนก่อน

    Mistakes in the video (the ones that I know about 😅): - I changed the implementation for the neighbor_sampler function. It previously takes one minute to sample from Cora dataset. Now, it samples in around 10 seconds. That’s because I removed the unique function and used sparse matrices indexing. The link in the description has the new code. - In 32:54 I said that the probability of picking an edge is 1. That's actually not true, it's 1/|E|, where |E| is the number of edges in the graph. I fixed it on my github notes.

  • @Savi_Ann
    @Savi_Ann 4 หลายเดือนก่อน

    Nice visualization!

  • @oneplus383
    @oneplus383 4 หลายเดือนก่อน

    Short but informative

  • @AbdulQadeerRasooli-l8k
    @AbdulQadeerRasooli-l8k 4 หลายเดือนก่อน

    Hello sir if possible please make a video related to the practical work of this paper Paper Title: GCN-FFNN: A two-stream deep model for learning solution to partial differential equations

    • @mashaan14
      @mashaan14 4 หลายเดือนก่อน

      Thanks for your suggestion. I skimmed though the code on github. I think the main contribution is class Ensemble(), which you can find in (models.py). That class concatenates the outputs from GCN and FFNN. I’ll try to fit the paper in one of my upcoming videos.

  • @tomoki-v6o
    @tomoki-v6o 5 หลายเดือนก่อน

    great trick . i would use (X-XT)**2+(Y+YT)**2

    • @mashaan14
      @mashaan14 5 หลายเดือนก่อน

      That's the beauty of coding, it can be done in different ways..

  • @mashaan14
    @mashaan14 5 หลายเดือนก่อน

    Code walkthrough: th-cam.com/video/y1ZmMcMYjkY/w-d-xo.html You can access the notebook in github: github.com/mashaan14/VisionTransformer-MNIST/blob/main/VisionTransformer_MNIST.ipynb

  • @mashaan14
    @mashaan14 5 หลายเดือนก่อน

    After playing with jax I don’t feel comfortable linking the notebook I showed in the video. Most of the video content is still valid because it shows the difference between pyg and jraph. However, in the notebook I used haiku which is not recommended by Google DeepMind. They recommend using flax instead. So, I linked a new notebook showing GCN code in JAX/Flax: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_03_21_jraph_GCN.ipynb Here’s another video where I explained graph attention code in JAX/Flax: th-cam.com/video/O1zGWMEgW7A/w-d-xo.html

  • @mehmeterenbulut6076
    @mehmeterenbulut6076 5 หลายเดือนก่อน

    Hi man, beautiful video explaining both libraries! Loved your explanation; clear and on point. About the issue that the test results of PyG and Jraph being different, I think it is because even though both obtained 100% training accuracy (which also means they overfitted the data), the decision boundary they draw for the training set is not necessarily the same. One reason that might lead them to be different is that PyG's and Jraph's GNN weights are probably initialized randomly. Therefore, their different decision boundaries can easily result in 2 different results on the test set.

    • @mashaan14
      @mashaan14 5 หลายเดือนก่อน

      I loved your explanation, yeah it totally makes sense. But it was far simpler than that, I was training on two different feature matrices. If you notice in the jraph part, I passed this command: nodes=jnp.eye(data_Cora.x.shape[0]) I was training jraph on the identity matrix while training pyg on the feature matrix. I know it’s crazy how jraph got so close with only the identity matrix. Anyways, I couldn’t fix the notebook in the video because it was written in haiku. So I took it down and write a new one with JAX/Flax: github.com/mashaan14/TH-cam-channel/blob/main/notebooks/2024_03_21_jraph_GCN.ipynb I’d love if you can take a look at the new code.

  • @mashaan14
    @mashaan14 5 หลายเดือนก่อน

    There was a mistake in the video. I accidentally used the identity matrix instead of the feature matrix when I packaged the graph in a jraph.GraphsTuple. This line of code: nodes=jnp.eye(data_Cora.x.shape[0]), should be changed to: nodes=jnp.asarray(data_Cora.x), I fixed it on github and the notebook should work fine.

  • @oneplus383
    @oneplus383 5 หลายเดือนก่อน

    Very good explanation. 😊

    • @mashaan14
      @mashaan14 5 หลายเดือนก่อน

      Glad it was helpful!

  • @doublesami
    @doublesami 6 หลายเดือนก่อน

    very informative, Can you please make a video on vision mamba or Vmamba and explain the theoretical as well as the implementation part ? Looking forward

    • @mashaan14
      @mashaan14 6 หลายเดือนก่อน

      Thanks, I just checked VMamba on github. Sure I'll add it to my todo list. The thing is I'm recording a series on graph neural networks. Once I'm done with that, I'll get back to vision transformers.

  • @oneplus383
    @oneplus383 6 หลายเดือนก่อน

    How can we convert pytoch model into jax?

    • @mashaan14
      @mashaan14 6 หลายเดือนก่อน

      actually I'm not aware of any tool that can convert pytorch to jax.

    • @oneplus383
      @oneplus383 6 หลายเดือนก่อน

      @@mashaan14 can you give me some idea about what can I change to run PyTorch to jax

    • @mashaan14
      @mashaan14 6 หลายเดือนก่อน

      @@oneplus383 Instead of torch tensors you have to use jax.numpy arrays. For each pytorch layer, check its equivalent in jax documentation. Start with something small and build your way up. Here's how to code a simple NN in jax th-cam.com/video/GNLOa4riys8/w-d-xo.html

    • @oneplus383
      @oneplus383 6 หลายเดือนก่อน

      Thansk mashaan brother

    • @oneplus383
      @oneplus383 6 หลายเดือนก่อน

      You live in United States?

  • @nill513
    @nill513 6 หลายเดือนก่อน

    Kudos to you, sir!

    • @mashaan14
      @mashaan14 6 หลายเดือนก่อน

      happy to help..

  • @Sridhar.SubramanianMtech2023
    @Sridhar.SubramanianMtech2023 6 หลายเดือนก่อน

    Thank you for this insightful video 😊

  • @dossantos4415
    @dossantos4415 6 หลายเดือนก่อน

    Could you the same but for NLP

    • @mashaan14
      @mashaan14 6 หลายเดือนก่อน

      I guess you want maps similar to the ones in this paper: Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointly learning to align and translate, 2014 If that's the case, I'm sorry I'm not familiar with that topic.

  • @vivekdabholkar5965
    @vivekdabholkar5965 6 หลายเดือนก่อน

    Mashaan, Very proud of you, your accomplishments, and sharing of knowledge!

    • @mashaan14
      @mashaan14 6 หลายเดือนก่อน

      thanks Vivek.. great to reconnect with you. It’s been a long time!!

  • @Falconoo7383
    @Falconoo7383 7 หลายเดือนก่อน

    Very informative video...

    • @mashaan14
      @mashaan14 7 หลายเดือนก่อน

      Glad you liked it

  • @Falconoo7383
    @Falconoo7383 7 หลายเดือนก่อน

    Thank you for the informative video to explain softmax. Can you please explain how to install manim easily?

    • @mashaan14
      @mashaan14 7 หลายเดือนก่อน

      I’m installing manim into google colab using the command: !pip install manim You can check the code I used to make this video: github.com/mashaan14/manim/blob/main/manim_visualizeSoftmax.ipynb

    • @Falconoo7383
      @Falconoo7383 7 หลายเดือนก่อน

      @@mashaan14 Thank you. Last time, I don't know why but when i used its not working properly. Let me check again.

  • @alexanderzikal7244
    @alexanderzikal7244 7 หลายเดือนก่อน

    Thank You, I tried it out! 1 mistake I found: Inside the class ToyDataset the function "def make_nested_classes(self):" has wrong brackets by n_samples=(int(self.size*0.6), int(self.size*0.2)) on github.

    • @mashaan14
      @mashaan14 7 หลายเดือนก่อน

      Thanks for bringing it up. You're right. It should be square brackets like this: n_samples=[int(self.size*0.6), int(self.size*0.2)]

    • @alexanderzikal7244
      @alexanderzikal7244 7 หลายเดือนก่อน

      A Tupel is needed -> round brackets, then all works fine.

    • @mashaan14
      @mashaan14 7 หลายเดือนก่อน

      I just checked it on scikit-learn . It does need a tuple as an input. That’s strange!! when I run it on colab, it didn’t throw an error. Anyways, I’ll fix it on github.

  • @oneplus383
    @oneplus383 8 หลายเดือนก่อน

    salam alaikum brother Mashan I am a student and I have found your tutorials via Pytorch Geometric, Since you are perfect in Graph Neural networks I wanna ask you something and hope I will find you kind in this regard, I wanna learn Graph neural networks using Jraph library would you please let me know about, How Can I learn it quickly.

    • @mashaan14
      @mashaan14 8 หลายเดือนก่อน

      وعليكم السلام، To be honest I just heard about Jraph library from you. I’m trying to learn Jax, but I haven’t made any tutorials using Jax yet. Anyway, to learn Jraph, I suggest starting with something that you already know the outcome, for example a graph with 10 nodes. I’ll make sure to put Jraph on my todo list ✅.

    • @oneplus383
      @oneplus383 8 หลายเดือนก่อน

      @@mashaan14 brother I appreciate your reply. Can you suggest me steps or things to study to get into the Graph neural network in short terms.

    • @mashaan14
      @mashaan14 8 หลายเดือนก่อน

      @@oneplus383I think Stanford CS224W is good way to start: th-cam.com/video/JAB_plj2rbA/w-d-xo.html

  • @khanfor
    @khanfor 9 หลายเดือนก่อน

    Great explaination Mashaan.

    • @mashaan14
      @mashaan14 9 หลายเดือนก่อน

      thank you 🙏

  • @mohsinaljoaithen2343
    @mohsinaljoaithen2343 9 หลายเดือนก่อน

    Very informative, keep the good work.

    • @mashaan14
      @mashaan14 9 หลายเดือนก่อน

      thank you 🙏

  • @Islam_peacefull_Religion
    @Islam_peacefull_Religion 10 หลายเดือนก่อน

    🥲i wish I could have GPU in my lappi(laptop)

    • @mashaan14
      @mashaan14 10 หลายเดือนก่อน

      You can start playing with GPUs in google colab. But they have limitations on overusing GPUs.

  • @aboudramanediarra7086
    @aboudramanediarra7086 10 หลายเดือนก่อน

    Hello, thank you very much for this beautiful presentation. I work with image data. I'd like to represent them in the form of a graph and predict the links between the pixels. Can you help me with some ideas or a piece of source code? Thanks in advance.

    • @mashaan14
      @mashaan14 10 หลายเดือนก่อน

      I think you need to define some similarities between image pixels. For example, color and position similarities. Those similarities will serve as graph edges. Here’s a good reference on modeling image data as graphs, it summarizes a decade long of research: Arbelaez, P., Maire, M., Fowlkes, C., & Malik, J. (2010). Contour detection and hierarchical image segmentation. IEEE transactions on pattern analysis and machine intelligence, 33(5), 898-916.

    • @darpan2648
      @darpan2648 5 หลายเดือนก่อน

      Thank you..helpful video. Can i use GCN for some biological dataset? How it give results,pls suggest and help

    • @mashaan14
      @mashaan14 5 หลายเดือนก่อน

      @@darpan2648 GCN can be used wherever you got a feature matrix X and an adjacency matrix A. For example, one of the most used datasets in GCN is Cora dataset, which contains a set of documents. The words inside the documents represent the feature matrix X. The citation links between the documents represent the adjacency matrix A. So, before running GCN on the biological dataset, you need to identify what are the feature and adjacency matrices.

  • @guanenteng4870
    @guanenteng4870 10 หลายเดือนก่อน

    thank you. i hope you can create tutorial for spatial temporal data.

    • @mashaan14
      @mashaan14 10 หลายเดือนก่อน

      So glad it helped you. Thanks for your suggestion, I saw several papers using GCN for spatial temporal data. Hopefully, I can write a tutorial on one of these.

  • @bigjeffystyle7011
    @bigjeffystyle7011 11 หลายเดือนก่อน

    Great video. The code is really easy to understand for this application. Thanks for taking the time to summarize and condense a ton of information. Q: Have you seen any implementations in PyTorch Lightning or with DataLoaders for larger applications at scale? I'm looking at using this for a larger production application and any additional sources or packages would be helpful.

    • @mashaan14
      @mashaan14 11 หลายเดือนก่อน

      Glad it was helpful! Actually, I haven’t come across a GCN for large scale applications. But there’s an interesting algorithmic change to GCN. The authors claim it runs faster with the same accuracy. paper: proceedings.mlr.press/v97/wu19e/wu19e.pdf code: github.com/Tiiiger/SGC