Matchue
Matchue
  • 26
  • 171 332
Double Kill VR Gameplay December 2021
Double Kill is a Looter Shooter I began working on for a class project. It was previously named Dual Wield, and will probably be something different in the future.
It is a looter shooter game, which takes place on procedurally generated islands. Wave of enemies attack you, and you must collect guns (from killing enemies and buying from the shopkeeper), which you'll use to defend yourself.
New Features:
- Day night cycle
- Skills (based on level, including mostly kill skills)
- Better guns (with ricocheting bullets, exploding bullets, and more)
- Different plane enemies (bomber and biplane)
- Improved graphics (fog, gun lights, better particles, and more)
มุมมอง: 178

วีดีโอ

Anomaly Detection on Image Data using Generative Adversarial Networks
มุมมอง 4.4K2 ปีที่แล้ว
My Second Undergraduate Honours Seminar at the University of Regina.
Dual Wield VR Gameplay
มุมมอง 1213 ปีที่แล้ว
Dual Wield is a Looter Shooter I began working on for a class project. It is a looter shooter game, which takes place on procedurally generated islands. Wave of enemies attack you, and you must collect guns (from killing enemies and buying from the shopkeeper), which you'll use to defend yourself.
StarGAN explained in 5 minutes! Guest Video: Kallin
มุมมอง 4.9K3 ปีที่แล้ว
StarGAN is a deep learning model for image translation, which uses an intuitive architecture to translate between any number of classes. This video was created and presented by my good friend and coworker, Kallin Kehrig! CycleGAN: arxiv.org/abs/1703.10593 StarGAN: arxiv.org/abs/1711.09020 Adaptive Instance Normalization: arxiv.org/abs/1703.06868 Modulated Convolutions(StyleGAN 2): arxiv.org/abs...
Utilizing Generative Deep Learning on Small Datasets - Honours Seminar
มุมมอง 5623 ปีที่แล้ว
Utilizing Generative Deep Learning on Small Datasets. Honours Seminar given at University of Regina by Matthew Mann. This video is a draft for the actual seminar presentation, before review and revision.
Using AI to Create Stunningly Beautiful Worlds | StyleGAN 2 with ADA
มุมมอง 1.3K3 ปีที่แล้ว
NVidia's StyleGAN 2 with ADA can be used to create stunning, picturesque places that don't exist in real life. Music: Feels - Patrick Patrikios Dolphin-esque - Godmode
What is the Lottery Ticket Hypothesis, and why is it important?
มุมมอง 6K3 ปีที่แล้ว
The Lottery Ticket Hypothesis suggests that, even before training, a neural network can be pruned to a fraction of its original size, and still achieve comparable results. It's implications are very important for architecture design and training efficiency. I'll be creating a video on the update to StyleGAN 2 (ADA) soon, so be sure to watch for that! Sources: Paper: arxiv.org/abs/1803.03635 Mus...
Absurdity and the Myth of Sisyphus
มุมมอง 1803 ปีที่แล้ว
Absurdity arises from the fundamental disharmony between one’s search for meaning and the lack of meaning to be found in the Universe. This video describes Albert Camus's take on Absurdity from "The Myth Of Sisyphus." The Myth Of Sisyphus uses the ancient story of Sisyphus to illustrate how, through acceptance of the Absurd, we are able to live on happily in spite of it.
BiGAN Explained In 3 Minutes!
มุมมอง 6K3 ปีที่แล้ว
BiGAN (Bidirectional GAN) is an architecture designed to unsupervised feature learning. Here's BiGAN's main concepts explained simply in under 4 minutes. Thanks for watching! Sources: Original BiGAN Paper: arxiv.org/abs/1605.09782 Follow-Up Paper (BigBiGAN): arxiv.org/abs/1907.02544 Music: With a Stamp by Twin Musicom is licensed under a Creative Commons Attribution license (creativecommons.org...
CycleGAN Explained in 5 Minutes!
มุมมอง 37K4 ปีที่แล้ว
CycleGAN is an architecture designed to perform unpaired image-to-image translation. Here's CycleGAN's main concepts explained simply in under 5 minutes. Thanks for watching! Find my implementations of CycleGAN in Tensorflow 2.0 here: github.com/manicman1999/CycleGAN-Tensorflow-2.0 Sources: Original CycleGAN: arxiv.org/abs/1703.10593 Pix2Pix: arxiv.org/abs/1611.07004 BicycleGAN: arxiv.org/abs/1...
How Does StyleGAN 2 Work? Replacing Growth
มุมมอง 3.2K4 ปีที่แล้ว
How does StyleGAN 2 Work? This is the second video of a three part series outlining the main improvements StyleGAN 2 made to StyleGAN. In this video I go through the theory behind the Generator's new skip connections, and the Discriminator's residual blocks. My implementation for StyleGAN 2 in TensorFlow 2.0 can be found here: github.com/manicman1999/StyleGAN2-Tensorflow-2.0 Music: With a Stamp...
How Does StyleGAN 2 Work? Modulated Convolution Tutorial & Code
มุมมอง 10K4 ปีที่แล้ว
How does StyleGAN 2 work? In the first part of a three part series, I go through the theory behind modulated/demodulated convolution; a replacement for adaptive instance normalization in StyleGAN. I include a high level explanation, then an implementation in Tensorflow 2.0. My full implementation can be found here: github.com/manicman1999/StyleGAN2-Tensorflow-2.0 Music: Brittle Rille - Reunited...
NVidia just released StyleGAN 2 - And It's Mind Blowing!
มุมมอง 62K4 ปีที่แล้ว
StyleGAN 2 generates beautiful looking images of human faces. Released as an improvement to the original, popular StyleGAN by NVidia, StyleGAN 2 improves on the quality of images, as well as providing a novel way to detect whether an image is real or fake. The results from this work are absolutely incredible! Music Musik Von Melodies by Mylar Melodies Links StyleGAN 2 - nvlabs.github.io/stylega...
StyleGAN 2 Truncation
มุมมอง 3.5K4 ปีที่แล้ว
StyleGAN 2 trained on images of landscapes, with varying levels of truncation. At the beginning, all images have been fully truncated, showing the "average" landscape of all generated landscapes. When ψ = 1, no truncation is applied, showing full diversity across the distribution of images. When ψ = 1.5, generated images can reach beyond its own distribution, lowering image quality significantl...
These Cats Do Not Exist
มุมมอง 1.1K4 ปีที่แล้ว
Cat images, made using my own custom implementation of StyleGAN! Song: Peace by HOVATOFF
New York, Stylized
มุมมอง 4.5K4 ปีที่แล้ว
New York, Stylized
Creative AI - Generating Flowers (StyleGAN)
มุมมอง 1.3K5 ปีที่แล้ว
Creative AI - Generating Flowers (StyleGAN)
StyleGAN Part 2: Noise And Constant Input - Matchue GANs
มุมมอง 4.4K5 ปีที่แล้ว
StyleGAN Part 2: Noise And Constant Input - Matchue GANs
StyleGAN Part 1: Adaptive Instance Normalization - Matchue GANs
มุมมอง 11K5 ปีที่แล้ว
StyleGAN Part 1: Adaptive Instance Normalization - Matchue GANs
More with Less: Data Augmentation and Gradient Penalties - Matchue GANs
มุมมอง 2K5 ปีที่แล้ว
More with Less: Data Augmentation and Gradient Penalties - Matchue GANs
Generative Adversarial Networks - Matchue GANs
มุมมอง 2.9K5 ปีที่แล้ว
Generative Adversarial Networks - Matchue GANs
Automatic Generation of Video Game Character Images using Augmented Structure-and-Style Networks
มุมมอง 4105 ปีที่แล้ว
Automatic Generation of Video Game Character Images using Augmented Structure-and-Style Networks
StyleGAN Flowers Interpolation
มุมมอง 1.2K5 ปีที่แล้ว
StyleGAN Flowers Interpolation
StyleGAN Landscapes Interpolation
มุมมอง 3.1K5 ปีที่แล้ว
StyleGAN Landscapes Interpolation
Creating An MNIST Classifier | Exercise 1 | NaN to GAN
มุมมอง 3385 ปีที่แล้ว
Creating An MNIST Classifier | Exercise 1 | NaN to GAN
Deep Learning | Episode 1 | NaN to GAN
มุมมอง 5515 ปีที่แล้ว
Deep Learning | Episode 1 | NaN to GAN

ความคิดเห็น

  • @Hateusernamearentu
    @Hateusernamearentu หลายเดือนก่อน

    very clear. thanks. especially the cylcle picture at the end. clears up everything.

  • @hamzashahzad9507
    @hamzashahzad9507 4 หลายเดือนก่อน

    LUN KHAAAAA

  • @hamzashahzad9507
    @hamzashahzad9507 4 หลายเดือนก่อน

    GHALAT BTA RHA HAI TU

  • @tejasmathursriganesh3356
    @tejasmathursriganesh3356 5 หลายเดือนก่อน

    chris griffin

  • @TimelessHipHop1987
    @TimelessHipHop1987 5 หลายเดือนก่อน

    All the world's technology started when Lucifer was cast out of Heaven by our heavenly Truly God

  • @patrickleddy4416
    @patrickleddy4416 7 หลายเดือนก่อน

    Are you aware of any AI that can intake a technical specification image (lets say it's a complex chart or graph, with hundreds of separate lines of text), and reproduce that tech specification image with variations that make it unique?

  • @reshankjain6622
    @reshankjain6622 9 หลายเดือนก่อน

    I'm writing a research paper on GAN usage and I would love to be able to cite you. You have compiled days worth of information into a 5 minute video. Simply phenomenal.

  • @fadaontop
    @fadaontop 10 หลายเดือนก่อน

    Great Job. I would like to ask few questions from you about using MONAI for unpair medical images. Since I have few data to work with and I am new to Machine learning . I intend to downsample the few images (about 500) I have and used the two scenario (Gen and Dis) for the two data sets to be trained. Can this approach work ? Which should I place at the Generative model and Discriminator?

  • @mrudulpawar7944
    @mrudulpawar7944 10 หลายเดือนก่อน

    Thanks a lot for this !

  •  11 หลายเดือนก่อน

    amazing content. It helped me a lot

  • @araldjean-charles3924
    @araldjean-charles3924 ปีที่แล้ว

    For the initial conditions that work, have anybody look at how much wiggle room you have. Is there an epsilon-neighborhood of the initial state you can safely start from, and how small is epsilon?

  • @andycao966
    @andycao966 ปีที่แล้ว

    Great job! Do you think we can make a lottery prediction by using the LTH method? Thanks.

  • @bertgallager1894
    @bertgallager1894 ปีที่แล้ว

    Great video

  • @henkjekel4081
    @henkjekel4081 ปีที่แล้ว

    Lol, its completely wrong what you say. Go study first before you make a video mate

    • @kkkk-sh1sb
      @kkkk-sh1sb ปีที่แล้ว

      can you please tell me where is wrong, cos i m interested in this concept. thx

  • @psychologienerd7546
    @psychologienerd7546 ปีที่แล้ว

    The final bit causes problems: if __name__ == "__main__": model = WGAN(lr = 0.0001, silent = False) while(True): model.train() see below

  • @psychologienerd7546
    @psychologienerd7546 ปีที่แล้ว

    ValueError: `tf.compat.v1.keras` Optimizer (<__main__.Adam_lr_mult object at 0x7fb49b609fa0>) is not supported when eager execution is enabled. Use a `tf.keras` Optimizer instead, or disable eager execution.

  • @hardikvegad3508
    @hardikvegad3508 ปีที่แล้ว

    Hey can you help me understand how we can find anomaly on text document, and how we can specify something specifically as anomaly in an documnet.

  • @돌망-c7k
    @돌망-c7k ปีที่แล้ว

    hi how to make this video?

  • @michaelyoon7061
    @michaelyoon7061 2 ปีที่แล้ว

    I absolutely loved your video, thank you so much for helping me understand cycle gan. Subscribed :).

  • @kobruh9642
    @kobruh9642 2 ปีที่แล้ว

    So far so good!

  • @masonmiao849
    @masonmiao849 2 ปีที่แล้ว

    Thanks

  • @chrisdutoit6497
    @chrisdutoit6497 2 ปีที่แล้ว

    Really nice figures!

  • @radovicmiloskg
    @radovicmiloskg 2 ปีที่แล้ว

    You nailed it!

  • @hazed834
    @hazed834 2 ปีที่แล้ว

    nicely explained

  • @joboring8397
    @joboring8397 2 ปีที่แล้ว

    1:27 "Generated...from absolutely nothing." Yeah, absolutely nothing...except for the millions of training images and roughly 0.7 megawatt hours of electricity used during training (aka "absolutely nothing"). Not to mention the 131.61 MWh of electricity (~$13,000) used to power multiple $150,000 GPUs during the development of StyleGAN2 (starting from StyleGAN).

  • @tijovvinu
    @tijovvinu 2 ปีที่แล้ว

    Thanks for the presentation. May i get the code and dataset you used here?

  • @erknee6071
    @erknee6071 2 ปีที่แล้ว

    I just want to ask why you don't use reshape directly to fuse batch & channel but change dimension formulation to (B,C,H,W) first ? And normalize kernel group by norm do not compute variance along the group because the mean is not zero

  • @car6647
    @car6647 2 ปีที่แล้ว

    thanks!

  • @chenweicui7887
    @chenweicui7887 2 ปีที่แล้ว

    This video is really informative, includes definition, intuition, and experiments done by the creator himself, while still being short. Thank you for the good work!

  • @ivanxyz1
    @ivanxyz1 2 ปีที่แล้ว

    Man oh man! Geez!

  • @kuruvillaabraham9463
    @kuruvillaabraham9463 2 ปีที่แล้ว

    Hi Matchue ,Nice Video ,Are we able to change color of the Pokemon

  • @Aniket7Tomar
    @Aniket7Tomar 2 ปีที่แล้ว

    Does the paper really state that you can prune without training to convergence? I thought the big deal was that they showed the existence of a smaller network that can be trained if we use the same initialization. I don't think they were able to find this network without training to convergence.

  • @hdubbs9174
    @hdubbs9174 2 ปีที่แล้ว

    Hello, in your TF2.0 implementation I'm not really sure how the Style Mapping is actually implemented? It doesn't look like it's actually used anywhere. Could you please explain? Thanks for the video and sharing. You explained the concepts very nicely.

  • @mingxintian9642
    @mingxintian9642 2 ปีที่แล้ว

    Hi Matchue thank you for your presentation. Could you provide an example of a practical application using this method in the medical field?

  • @tahminaazmin655
    @tahminaazmin655 2 ปีที่แล้ว

    Thanks for your nice presentation. Into the Anomaly-Agnostic Method slide, how the method can equally perform on any subdomain of anomalous samples?

  • @iviekokobili2434
    @iviekokobili2434 2 ปีที่แล้ว

    Thanks for the presentation, what are some of the current real world application of anomaly detection on image data using generative adversarial networks

  • @neelpatel2600
    @neelpatel2600 2 ปีที่แล้ว

    Nice presentation. Is there any image preprocessing done before giving it as an input to model, like converting to grayscale, resizing, applying filters, etc?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      Yes. The images are all resized to the same size (here it was 128x128), then duplicated and mirrored horizontally to expand the size of the dataset.

  • @HarshPatel-zq5uh
    @HarshPatel-zq5uh 2 ปีที่แล้ว

    Nice Presentation. Are there any limitations of using the Anomaly-Agnostic method?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      This method simply doesn’t perform as well as the classification method. As such, it should only be used if one has few or no sample anomalies.

  • @ryanlaube5486
    @ryanlaube5486 2 ปีที่แล้ว

    Thanks for the presentation Matthew! I was wondering with which processes does the discriminator in a GAN learn to classify between real and fake images.

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      Typically, simple objective functions can be used to train a Discriminator to classify between its two options. Cross Entropy can be used, although simpler functions like Hinge have been used as well; ReLU(1 + label) where label is either -1 or 1.

  • @bennetteidsness3275
    @bennetteidsness3275 2 ปีที่แล้ว

    Would you consider the anomaly-agnostic method or the classification method to be more effective?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      The classification would be the most effective method. However, it can only be used if you have examples of anomalies.

  • @kajalshah5478
    @kajalshah5478 2 ปีที่แล้ว

    In the BiGAN slide you mentioned that the method of training is different? Can you please explain it that what is the difference?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      Sure! Since there’s a new model, the Encoder, that model now also has to be trained using the Discriminator. As such, the Discriminator looks at an input pair; either (real image, image encoding) or (fake image, latent vector used to generate image). Now both the generator and encoder must reliably fool the discriminator, by learning to correspond.

  • @henriksteenbock2628
    @henriksteenbock2628 2 ปีที่แล้ว

    You mention that this method could also be used to detect medical anomalies/emergencies. However, finding out that you have an anomaly is only the first step, its just as important to find out where that anomaly is located. Do you think your high-level feature vector can also help locate such anomalies, or at the very least determine a rough center point of them in an image?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      Very interesting question! I hadn’t thought much into this, but I suppose reverse engineering the model (e.g. testing which pixels have the largest gradients w.r.t. the final classification) could give an indication. A fellow student I’ve worked with is researching LIME, which could also be another way to go; this could give a rough indication what portions of the image have to do with the classification.

  • @jayjoshi4618
    @jayjoshi4618 2 ปีที่แล้ว

    You mentioned that autoencoders have their limitations. Are these limitations related to distance measures? What other reconstruction loss functions we can use in GAN's? I guess you are using something similar to mse?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      These limitations are related to the fact distance measures are being used at all. The reason GANs perform better on this task is because it doesn’t directly compare any two images, but rather aims to generate realistic looking dataset images in general, allowing for features to be much more relevant to the dataset. Using anything that relies on distance measures will lead to the mapped features being reliant on low-level features.

  • @sushitharajeev6836
    @sushitharajeev6836 2 ปีที่แล้ว

    That's a nice presentation! How does it map on each pixel in the slide Image distribution mapping and Does it use any Image Processing technique for altering the pixel?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      The network outputs a SxSx3 tensor, where S is the side length of the image. These values can be anything, but are clipped afterward to be between 0 and 1, like RGB images typically are handled.

  • @aymenbensaid1121
    @aymenbensaid1121 2 ปีที่แล้ว

    In the classification results, is there a justification of using those threshold values (0.2, 0.5, 0.8)? maybe due to some characteristics of the used data, or they were randomly chosen?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      I chose 0.5, as it made the most logical sense to split numbers ranging from 0 to 1 at the halfway point. 0.2 and 0.8 were chosen arbitrarily to see the effect of having a threshold near 0 and near 1.

  • @amalluminati
    @amalluminati 2 ปีที่แล้ว

    Hi Matthew, thank you for your presentation. The generator of the GAN architecture takes random numbers from a normal distribution and transforms it into an image, how are the numbers used to create these images if you could explain in a simple way, does it play around with the rgb values of pixels by assigning the numbers to them ?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      The GAN Generator uses various deep learning layers (specifically convolution, activation and upsampling layers) to translate the noise into the shape of an image, with RGB values for each pixel.

  • @kpdave25
    @kpdave25 2 ปีที่แล้ว

    Thank you for your presentation. Is there any example where they are using a similar system or these can be used?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      These systems are often used in dataset cleaning. Other times, these systems can be used to faulty infrastructure, machines, etc.

  • @yixingchen8896
    @yixingchen8896 2 ปีที่แล้ว

    Hello, thanks for the presentation. Regarding your final results, is there any process to determine the threshold? Will there be a threshold where the three evaluation matrices reach a balance?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      To determine the threshold one could sample a bunch of different possible thresholds at regular intervals and choose the one that minimized false positive % + false negative %. I did this in the model-agnostic portion.

  • @ebastudios9360
    @ebastudios9360 2 ปีที่แล้ว

    We delivered. Please can you explain the results -False Positive and False Negative. In the Anomaly-Agnostic Method slide?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      A false positive would be a typical image labelled as anomalous. A false negative would be an anomalous image labelled as typical. These numbers can be adjusted by adjusting the threshold for classifying an image as either positive or negative.

  • @arminsadreddin9113
    @arminsadreddin9113 2 ปีที่แล้ว

    Nice presentation. 1- What is the input of the generating deep learning model of GAN? 2- Is it possible to use CNN for autoencoders to extract “high level features” for dimensionality reduction?

    • @Matchue624
      @Matchue624 2 ปีที่แล้ว

      The input to the Generator network in a GAN is a latent vector. This vector consists of some number of random noise, typically normally distributed noise with a mean of zero and standard deviation of one. While this is possible, its quality is significantly worse. Due to the autoencoder attempting to reconstruct each pixel, it will lead to high-level features that have to do only with color, rather than with what we’d want for this task (such as objects or dataset-related features).