BERT for Topic Modeling - EXPLAINED!

แชร์
ฝัง
  • เผยแพร่เมื่อ 15 ธ.ค. 2024

ความคิดเห็น • 35

  • @nextXstep
    @nextXstep 3 ปีที่แล้ว +1

    Couldn't find a single article explaining the codes behind bertopic, the explanation in this video is absolutely perfect thanks!!

  • @pierrickleroy780
    @pierrickleroy780 2 ปีที่แล้ว +1

    Another clear explanation of multiple sota useful concepts, cheers man I really like your videos and way of communicating things !

  • @IRiViI
    @IRiViI 3 ปีที่แล้ว +2

    I just wanted to say that I love your videos!

  • @MLA263
    @MLA263 ปีที่แล้ว

    This is amazing, thank you, you hero

  • @chetanetrx2005
    @chetanetrx2005 3 ปีที่แล้ว +4

    Nice video..
    Can you please explain how sentence transformer works at inference time when we have only one sentence?

  • @lucyledezma709
    @lucyledezma709 2 ปีที่แล้ว

    Great video!!!🤗🤗🤗

  • @lepton555
    @lepton555 3 ปีที่แล้ว +1

    Awesome job, man! Absolutely horrific! Thank you!

  • @teetanrobotics5363
    @teetanrobotics5363 3 ปีที่แล้ว

    Amazing content bro. Please keep updating your playlists.

  • @gunavardhinip
    @gunavardhinip ปีที่แล้ว

    Brother will you explaint bart for text summarization

  • @yazdipour
    @yazdipour 3 ปีที่แล้ว

    love ur bert/nlp contents

    • @CodeEmporium
      @CodeEmporium  3 ปีที่แล้ว

      Glad you do. I'll try making more of this :)

  • @kabeerjaffri4015
    @kabeerjaffri4015 3 ปีที่แล้ว +4

    Hey can you make a video on unsupervised temporal action localization in video. like when you search google for somthing and they show time intervals where video content matches your query.
    I think its a great topic and may spark your interest. BTW great content as usual

  • @emanuelgerber
    @emanuelgerber 3 ปีที่แล้ว

    Thanks for making this video! Helps a lot

  • @alexnim4873
    @alexnim4873 3 ปีที่แล้ว

    Amazing video bro. Very helpful. Thanks so much

  • @datasci4547
    @datasci4547 3 ปีที่แล้ว

    Very clear explanation. Thank you :)

  • @Daniel-gy1rc
    @Daniel-gy1rc 2 ปีที่แล้ว

    can u make a video about Google-Palm?

  • @user-or7ji5hv8y
    @user-or7ji5hv8y 3 ปีที่แล้ว

    Really well explained

  • @boubacarbah1455
    @boubacarbah1455 2 ปีที่แล้ว

    Hello , i'm trying to reproduce your exercice. But i got a problem when i tried to import BERTOPIC " from import bertopic ".I get this error " no module named "llvmlite.binding.dylib". And i could not fix it; Si i wonder if you have a solution ?

  • @wilfredomartel7781
    @wilfredomartel7781 ปีที่แล้ว

    Just a doubt. So tripplet dataset is better for improving embeddings. May a video of how to fine tuning a non english transformer?

    • @CodeEmporium
      @CodeEmporium  ปีที่แล้ว +1

      Sure thing. The goal is to make a series in BERT training and code soon from scratch. In the mean time, maybe you’ll enjoy the playlist called “Transformers from Scratch” where I build a translator for a non-English language. Though there is no “pre-training” and “fine tuning”, many components of BERT are similar to the transformer. So I recommend checking that out

    • @wilfredomartel7781
      @wilfredomartel7781 ปีที่แล้ว

      @@CodeEmporium thank you to take your time in resondijg my message. Maybe a link suggestion my friend.

    • @CodeEmporium
      @CodeEmporium  ปีที่แล้ว

      Transformers from scratch
      th-cam.com/play/PLTl9hO2Oobd97qfWC40gOSU8C0iu0m2l4.html

  • @maryamaziz3841
    @maryamaziz3841 3 ปีที่แล้ว

    Hi, How can use BERT to word embedding

  • @geoffreyanderson4719
    @geoffreyanderson4719 3 ปีที่แล้ว

    My spidey-sense tingles for me when more than half of the topics in the corpus are unclustered. Exploratory data analysis on those might reveal some easily fixed errors. Like, I might want to see what happens when you topic model just those bad bois, letting all the rest of the good bois to go home on the regular schoolbus. Maybe the Breakfast Club misfits have something in common with each other after all.

    • @CodeEmporium
      @CodeEmporium  3 ปีที่แล้ว

      Aye good eye. I was mostly just trying to illustrate BERT. But some analysis on this would have been nice too as a follow up

  • @Burakyesilvlogs
    @Burakyesilvlogs 3 ปีที่แล้ว

    Great video

  • @pranaymehta8471
    @pranaymehta8471 3 ปีที่แล้ว

    Have you tried to work with BERTopic with datasets that are bigger in size ? For instance 100k, 500k data sizes ? From what I have seen, the sentence transformer takes a lot of time to create the N dimensional embeddings . I am not sure if berTopic runs things in parallel.

    • @CodeEmporium
      @CodeEmporium  3 ปีที่แล้ว +2

      Good question. Sorry I'm late to this. I've worked with Sentence Transformers in general and i can say the speed and quality really depends on which sentence transformer you use. You'll just need to choose the one that balances both qualities if speed is Essential (like for online applications as opposed to postmortem analysis)
      Also BERT and hence the sentence transformer process information in parallel. I know Sentence Transformers have this method called "embedding()" where you can pass in a list and we fetch the embeddings in parallel

  • @anonxnor
    @anonxnor 3 ปีที่แล้ว

    You should be able to do this with the CLS token embeddings, instead of sentence embeddings from S-BERT, if you use regular BERT right?

  • @greenital
    @greenital 3 ปีที่แล้ว

    in love ** divine * :D