Angela Fan explains NLLB-200: High-quality Machine Translation for Low-Resource Languages

แชร์
ฝัง
  • เผยแพร่เมื่อ 24 ต.ค. 2024

ความคิดเห็น • 6

  • @TheAIEpiphany
    @TheAIEpiphany ปีที่แล้ว +1

    Great interview! Angela was a bit misleading in the "open source" section though, by omitting what is not open-sourced and instead focusing on what is open-sourced. Namely none of their models are open sourced (nor datasets to the best of my knowledge), meaning you can't use any of them for commercial applications and you have to retrain the models from scratch (due to the CC-BY-NC 4.0 license). Which means you have to have 100+ GPUs if you want to train 54B MoE in any reasonable timeframe. With 100 A100 GPUs you need roughly a month. :) (assuming of course you have MT experts and a crystal ball at hand otherwise a lot of additional inefficiencies creep in)

  • @jimbarton36
    @jimbarton36 2 ปีที่แล้ว +1

    Love hearing what Angela Fan has to say!

  • @aliaraabi7367
    @aliaraabi7367 ปีที่แล้ว

    Very nice and inspiring. Thanks Jakub.

  • @RaviAnnaswamy
    @RaviAnnaswamy 2 ปีที่แล้ว

    Very impressive work and great interview too thanks a lot

    • @RaviAnnaswamy
      @RaviAnnaswamy 2 ปีที่แล้ว

      Very great contribution to humanity

  • @elyseemanimpiregasana2117
    @elyseemanimpiregasana2117 2 ปีที่แล้ว

    Very nice Angela, Miss you