Stanford CS25: V4 I Demystifying Mixtral of Experts

แชร์
ฝัง
  • เผยแพร่เมื่อ 15 พ.ค. 2024
  • April 25, 2024
    Speaker: Albert Jiang, Mistral AI / University of Cambridge
    Demystifying Mixtral of Experts
    In this talk I will introduce Mixtral 8x7B, a Sparse Mixture of Experts (SMoE) language model. Mixtral has the same architecture as Mistral 7B, with the difference that each layer is composed of 8 feedforward blocks (i.e. experts). For every token, at each layer, a router network selects two experts to process the current state and combines their outputs. Even though each token only sees two experts, the selected experts can be different at each timestep. As a result, each token has access to 47B parameters, but only uses 13B active parameters during inference. I will go into the architectural details and analyse the expert routing decisions made by the model.
    About the speaker:
    Albert Jiang is an AI scientist at Mistral AI, and a final-year PhD student at the computer science department of Cambridge University. He works on language model pretraining and reasoning at Mistral AI, and language models for mathematics at Cambridge.
    More about the course can be found here: web.stanford.edu/class/cs25/
    View the entire CS25 Transformers United playlist: • Stanford CS25 - Transf...

ความคิดเห็น • 4

  • @marknuggets
    @marknuggets 28 วันที่ผ่านมา +1

    Cool format, Stanford quickly becomes my favorite blogger lol

  • @user-uy4rx3hs3x
    @user-uy4rx3hs3x 22 วันที่ผ่านมา +1

    where to get slides

  • @acoustic_boii
    @acoustic_boii 28 วันที่ผ่านมา +1

    Dear Stanford online recently I have completed product management course from Stanford online but i haven't got the certificate help me please how will I get the certificate

    • @Ethan_here230
      @Ethan_here230 27 วันที่ผ่านมา +1

      Wait u will get it
      - Ethan from Stanford