Self host Mixtral-8x7B MoE LLM on Mac+cross devices, 2MB AI inference app fully portable

แชร์
ฝัง

ความคิดเห็น • 4

  • @MetodNovak
    @MetodNovak 5 หลายเดือนก่อน +1

    That's impressive. Congrats! Would it be possible to use WasmEdge+Mixtral as a replacement for LMStudio and run the open interpreter with the command 'interpreter --local'? So, instead of LLM Studio, it would use your implementation. That would be something awesome.

    • @SecondStateInc
      @SecondStateInc  5 หลายเดือนก่อน

      Yes, absolutely! This article may be helpful www.secondstate.io/articles/mixtral-8-7b/. Please let us know if you have any questions.

  • @antferdom
    @antferdom 5 หลายเดือนก่อน +1

    Does the WASM model store the whole model checkpoint within his memory space, and if yes, how can a 7B model fit the 4 GB maximum WASM linear memory? How does the WASM module use the CUDA, and in this case, the Apple Metal APIs? Great video!

    • @antferdom
      @antferdom 5 หลายเดือนก่อน

      Any insight?