Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Mixture of Experts Explained (huggingface.co)
29 points by osanseviero on Dec 11, 2023 | hide | past | favorite | 2 comments


The thing I’m confused about is if MoE is supported by Metal? More specifically, can I run the 8x7B on a MacBook (with enough RAM, which I believe is 48GB required?)?


There's nothing about MoE that is incompatible with Metal.

There is a prototype GGUF implementation (https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF) that is out this morning, but it requires a branch that hasn't been merged into the main llamacpp release.

In fact, the #1 challenge with MoE is increased RAM usage, which is something that the unified memory of Macs specializes in.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: