"GPT-4 uses a simple top-2 Token Choice router for MLP MoE layers. It does not use MoE for attention."
GPT won't fix, since "tokens being dropped are generally good for the performance of MoE models."
https://152334h.github.io/blog/knowing-enough-about-moe/#con...
"GPT-4 uses a simple top-2 Token Choice router for MLP MoE layers. It does not use MoE for attention."
GPT won't fix, since "tokens being dropped are generally good for the performance of MoE models."
https://152334h.github.io/blog/knowing-enough-about-moe/#con...