You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hey. I'm starting to try to implement M2M in vLLM and noticed that all the currently supported models use Causal language models (decoder only), while M2M is non-autoregressive and has an encoder-decoder architecture. Is it still possible to implement a vLLM version of M2M or should I just give up 🥲 Appreciate any help with this : )
The text was updated successfully, but these errors were encountered:
I mean specifically M2M-100, the translation model by Meta AI.arXiv paper418M params model card on HF1.2B params model card on HF. Sorry for not being clear enough previously. I'm mostly concerned about the fact that non-autoregressive models include a decoder with separate attention etc so I'm wondering if it's still feasible.
Hey. I'm starting to try to implement M2M in vLLM and noticed that all the currently supported models use Causal language models (decoder only), while M2M is non-autoregressive and has an encoder-decoder architecture. Is it still possible to implement a vLLM version of M2M or should I just give up 🥲 Appreciate any help with this : )
The text was updated successfully, but these errors were encountered: