Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feasible to implement non-autoregressive LMs? (M2M) #1072

Closed
Axreub opened this issue Sep 17, 2023 · 3 comments
Closed

Feasible to implement non-autoregressive LMs? (M2M) #1072

Axreub opened this issue Sep 17, 2023 · 3 comments

Comments

@Axreub
Copy link

Axreub commented Sep 17, 2023

Hey. I'm starting to try to implement M2M in vLLM and noticed that all the currently supported models use Causal language models (decoder only), while M2M is non-autoregressive and has an encoder-decoder architecture. Is it still possible to implement a vLLM version of M2M or should I just give up 🥲 Appreciate any help with this : )

@irasin
Copy link
Contributor

irasin commented Sep 18, 2023

Hi, @Axreub, I just wonder what is the non-autoregressive LMs(M2M), can you share the paper or code about it? Thanks a lot.

@Axreub
Copy link
Author

Axreub commented Sep 18, 2023

I mean specifically M2M-100, the translation model by Meta AI.arXiv paper 418M params model card on HF 1.2B params model card on HF. Sorry for not being clear enough previously. I'm mostly concerned about the fact that non-autoregressive models include a decoder with separate attention etc so I'm wondering if it's still feasible.

@hmellor
Copy link
Member

hmellor commented Mar 8, 2024

Closing as duplicate of #187

@hmellor hmellor closed this as completed Mar 8, 2024
@hmellor hmellor closed this as completed Feb 27, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants