模型:
RUCAIBox/elmer
The ELMER model was proposed in ELMER: A Non-Autoregressive Pre-trained Language Model for Efficient and Effective Text Generation by Junyi Li, Tianyi Tang, Wayne Xin Zhao, Jian-Yun Nie and Ji-Rong Wen.
The detailed information and instructions can be found https://github.com/RUCAIBox/ELMER .
ELMER is an efficient and effective PLM for NAR text generation, which generates tokens at different layers by leveraging the early exit technique.
The architecture of ELMER is a variant of the standard Transformer encoder-decoder and poses three technical contributions:
To fine-tune ELMER on non-autoregressive text generation:
>>> from transformers import BartTokenizer as ElmerTokenizer >>> from transformers import BartForConditionalGeneration as ElmerForConditionalGeneration >>> tokenizer = ElmerTokenizer.from_pretrained("RUCAIBox/elmer") >>> model = ElmerForConditionalGeneration.from_pretrained("RUCAIBox/elmer")
@article{lijunyi2022elmer, title={ELMER: A Non-Autoregressive Pre-trained Language Model for Efficient and Effective Text Generation}, author={Li, Junyi and Tang, Tianyi and Zhao, Wayne Xin and Nie, Jian-Yun and Wen, Ji-Rong}, booktitle={EMNLP 2022}, year={2022} }