模型:
cambridgeltl/mirror-bert-base-uncased-word
language: en
tags:
An unsupervised word encoder proposed by Liu et al. (2021) . Trained with a set of unlabelled words, using bert-base-uncased as the base model. Please use [CLS] as the representation of the input.
@inproceedings{ liu2021fast, title={Fast, Effective and Self-Supervised: Transforming Masked LanguageModels into Universal Lexical and Sentence Encoders}, author={Liu, Fangyu and Vuli{\'c}, Ivan and Korhonen, Anna and Collier, Nigel}, booktitle={EMNLP 2021}, year={2021} }