模型:
cambridgeltl/sst_mobilebert-uncased
This model provides a MobileBERT (Sun et al., 2020) fine-tuned on the SST data with three sentiments (0 -- negative, 1 -- neutral, and 2 -- positive).
Below, we provide illustrations on how to use this model to make sentiment predictions.
import torch from transformers import AutoTokenizer, AutoConfig, MobileBertForSequenceClassification # load model model_name = r'cambridgeltl/sst_mobilebert-uncased' tokenizer = AutoTokenizer.from_pretrained(model_name) config = AutoConfig.from_pretrained(model_name) model = MobileBertForSequenceClassification.from_pretrained(model_name, config=config) model.eval() ''' labels: 0 -- negative 1 -- neutral 2 -- positive ''' # prepare exemplar sentences batch_sentences = [ "in his first stab at the form , jacquot takes a slightly anarchic approach that works only sporadically .", "a valueless kiddie paean to pro basketball underwritten by the nba .", "a very well-made , funny and entertaining picture .", ] # prepare input inputs = tokenizer(batch_sentences, max_length=256, truncation=True, padding=True, return_tensors='pt') input_ids, attention_mask = inputs.input_ids, inputs.attention_mask # make predictions outputs = model(input_ids=input_ids, attention_mask=attention_mask) predictions = torch.argmax(outputs.logits, dim = -1) print (predictions) # tensor([1, 0, 2])
If you find this model useful, please kindly cite our model as
@misc{susstmobilebert, author = {Su, Yixuan}, title = {A MobileBERT Fine-tuned on SST}, howpublished = {\url{https://huggingface.co/cambridgeltl/sst_mobilebert-uncased}}, year = 2022 }