模型:
DTAI-KULeuven/robbertje-1-gb-non-shuffled
任务:
填充掩码数据集:
oscar dbrd lassy-ud europarl-mono conll2002 3Aconll2002 3Aeuroparl-mono 3Alassy-ud 3Adbrd 3Aoscar语言:
nl预印本库:
arxiv:2101.05716许可:
mitRobBERTje is a collection of distilled models based on RobBERT . There are multiple models with different sizes and different training settings, which you can choose for your use-case.
We are also continuously working on releasing better-performing models, so watch the repository for updates.
Model | Description | Parameters | Training size | Huggingface id |
---|---|---|---|---|
Non-shuffled | Trained on the non-shuffled variant of the oscar corpus, without any operations to preserve this order during training and distillation. | 74 M | 1 GB | this model |
Shuffled | Trained on the publicly available and shuffled OSCAR corpus. | 74 M | 1 GB | DTAI-KULeuven/robbertje-1-gb-shuffled |
Merged (p=0.5) | Same as the non-shuffled variant, but sequential sentences of the same document are merged with a probability of 50%. | 74 M | 1 GB | DTAI-KULeuven/robbertje-1-gb-merged |
BORT | A smaller version with 8 attention heads instead of 12 and 4 layers instead of 6 (and 12 for RobBERT). | 46 M | 1 GB | DTAI-KULeuven/robbertje-1-gb-bort |
We calculated the pseudo perplexity (PPPL) from cite , which is a built-in metric in our distillation library. This metric gives an indication of how well the model captures the input distribution.
Model | PPPL |
---|---|
RobBERT (teacher) | 7.76 |
Non-shuffled | 12.95 |
Shuffled | 18.74 |
Merged (p=0.5) | 17.10 |
BORT | 26.44 |
We also evaluated our models on sereral downstream tasks, just like the teacher model RobBERT. Since that evaluation, a Dutch NLI task named SICK-NL was also released and we evaluated our models with it as well.
Model | DBRD | DIE-DAT | NER | POS | SICK-NL |
---|---|---|---|---|---|
RobBERT (teacher) | 94.4 | 99.2 | 89.1 | 96.4 | 84.2 |
Non-shuffled | 90.2 | 98.4 | 82.9 | 95.5 | 83.4 |
Shuffled | 92.5 | 98.2 | 82.7 | 95.6 | 83.4 |
Merged (p=0.5) | 92.9 | 96.5 | 81.8 | 95.2 | 82.8 |
BORT | 89.6 | 92.2 | 79.7 | 94.3 | 81.0 |