模型:
klue/bert-base
Model Description: KLUE BERT base is a pre-trained BERT Model on Korean Language. The developers of KLUE BERT base developed the model in the context of the development of the Korean Language Understanding Evaluation (KLUE) Benchmark .
from transformers import AutoModel, AutoTokenizer model = AutoModel.from_pretrained("klue/bert-base") tokenizer = AutoTokenizer.from_pretrained("klue/bert-base")
The model can be used for tasks including topic classification, semantic textual similarity, natural language inference, named entity recognition, and other tasks outlined in the KLUE Benchmark .
Misuse and Out-of-scope UseThe model should not be used to intentionally create hostile or alienating environments for people. In addition, the model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.
Significant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021) ). The model developers discuss several ethical considerations related to the model in the paper , including:
For ethical considerations related to the KLUE Benchmark, also see the paper .
The authors use the following pretraining corpora for the model, described in the associated paper :
We gather the following five publicly available Korean corpora from diverse sources to cover a broad set of topics and many different styles. We combine these corpora to build the final pretraining corpus of size approximately 62GB.
The authors also describe ethical considerations related to the pretraining corpora in the associated paper .
Training Procedure PreprocessingThe authors describe their preprocessing procedure in the associated paper :
We filter noisy text and non-Korean text using the same methods from Section 2.3 (of the paper). Each document in the corpus is split into sentences using C++ implementation (v1.3.1.) of rule-based Korean Sentence Splitter (KSS) . For CC-100-Kor and NEWSCRAWL, we keep sentences of length greater than equal to 200 characters, as a heuristics to keep well-formed sentences. We then remove sentences included in our benchmark task datasets, using BM25 as a sentence similarity metric ( reference ).
TokenizationThe authors describe their tokenization procedure in the associated paper :
We design and use a new tokenization method, morpheme-based subword tokenization. When building a vocabulary, we pre-tokenize a raw text into morphemes using a morphological analyzer, and then we apply byte pair encoding (BPE) ( Senrich et al., 2016 ) to get the final vocabulary. For morpheme segmentation, we use Mecab-ko , MeCab ( Kudo, 2006 ) adapted for Korean, and for BPE segmentation, we use the wordpiece tokenizer from Huggingface Tokenizers library . We specify the vocabulary size to 32k. After building the vocabulary, we only use the BPE model during inference, which allows us to tokenize a word sequence by reflecting morphemes without a morphological analyzer. This improves both usability and speed.
The training configurations are further described in the paper .
The model was evaluated on the KLUE Benchmark . The tasks and metrics from the KLUE Benchmark that were used to evaluate this model are described briefly below. For more information about the KLUE Benchmark, see the data card , Github Repository , and associated paper .
Task: Topic Classification (TC) - Yonhap News Agency Topic Classification (YNAT), Metrics: Macro F1 score, defined as the mean of topic-wise F1 scores, giving the same importance to each topic.
Task: Semantic Textual Similarity (STS), Metrics: Pearsons' correlation coefficient (Pearson’ r) and F1 score
Task: Natural Language Inference (NLI), Metrics: Accuracy
Task: Named Entity Recognition (NER), Metrics: Entity-level macro F1 (Entity F1) and character-level macro F1 (Char F1) scores
Task: Relation Extraction (RE), Metrics: Micro F1 score on relation existing cases and area under the precision- recall curve (AUPRC) on all classes
Task: Dependency Parsing (DP), Metrics: Unlabeled attachment score (UAS) and labeled attachment score (LAS)
Task: Machine Reading Comprehension (MRC), Metrics: Exact match (EM) and character-level ROUGE-W (ROUGE), which can be viewed as longest common consecutive subsequence (LCCS)-based F1 score.
Task: Dialogue State Tracking (DST), Metrics: Joint goal accuracy (JGA) and slot micro F1 score (Slot F1)
Task | TC | STS | NLI | NER | RE | DP | MRC | DST | ||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Metric | F1 | Pearsons' r | F1 | ACC | Entity F1 | Char F1 | F1 | AUPRC | UAS | LAS | EM | ROUGE | JGA | Slot F1 |
85.73 | 90.85 | 82.84 | 81.63 | 83.97 | 91.39 | 66.44 | 66.17 | 89.96 | 88.05 | 62.32 | 68.51 | 46.64 | 91.61 |
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019) . We present the hardware type based on the associated paper .
See the associated paper for details on the modeling architecture (BERT), objective, compute infrastructure, and training details.
@misc{park2021klue, title={KLUE: Korean Language Understanding Evaluation}, author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho}, year={2021}, eprint={2105.09680}, archivePrefix={arXiv}, primaryClass={cs.CL} }