模型:
alabnii/jmedroberta-base-manbyo-wordpiece
This is a Japanese RoBERTa base model pre-trained on academic articles in medical sciences collected by Japan Science and Technology Agency (JST).
This model is released under the Creative Commons 4.0 International License (CC BY-NC-SA 4.0).
ReferenceJa:
@InProceedings{sugimoto_nlp2023_jmedroberta, author = "杉本海人 and 壹岐太一 and 知田悠生 and 金沢輝一 and 相澤彰子", title = "J{M}ed{R}o{BERT}a: 日本語の医学論文にもとづいた事前学習済み言語モデルの構築と評価", booktitle = "言語処理学会第29回年次大会", year = "2023", url = "https://www.anlp.jp/proceedings/annual_meeting/2023/pdf_dir/P3-1.pdf" }
En:
@InProceedings{sugimoto_nlp2023_jmedroberta, author = "Sugimoto, Kaito and Iki, Taichi and Chida, Yuki and Kanazawa, Teruhito and Aizawa, Akiko", title = "J{M}ed{R}o{BERT}a: a Japanese Pre-trained Language Model on Academic Articles in Medical Sciences (in Japanese)", booktitle = "Proceedings of the 29th Annual Meeting of the Association for Natural Language Processing", year = "2023", url = "https://www.anlp.jp/proceedings/annual_meeting/2023/pdf_dir/P3-1.pdf" }
Before using the model, make sure that Manbyo Dictionary has been downloaded under /usr/local/lib/mecab/dic/userdic .
# download Manbyo-Dictionary mkdir -p /usr/local/lib/mecab/dic/userdic wget https://sociocom.jp/~data/2018-manbyo/data/MANBYO_201907_Dic-utf8.dic mv MANBYO_201907_Dic-utf8.dic /usr/local/lib/mecab/dic/userdic
Note: If you don't have root privileges and find it difficult to download the Manbyo Dictionary to /usr/local/lib/mecab/dic/userdic , you can still load our model by overriding tokenizer settings as follows:
# download Manbyo-Dictionary wherever you like wget https://sociocom.jp/~data/2018-manbyo/data/MANBYO_201907_Dic-utf8.dic mv MANBYO_201907_Dic-utf8.dic /anywhere/you/like
from transformers import AutoModelForMaskedLM, AutoTokenizer model = AutoModelForMaskedLM.from_pretrained("alabnii/jmedroberta-base-manbyo-wordpiece") tokenizer = AutoTokenizer.from_pretrained("alabnii/jmedroberta-base-manbyo-wordpiece", **{ "mecab_kwargs": { "mecab_option": "-u /anywhere/you/like/MANBYO_201907_Dic-utf8.dic" } })
Input text must be converted to full-width characters(全角)in advance.
You can use this model for masked language modeling as follows:
from transformers import AutoModelForMaskedLM, AutoTokenizer model = AutoModelForMaskedLM.from_pretrained("alabnii/jmedroberta-base-manbyo-wordpiece") model.eval() tokenizer = AutoTokenizer.from_pretrained("alabnii/jmedroberta-base-manbyo-wordpiece") texts = ['この患者は[MASK]と診断された。'] inputs = tokenizer.batch_encode_plus(texts, return_tensors='pt') outputs = model(**inputs) tokenizer.convert_ids_to_tokens(outputs.logits[0][1:-1].argmax(axis=-1)) # ['この', '患者', 'は', 'ALS', 'と', '診断', 'さ', 'れ', 'た', '。']
Alternatively, you can employ Fill-mask pipeline .
from transformers import pipeline fill = pipeline("fill-mask", model="alabnii/jmedroberta-base-manbyo-wordpiece", top_k=10) fill("この患者は[MASK]と診断された。") #[{'score': 0.020739275962114334, # 'token': 11474, # 'token_str': 'ALS', # 'sequence': 'この 患者 は ALS と 診断 さ れ た 。'}, # {'score': 0.0193060003221035, # 'token': 10777, # 'token_str': '統合失調症', # 'sequence': 'この 患者 は 統合失調症 と 診断 さ れ た 。'}, # {'score': 0.014001614414155483, # 'token': 27318, # 'token_str': 'Fabry病', # 'sequence': 'この 患者 は Fabry病 と 診断 さ れ た 。'}, # ...
You can fine-tune this model on downstream tasks.
See also sample Colab notebooks: https://colab.research.google.com/drive/1yqUaqLf0Lf_imRT9TXPXEt1dowfK_2CS?usp=sharing
Mecab (w/ IPAdic & Manbyo Dictionary ) was used for pre-training. Each word is tokenized into tokens by WordPiece .
The vocabulary consists of 30000 tokens including words (IPAdic & Manbyo Dictionary ) and subwords induced by WordPiece .
The following hyperparameters were used during pre-training:
As the config file suggests, our model is based on HuggingFace's BertForMaskedLM class. However, we consider our model as RoBERTa for the following reasons:
This work was supported by Japan Japan Science and Technology Agency (JST) AIP Trilateral AI Research (Grant Number: JPMJCR20G9), and Joint Usage/Research Center for Interdisciplinary Large-scale Information Infrastructures (JHPCN) (Project ID: jh221004), in Japan. In this research work, we used the " mdx: a platform for the data-driven future ".