Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

A Pre-Training Technique to Localize Medical BERT and to Enhance Biomedical BERT

View through CrossRef
Abstract Background: Pre-training large-scale neural language models on raw texts has been shown to make a significant contribution to a strategy for transfer learning in natural language processing (NLP). With the introduction of transformer-based language models, such as Bidirectional Encoder Representations from Transformers (BERT), the performance of information extraction from free text by NLP has significantly improved for both the general domain and the medical domain; however, it is difficult for languages in which there are few publicly available medical databases with a high quality and a large size to train medical BERT models that perform well.Method: We introduce a method to train a BERT model using a small medical corpus both in English and in Japanese. Our proposed method consists of two interventions: simultaneous pre-training, which is intended to encourage masked language modeling and next-sentence prediction on the small medical corpus, and amplified vocabulary, which helps with suiting the small corpus when building the customized corpus by byte-pair encoding. Moreover, we used whole PubMed abstracts and developed a high-performance BERT model, Bidirectional Encoder Representations from Transformers for Biomedical Text Mining by Osaka University (ouBioBERT), in English via our method. We then evaluated the performance of our BERT models and publicly available baselines and compared them.Results: We confirmed that our Japanese medical BERT outperforms conventional baselines and the other BERT models in terms of the medical-document classification task and that our English BERT pre-trained using both the general and medical domain corpora performs sufficiently for practical use in terms of the biomedical language understanding evaluation (BLUE) benchmark. Moreover, ouBioBERT shows that the total score of the BLUE benchmark is 1.1 points above that of BioBERT and 0.3 points above that of the ablation model trained without our proposed method.Conclusions: Our proposed method makes it feasible to construct a practical medical BERT model in both Japanese and English, and it has a potential to produce higher performing models for biomedical shared tasks.
Title: A Pre-Training Technique to Localize Medical BERT and to Enhance Biomedical BERT
Description:
Abstract Background: Pre-training large-scale neural language models on raw texts has been shown to make a significant contribution to a strategy for transfer learning in natural language processing (NLP).
With the introduction of transformer-based language models, such as Bidirectional Encoder Representations from Transformers (BERT), the performance of information extraction from free text by NLP has significantly improved for both the general domain and the medical domain; however, it is difficult for languages in which there are few publicly available medical databases with a high quality and a large size to train medical BERT models that perform well.
Method: We introduce a method to train a BERT model using a small medical corpus both in English and in Japanese.
Our proposed method consists of two interventions: simultaneous pre-training, which is intended to encourage masked language modeling and next-sentence prediction on the small medical corpus, and amplified vocabulary, which helps with suiting the small corpus when building the customized corpus by byte-pair encoding.
Moreover, we used whole PubMed abstracts and developed a high-performance BERT model, Bidirectional Encoder Representations from Transformers for Biomedical Text Mining by Osaka University (ouBioBERT), in English via our method.
We then evaluated the performance of our BERT models and publicly available baselines and compared them.
Results: We confirmed that our Japanese medical BERT outperforms conventional baselines and the other BERT models in terms of the medical-document classification task and that our English BERT pre-trained using both the general and medical domain corpora performs sufficiently for practical use in terms of the biomedical language understanding evaluation (BLUE) benchmark.
Moreover, ouBioBERT shows that the total score of the BLUE benchmark is 1.
1 points above that of BioBERT and 0.
3 points above that of the ablation model trained without our proposed method.
Conclusions: Our proposed method makes it feasible to construct a practical medical BERT model in both Japanese and English, and it has a potential to produce higher performing models for biomedical shared tasks.

Related Results

Big Data as the Foundation of a Novel Training Platform for Biomedical Researchers in Qatar
Big Data as the Foundation of a Novel Training Platform for Biomedical Researchers in Qatar
BackgroundTechnological breakthroughs witnessed over the past decade have led to an explosive increase in molecular profiling capabilities. This has ushered a new “data-rich era” f...
Pelatihan Peramalan Target Indikator Kinerja Daerah
Pelatihan Peramalan Target Indikator Kinerja Daerah
The purpose of this community service in the form of training is to improve the ability of functional planner staff in forecasting indicator targets of regional performance. This t...
Advancements in Biomedical and Bioinformatics Engineering
Advancements in Biomedical and Bioinformatics Engineering
Abstract: The field of biomedical and bioinformatics engineering is witnessing rapid advancements that are revolutionizing healthcare and medical research. This chapter provides a...
TOO-BERT: A Trajectory Order Objective BERT for self-supervised representation learning of temporal healthcare data
TOO-BERT: A Trajectory Order Objective BERT for self-supervised representation learning of temporal healthcare data
Abstract The growing availability of Electronic Health Records (EHRs) presents an opportunity to enhance patient care by uncovering hidden health risks and improving inform...
Biomedical Engineering International joins the Family of Platinum Open Access Journals
Biomedical Engineering International joins the Family of Platinum Open Access Journals
We are delightfully announcing the launch of Biomedical Engineering International, a new interdisciplinary international scholarly open-access journal dedicated to publishing origi...
Semantic Similarity Caculating based on BERT
Semantic Similarity Caculating based on BERT
The exploration of semantic similarity is a fundamental aspect of natural language processing, as it aids in comprehending the significance and usage of vocabulary present in a lan...
МЕТОД СЕМАНТИЧНОГО АНАЛІЗУ ДАНИХ ДЛЯ ВИЗНАЧЕННЯ МАРКЕРНИХ СЛІВ ПРИ ОБРОБЛЕННІ РЕЗУЛЬТАТІВ ОЦІНКИ ВІЗИТОРІВ В ІНТЕРАКТИВНОМУ МИСТЕЦТВІ
МЕТОД СЕМАНТИЧНОГО АНАЛІЗУ ДАНИХ ДЛЯ ВИЗНАЧЕННЯ МАРКЕРНИХ СЛІВ ПРИ ОБРОБЛЕННІ РЕЗУЛЬТАТІВ ОЦІНКИ ВІЗИТОРІВ В ІНТЕРАКТИВНОМУ МИСТЕЦТВІ
Предметом дослідження є поглиблений семантичний аналіз даних, що базується на інтеграції методологій латентного розподілу Діріхле (LDA) та двонаправленого кодувального представленн...

Back to Top