Re-train or Train from Scratch? Comparing Pre-training Strategies of BERT in the Medical Domain - Archive ouverte HAL Access content directly
Conference Poster Year : 2022

Re-train or Train from Scratch? Comparing Pre-training Strategies of BERT in the Medical Domain

(1, 2) , (3) , (1, 2) , (1, 2)
1
2
3

Abstract

BERT models used in specialized domains all seem to be the result of a simple strategy: initializing with the original BERT then resuming pre-training on a specialized corpus. This method yields rather good performance (e.g. BioBERT (Lee et al., 2020), SciBERT (Beltagy et al., 2019), BlueBERT (Peng et al., 2019)). However, it seems reasonable to think that training directly on a specialized corpus, using a specialized vocabulary, could result in more tailored embeddings and thus help performance. To test this hypothesis, we train BERT models from scratch using many configurations involving general and medical corpora. Based on evaluations using four different tasks, we find that the initial corpus only has a weak influence on the performance of BERT models when these are further pre-trained on a medical corpus.
Fichier principal
Vignette du fichier
2022.lrec-1.281.pdf (791.22 Ko) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

cea-03745341 , version 1 (04-08-2022)

Identifiers

  • HAL Id : cea-03745341 , version 1

Cite

Hicham El Boukkouri, Olivier Ferret, Thomas Lavergne, Pierre Zweigenbaum. Re-train or Train from Scratch? Comparing Pre-training Strategies of BERT in the Medical Domain. LREC 2022, Jun 2022, Marseille, France. pp.2626-2633, 2022, LREC 2022. ⟨cea-03745341⟩
67 View
36 Download

Share

Gmail Facebook Twitter LinkedIn More