OpenMethods introduction to: BERT for Humanists: a deep learning language model meets DH

Fiche du document

Type de document
Périmètre
Langue
Identifiants
  • handle:  10670/1.kq40md
  • https://openmethods.dariah.eu/2021/11/09/bert-for-humanists-a-deep-learning-language-model-meets-dh/
Organisation

DARIAH




Citer ce document

Marinella Testori, « OpenMethods introduction to: BERT for Humanists: a deep learning language model meets DH », OpenMethods: Highlighting Digital Humanities Methods and Tools, ID : 10670/1.kq40md


Métriques


Partage / Export

Résumé 0

Introduction: Awarded as Best Long Paper at the 2019 NACCL (North American Chapter of the Association for Computational Linguistics) Conference, the contribution by Jacob Devlin et al. provides an illustration of “BERT: Pre-training of Deep Biredictional Transformers for Language Understanding” (https://aclanthology.org/N19-1423/). As highlighted by the authors in the abstract, BERT is a “new language representation model” and, in the past few years, it has become widespread in various NLP applications; for example, a project exploiting it is CamemBERT (https://camembert-model.fr/), regarding French.  In June 2021, a workshop organized by David Mimno, Melanie Walsh and Maria Antoniak (https://melaniewalsh.github.io/BERT-for-Humanists/workshop/) pointed out how to use BERT in projects related to digital humanities, in order to deal with word similarity and classification classification while relying on Phyton-based HuggingFace transformers library. (https://melaniewalsh.github.io/BERT-for-Humanists/tutorials/ ). A further advantage of this training resource is that it has been written with sensitivity towards the target audience in mind:  in a way that it provides a gentle introduction to complexities of language models to scholars with education and background other than Computer Science. Along with the Tutorials, the same blog includes Introductions about BERT in general and in its specific usage in a Google Colab notebook, as well as a constantly-updated bibliography and a glossary of the main terms ('attention', 'Fine-Tune', 'GPU', 'Label', 'Task', 'Transformers', 'Token', 'Type', 'Vector').

document thumbnail

Par les mêmes auteurs

Sur les mêmes sujets

Sur les mêmes disciplines

Exporter en