Ensemble model of Bidirectional Encoder Representation from Transformers for Named Entity Recognition

dc.contributor.authorJendle, Carl
dc.contributor.authorSchönbeck, Linus
dc.contributor.departmentChalmers tekniska högskola / Institutionen för data och informationstekniksv
dc.contributor.examinerAxelson-Fisk, Marina
dc.contributor.supervisorBrown-Cohen, Jonah
dc.date.accessioned2021-08-20T12:49:26Z
dc.date.available2021-08-20T12:49:26Z
dc.date.issued2021sv
dc.date.submitted2020
dc.description.abstractNamed entity recognition (NER) has been widely modeled using Bidirectional En coder Representations from Transformers (BERT) in state of the art implementations since its appearance in 2018. Various configurations based on BERT models currently hold 4 out of 5 top positions on the GLUE leaderboard, an acknowledged benchmark for natural language processing and understanding. Relying on BERT architecture, a range of NER model designs were investigated to predict entities in a comparatively small set of medical press releases. The performance of all investigated model designs proved to be boosted with transfer learning using the publicly available datasets Conll2003 and BC5CDR early on in the project. Transfer learning was therefore implemented in the best named entity recognition system found, the separate submodel system under Section 6.3.6. This final design consisted of two submodels, each classifying different entity subsets independently. The Conll and BC5CDR datasets were used for transfer learning in the respective submodels prior to the introduction of medical press release data. The separate submodel system reached an F1-score of 0.79 (Conll model) and 0.78 (BC5CDR model). The effect of pre-training a selection of publicly available BERT models on the medical press releases was also investigated, but was given less emphasis due to insufficient amounts of data.sv
dc.identifier.coursecodeMPDSCsv
dc.identifier.urihttps://hdl.handle.net/20.500.12380/303941
dc.language.isoengsv
dc.setspec.uppsokTechnology
dc.subjectTransfer learningsv
dc.subjectnatural language processingsv
dc.subjectnamed entity recognitionsv
dc.subjectBERTsv
dc.subjectconditional random fieldsv
dc.titleEnsemble model of Bidirectional Encoder Representation from Transformers for Named Entity Recognitionsv
dc.type.degreeExamensarbete för masterexamensv
dc.type.uppsokH
Ladda ner
Original bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
CSE 21-54 Jendle Schönbeck.pdf
Storlek:
1.96 MB
Format:
Adobe Portable Document Format
Beskrivning:
License bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
license.txt
Storlek:
1.51 KB
Format:
Item-specific license agreed upon to submission
Beskrivning: