Using pre-trained language models for extractive text summarisation of academic papers

dc.contributor.authorHermansson, Erik
dc.contributor.authorBoddien, Charlottte
dc.contributor.departmentChalmers tekniska högskola / Institutionen för mekanik och maritima vetenskapersv
dc.contributor.examinerSelpi, Selpi
dc.contributor.supervisorSelpi, Selpi
dc.date.accessioned2020-03-02T12:28:38Z
dc.date.available2020-03-02T12:28:38Z
dc.date.issued2020sv
dc.date.submitted2019
dc.description.abstractGiven the overwhelming amount of textual information on the internet and elsewhere, the automatic creation of high-quality summaries is becoming increasingly important. With the development of neural networks and pre-trained models within natural language processing in recent years, such as BERT and its derivatives, the field of automatic text summarisation has seen a lot of progress. These models have been pre-trained on large amounts of data for general knowledge and are then finetuned for specific tasks. Datasets are a limiting factor for training summarisation models, as they require a large amount of manual summaries to be created. Most of the current summarisation models have been trained and evaluated using textual data mostly from the news domain. However, pre-trained models, fine-tuned on data from the news domain, could potentially also be able to generalize and perform well on other data as well. The main objective of this thesis is to investigate the suitability of several pre-trained language models for automatic text summarisation. The chosen models were finetuned on readily available news data, and evaluated on a very different dataset of academic texts to determine their ability to generalise. There were only slight differences between the models on the news data. But more interestingly, the results on the academic texts showed significant differences between the models. The results indicate that the more robustly pre-trained models are able to generalise better and according to the metrics perform quite well. However, human evaluation puts this into question, showing that even the high-scoring summaries did not necessarily read well. This highlights the need for better evaluation methods and metrics.sv
dc.identifier.coursecodeMMSX30sv
dc.identifier.urihttps://hdl.handle.net/20.500.12380/300719
dc.language.isoengsv
dc.relation.ispartofseries2020:01sv
dc.setspec.uppsokTechnology
dc.subjectnatural language processingsv
dc.subjectnlpsv
dc.subjectmachine learningsv
dc.subjectdeep learningsv
dc.subjectautomatic text summarisationsv
dc.subjectextractive summarisationsv
dc.subjecttransformersv
dc.subjectbertsv
dc.subjectrobertasv
dc.subjectxlnetsv
dc.titleUsing pre-trained language models for extractive text summarisation of academic paperssv
dc.type.degreeExamensarbete för masterexamensv
dc.type.uppsokH
local.programmeComputer systems and networks (MPCSN), MSc
Ladda ner
Original bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
2020-01 Erik Hermansson & Charlotte Boddien.pdf
Storlek:
1.28 MB
Format:
Adobe Portable Document Format
Beskrivning:
Master thesis
License bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
license.txt
Storlek:
1.14 KB
Format:
Item-specific license agreed upon to submission
Beskrivning: