Generating subtitles with controllable length using natural language processing

Examensarbete för masterexamen

Please use this identifier to cite or link to this item:
Download file(s):
File Description SizeFormat 
CSE 22-46 Svensson Troksch.pdf2.1 MBAdobe PDFView/Open
Bibliographical item details
Type: Examensarbete för masterexamen
Title: Generating subtitles with controllable length using natural language processing
Authors: Svensson, Joakim
Troksch, Victor
Abstract: Creating subtitles for video content is a task that has traditionally been performed manually by subtitlers. When creating a subtitle, there are rules and guidelines for how the text should be presented to the viewer. Therefore, a subtitle, translated from one language to another, often contains linguistic compression in the form of paraphrasing or removing parts of the dialogues. With advances in natural language processing, subtitlers now have tools for machine translation and automated speech recognition to assist them in their work. This thesis aims to explore various methods for how to control the generated output length of a sequence-to-sequence model, which are typically used for text generation and therefore also for machine translation. We apply different modifications to both the model itself and the data to control the output. Furthermore, this project makes use of transfer learning and pre-trained models with the Transformer architecture. The length ratio method produced the best results, in which it was possible to effectively control the output length of a generated subtitle. We also discover that it was also possible to apply this method for a translation model. Although it is a relatively simple method, it produced the desired results with linguistic correctness.
Keywords: Natural Language Processing;NLP;Transformer;seq2seq;text generation;BART;subtitles
Issue Date: 2022
Publisher: Chalmers tekniska högskola / Institutionen för data och informationsteknik
Collection:Examensarbeten för masterexamen // Master Theses

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.