Supervised Learning with Dynamic Network Architectures

Date

Type

Examensarbete för masterexamen

Programme

Model builders

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

There exists many different techniques for training neural networks, but few are designed to handle the case of lifelong learning. Most models are based on the assumption that there is a training phase with a finite amount of data. This thesis investigates and evaluates a brand new algorithm, namely the Lifelong Learning starting from zero (LL0) which can be used for lifelong learning where there is a continuous stream of data. The algorithm stems from biology, logical rules, and machine learning. The algorithm builds a dynamic artificial neural network architecture over time, based on four different concepts; extension, generalization, forgetting and backpropagation. These first three concepts all have their origin in biology, and can be found in animals and humans in the form of neuroplasticity. The model is evaluated and benchmarked against five other models on different datasets and problems. The obtained results act as a proof of concept for the algorithm. Lastly, the pros and cons of the model are discussed, followed by a discussion on future work. The model proposed outperforms all models on chosen benchmarks, primarily in the area of one-shot learning. The model manages to achieve about 90% accuracy on unseen data after only training on a small portion of the training set on multiple datasets. LL0 also shows promising results in the area of lifelong learning.

Description

Keywords

Machine Learning, Neural Networks, Dynamic Architectures, Supervised Learning, Lifelong Learning, One-Shot Learning, Transfer Learning, Computer Science

Citation

Architect

Location

Type of building

Build Year

Model type

Scale

Material / technology

Index

Collections

Endorsement

Review

Supplemented By

Referenced By