Generalizability of Representation Learning on Downstream Tasks
| dc.contributor.author | Levinsson, Anton | |
| dc.contributor.author | Wang, Ziyuan | |
| dc.contributor.department | Chalmers tekniska högskola / Institutionen för data och informationsteknik | sv |
| dc.contributor.department | Chalmers University of Technology / Department of Computer Science and Engineering | en |
| dc.contributor.examiner | D. Johansson, Fredrik | |
| dc.contributor.supervisor | Balcioglu, Ahmet | |
| dc.date.accessioned | 2025-11-05T13:39:41Z | |
| dc.date.issued | 2025 | |
| dc.date.submitted | ||
| dc.description.abstract | We study the theoretical generalizability of representations learned by contrastive learning by analyzing their performance in downstream linear regression tasks. To quantify the quality of these learned features, we provide rigorous proofs for the worst-case and expected downstream performances under specified assumptions. These results offer theoretical results for analyzing the quality of features by looking at downstream tasks. To empirically verify the theory, we conduct experiments on both simulated and real-world data, following time-contrastive learning (TCL) strategies, which solve the problem of nonlinear independent component analysis (nonlinear ICA). In the real-world setting, we construct nonlinear source separation tasks using mixed audio signals from different instrument categories. Then, we propose some specific downstream tasks to verify our theories using our learned features from the TCL method and compare with the observed features. The results show that most of the defined downstream tasks are located in the confidence level of the expected performance and are bounded by the worst case, which indicates that our theory aligns with the real-world setting. | |
| dc.identifier.coursecode | DATX05 | |
| dc.identifier.uri | http://hdl.handle.net/20.500.12380/310726 | |
| dc.language.iso | eng | |
| dc.relation.ispartofseries | CSE 25-60 | |
| dc.setspec.uppsok | Technology | |
| dc.subject | contrastive learning, generalizability in downstream, nonliear ICA, linear regression. | |
| dc.title | Generalizability of Representation Learning on Downstream Tasks | |
| dc.type.degree | Examensarbete för masterexamen | sv |
| dc.type.degree | Master's Thesis | en |
| dc.type.uppsok | H | |
| local.programme | Computer science – algorithms, languages and logic (MPALG), MSc | |
| local.programme | Data science and AI (MPDSC), MSc |
