On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models
Multi-Task Learning is today an interesting and promising field which many mention as a must for achieving the next level advancement within machine learning. However, in reality, Multi-Task Learning is much more rarely used in real-world implementations than its more popular cousin Transfer Learnin...
Main Author: | |
---|---|
Format: | Others |
Language: | English |
Published: |
Umeå universitet, Institutionen för datavetenskap
2020
|
Subjects: | |
Online Access: | http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172257 |
id |
ndltd-UPSALLA1-oai-DiVA.org-umu-172257 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-UPSALLA1-oai-DiVA.org-umu-1722572020-06-18T03:40:30ZOn The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning modelsengTovedal, SofieaUmeå universitet, Institutionen för datavetenskap2020Multi-Task LearningNatural Language ProcessingSupervised LearningGated Recurrent UnitLong Short Term MemoryEngineering and TechnologyTeknik och teknologierMulti-Task Learning is today an interesting and promising field which many mention as a must for achieving the next level advancement within machine learning. However, in reality, Multi-Task Learning is much more rarely used in real-world implementations than its more popular cousin Transfer Learning. The questionis why that is and if Multi-Task Learning outperforms its Single-Task counterparts. In this thesis different Multi-Task Learning architectures were utilized in order to build a model that can handle labeling real technical issues within two categories. The model faces a challenging imbalanced data set with many labels to choose from and short texts to base its predictions on. Can task-sharing be the answer to these problems? This thesis investigated three Multi-Task Learning architectures and compared their performance to a Single-Task model. An authentic data set and two labeling tasks was used in training the models with the method of supervised learning. The four model architectures; Single-Task, Multi-Task, Cross-Stitched and the Shared-Private, first went through a hyper parameter tuning process using one of the two layer options LSTM and GRU. They were then boosted by auxiliary tasks and finally evaluated against each other. Student thesisinfo:eu-repo/semantics/bachelorThesistexthttp://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172257UMNAD ; 1226application/pdfinfo:eu-repo/semantics/openAccess |
collection |
NDLTD |
language |
English |
format |
Others
|
sources |
NDLTD |
topic |
Multi-Task Learning Natural Language Processing Supervised Learning Gated Recurrent Unit Long Short Term Memory Engineering and Technology Teknik och teknologier |
spellingShingle |
Multi-Task Learning Natural Language Processing Supervised Learning Gated Recurrent Unit Long Short Term Memory Engineering and Technology Teknik och teknologier Tovedal, Sofiea On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models |
description |
Multi-Task Learning is today an interesting and promising field which many mention as a must for achieving the next level advancement within machine learning. However, in reality, Multi-Task Learning is much more rarely used in real-world implementations than its more popular cousin Transfer Learning. The questionis why that is and if Multi-Task Learning outperforms its Single-Task counterparts. In this thesis different Multi-Task Learning architectures were utilized in order to build a model that can handle labeling real technical issues within two categories. The model faces a challenging imbalanced data set with many labels to choose from and short texts to base its predictions on. Can task-sharing be the answer to these problems? This thesis investigated three Multi-Task Learning architectures and compared their performance to a Single-Task model. An authentic data set and two labeling tasks was used in training the models with the method of supervised learning. The four model architectures; Single-Task, Multi-Task, Cross-Stitched and the Shared-Private, first went through a hyper parameter tuning process using one of the two layer options LSTM and GRU. They were then boosted by auxiliary tasks and finally evaluated against each other. |
author |
Tovedal, Sofiea |
author_facet |
Tovedal, Sofiea |
author_sort |
Tovedal, Sofiea |
title |
On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models |
title_short |
On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models |
title_full |
On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models |
title_fullStr |
On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models |
title_full_unstemmed |
On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models |
title_sort |
on the effectiveness of multi-tasklearningan evaluation of multi-task learning techniques in deep learning models |
publisher |
Umeå universitet, Institutionen för datavetenskap |
publishDate |
2020 |
url |
http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172257 |
work_keys_str_mv |
AT tovedalsofiea ontheeffectivenessofmultitasklearninganevaluationofmultitasklearningtechniquesindeeplearningmodels |
_version_ |
1719321808051634176 |