On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models

Multi-Task Learning is today an interesting and promising field which many mention as a must for achieving the next level advancement within machine learning. However, in reality, Multi-Task Learning is much more rarely used in real-world implementations than its more popular cousin Transfer Learnin...

Full description

Bibliographic Details
Main Author: Tovedal, Sofiea
Format: Others
Language:English
Published: Umeå universitet, Institutionen för datavetenskap 2020
Subjects:
Online Access:http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172257
id ndltd-UPSALLA1-oai-DiVA.org-umu-172257
record_format oai_dc
spelling ndltd-UPSALLA1-oai-DiVA.org-umu-1722572020-06-18T03:40:30ZOn The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning modelsengTovedal, SofieaUmeå universitet, Institutionen för datavetenskap2020Multi-Task LearningNatural Language ProcessingSupervised LearningGated Recurrent UnitLong Short Term MemoryEngineering and TechnologyTeknik och teknologierMulti-Task Learning is today an interesting and promising field which many mention as a must for achieving the next level advancement within machine learning. However, in reality, Multi-Task Learning is much more rarely used in real-world implementations than its more popular cousin Transfer Learning. The questionis why that is and if Multi-Task Learning outperforms its Single-Task counterparts. In this thesis different Multi-Task Learning architectures were utilized in order to build a model that can handle labeling real technical issues within two categories. The model faces a challenging imbalanced data set with many labels to choose from and short texts to base its predictions on. Can task-sharing be the answer to these problems? This thesis investigated three Multi-Task Learning architectures and compared their performance to a Single-Task model. An authentic data set and two labeling tasks was used in training the models with the method of supervised learning. The four model architectures; Single-Task, Multi-Task, Cross-Stitched and the Shared-Private, first went through a hyper parameter tuning process using one of the two layer options LSTM and GRU. They were then boosted by auxiliary tasks and finally evaluated against each other. Student thesisinfo:eu-repo/semantics/bachelorThesistexthttp://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172257UMNAD ; 1226application/pdfinfo:eu-repo/semantics/openAccess
collection NDLTD
language English
format Others
sources NDLTD
topic Multi-Task Learning
Natural Language Processing
Supervised Learning
Gated Recurrent Unit
Long Short Term Memory
Engineering and Technology
Teknik och teknologier
spellingShingle Multi-Task Learning
Natural Language Processing
Supervised Learning
Gated Recurrent Unit
Long Short Term Memory
Engineering and Technology
Teknik och teknologier
Tovedal, Sofiea
On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models
description Multi-Task Learning is today an interesting and promising field which many mention as a must for achieving the next level advancement within machine learning. However, in reality, Multi-Task Learning is much more rarely used in real-world implementations than its more popular cousin Transfer Learning. The questionis why that is and if Multi-Task Learning outperforms its Single-Task counterparts. In this thesis different Multi-Task Learning architectures were utilized in order to build a model that can handle labeling real technical issues within two categories. The model faces a challenging imbalanced data set with many labels to choose from and short texts to base its predictions on. Can task-sharing be the answer to these problems? This thesis investigated three Multi-Task Learning architectures and compared their performance to a Single-Task model. An authentic data set and two labeling tasks was used in training the models with the method of supervised learning. The four model architectures; Single-Task, Multi-Task, Cross-Stitched and the Shared-Private, first went through a hyper parameter tuning process using one of the two layer options LSTM and GRU. They were then boosted by auxiliary tasks and finally evaluated against each other.
author Tovedal, Sofiea
author_facet Tovedal, Sofiea
author_sort Tovedal, Sofiea
title On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models
title_short On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models
title_full On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models
title_fullStr On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models
title_full_unstemmed On The Effectiveness of Multi-TaskLearningAn evaluation of Multi-Task Learning techniques in deep learning models
title_sort on the effectiveness of multi-tasklearningan evaluation of multi-task learning techniques in deep learning models
publisher Umeå universitet, Institutionen för datavetenskap
publishDate 2020
url http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-172257
work_keys_str_mv AT tovedalsofiea ontheeffectivenessofmultitasklearninganevaluationofmultitasklearningtechniquesindeeplearningmodels
_version_ 1719321808051634176