Abstract

Neural-based multi-task learning has been successfully used in many real-world large-scale applications such as recommendation systems. For example, in movie recommendations, beyond providing users movies which they tend to purchase and watch, the system might also optimize for users liking the movies afterwards. With multi-task learning, we aim to build a single model that learns these multiple goals and tasks simultaneously. However, the prediction quality of commonly used multi-task models is often sensitive to the relationships between tasks. It is therefore important to study the modeling tradeoffs between task-specific objectives and inter-task relationships. In this work, we propose a novel multi-task learning approach, Multi-gate Mixture-of-Experts (MMoE), which explicitly learns to model task relationships from data. We adapt the Mixture-of-Experts (MoE) structure to multi-task learning by sharing the expert submodels across all tasks, while also having a gating network trained to optimize each task. To validate our approach on data with different levels of task relatedness, we first apply it to a synthetic dataset where we control the task relatedness. We show that the proposed approach performs better than baseline methods when the tasks are less related. We also show that the MMoE structure results in an additional trainability benefit, depending on different levels of randomness in the training data and model initialization. Furthermore, we demonstrate the performance improvements by MMoE on real tasks including a binary classification benchmark, and a large-scale content recommendation system at Google.

Keywords

Computer scienceTask (project management)InitializationMachine learningArtificial intelligenceBenchmark (surveying)Multi-task learningTask analysisBaseline (sea)Artificial neural network

Affiliated Institutions

Related Publications

BERT4Rec

Fei Sun , Jun Liu , Jian Wu +4 more

Modeling users' dynamic preferences from their historical behaviors is challenging and crucial for recommendation systems. Previous methods employ sequential neural networks to ...

2019 Proceedings of the 28th ACM Internati... 1977 citations

Publication Info

Year
2018
Type
article
Pages
1930-1939
Citations
987
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

987
OpenAlex

Cite This

Jiaqi Ma, Zhe Zhao, Xinyang Yi et al. (2018). Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts. , 1930-1939. https://doi.org/10.1145/3219819.3220007

Identifiers

DOI
10.1145/3219819.3220007