Presentation + Paper
7 June 2024 BSMO: enhancing multitask learning through batch swapping optimization
Zephaniah Spencer, Gunar Schirner, Hamed Tabkhi
Author Affiliations +
Abstract
With the advent of deep learning, there has been an ever-growing list of applications to which Deep Convolutional Neural Networks (DCNNs) can be applied. The field of Multi-Task Learning (MTL) attempts to provide optimizations to many-task systems, improving performance by optimization algorithms and structural changes to these networks. However, we have found that current MTL optimization algorithms often impose burdensome computation overheads, require meticulously labeled datasets, and do not adapt to tasks with significantly different loss distributions. We propose a new MTL optimization algorithm: Batch Swapping with Multiple Optimizers (BSMO). We utilize single-task labeled data to train on a multi-task hard parameter sharing (HPS) network through swapping tasks at the batch level. This dramatically increases the flexibility and scalability of training on an HPS network by allowing for per-task datasets and augmentation pipelines. We demonstrate the efficacy of BSMO versus current SOTA algorithms by benchmarking across contemporary benchmarks & networks.
Conference Presentation
(2024) Published by SPIE. Downloading of the abstract is permitted for personal use only.
Zephaniah Spencer, Gunar Schirner, and Hamed Tabkhi "BSMO: enhancing multitask learning through batch swapping optimization", Proc. SPIE 13034, Real-Time Image Processing and Deep Learning 2024, 1303409 (7 June 2024); https://doi.org/10.1117/12.3013609
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Mathematical optimization

Artificial intelligence

Computer vision technology

Convolutional neural networks

Back to Top