[Submitted on 2 Jan 2025]
View a PDF of the paper titled A Multi-task Supervised Compression Model for Split Computing, by Yoshitomo Matsubara and 2 other authors
Abstract:Split computing ($neq$ split learning) is a promising approach to deep learning models for resource-constrained edge computing systems, where weak sensor (mobile) devices are wirelessly connected to stronger edge servers through channels with limited communication capacity. State-of-theart work on split computing presents methods for single tasks such as image classification, object detection, or semantic segmentation. The application of existing methods to multitask problems degrades model accuracy and/or significantly increase runtime latency. In this study, we propose Ladon, the first multi-task-head supervised compression model for multi-task split computing. Experimental results show that the multi-task supervised compression model either outperformed or rivaled strong lightweight baseline models in terms of predictive performance for ILSVRC 2012, COCO 2017, and PASCAL VOC 2012 datasets while learning compressed representations at its early layers. Furthermore, our models reduced end-to-end latency (by up to 95.4%) and energy consumption of mobile devices (by up to 88.2%) in multi-task split computing scenarios.
Submission history
From: Yoshitomo Matsubara [view email]
[v1]
Thu, 2 Jan 2025 18:59:05 UTC (4,098 KB)
Source link
lol