2023 journal article

Dynamic Network-Assisted D2D-Aided Coded Distributed Learning

IEEE TRANSACTIONS ON COMMUNICATIONS, 71(6), 3352–3367.

author keywords: Computational modeling; Device-to-device communication; Training; Data models; Collaboration; Distance learning; Computer aided instruction; Online distributed learning; device-to-device communications; Index Terms; coded computing; data compression; load balancing
TL;DR: This work designs a novel D2D-aided coded distributed learning method named D 2D-CDL for efficient load balancing across devices and derives an optimal compression rate, which minimizes the processing time. (via Semantic Scholar)
Source: Web Of Science
Added: July 31, 2023

Today, numerous machine learning (ML) applications offer continuous data processing and real-time data analytics at the edge of wireless networks. Distributed real-time ML solutions are highly susceptible to the so-called straggler effect caused by resource heterogeneity, which can be mitigated by various computation offloading mechanisms that severely impact communication efficiency, especially in large-scale scenarios. To reduce the communication overhead, we leverage device-to-device (D2D) connectivity, which enhances spectrum utilization and allows for efficient data exchange between proximate devices. In particular, we design a novel D2D-aided coded distributed learning method named D2D-CDL for efficient load balancing across devices. The proposed solution captures system dynamics, including data (time-varying learning model, irregular intensity of data arrivals), device (diverse computational resources and volume of training data), and deployment (different locations and D2D graph connectivity). To decrease the number of communication rounds, we derive an optimal compression rate, which minimizes the processing time. The resulting optimization problem provides suboptimal compression parameters that improve the total training time. Our proposed method is particularly beneficial for real-time collaborative applications, where users continuously generate training data.