Shuguang Cui: Communication Efficient Federated Learning for Wireless Networks
Communication Efficient Federated Learning for Wireless Networks
Buch
lieferbar innerhalb 2-3 Wochen
(soweit verfügbar beim Lieferanten)
(soweit verfügbar beim Lieferanten)
EUR 164,28*
Verlängerter Rückgabezeitraum bis 31. Januar 2025
Alle zur Rückgabe berechtigten Produkte, die zwischen dem 1. bis 31. Dezember 2024 gekauft wurden, können bis zum 31. Januar 2025 zurückgegeben werden.
- Springer Nature Switzerland, 02/2024
- Einband: Gebunden, HC runder Rücken kaschiert
- Sprache: Englisch
- ISBN-13: 9783031512650
- Bestellnummer: 11769144
- Umfang: 192 Seiten
- Nummer der Auflage: 24001
- Auflage: 1st ed. 2024
- Gewicht: 459 g
- Maße: 241 x 160 mm
- Stärke: 17 mm
- Erscheinungstermin: 20.2.2024
- Serie: Wireless Networks
Achtung: Artikel ist nicht in deutscher Sprache!
Klappentext
This book provides a comprehensive study of Federated Learning (FL) over wireless networks. It consists of three main parts: (a) Fundamentals and preliminaries of FL, (b) analysis and optimization of FL over wireless networks, and (c) applications of wireless FL for Internet-of-Things systems. In particular, in the first part, the authors provide a detailed overview on widely-studied FL framework. In the second part of this book, the authors comprehensively discuss three key wireless techniques including wireless resource management, quantization, and over-the-air computation to support the deployment of FL over realistic wireless networks. It also presents several solutions based on optimization theory, graph theory and machine learning to optimize the performance of FL over wireless networks. In the third part of this book, the authors introduce the use of wireless FL algorithms for autonomous vehicle control and mobile edge computing optimization.Machine learning and data-driven approaches have recently received considerable attention as key enablers for next-generation intelligent networks. Currently, most existing learning solutions for wireless networks rely on centralizing the training and inference processes by uploading data generated at edge devices to data centers. However, such a centralized paradigm may lead to privacy leakage, violate the latency constraints of mobile applications, or may be infeasible due to limited bandwidth or power constraints of edge devices. To address these issues, distributing machine learning at the network edge provides a promising solution, where edge devices collaboratively train a shared model using real-time generated mobile data. The avoidance of data uploading to a central server not only helps preserve privacy but also reduces network traffic congestion as well as communication cost. Federated learning (FL) is one of most important distributed learning algorithms. In particular, FL enables devices to train a shared machine learning model while keeping data locally. However, in FL, training machine learning models requires communication between wireless devices and edge servers over wireless links. Therefore, wireless impairments such as noise, interference, and uncertainties among wireless channel states will significantly affect the training process and performance of FL. For example, transmission delay can significantly impact the convergence time of FL algorithms. In consequence, it is necessary to optimize wireless network performance for the implementation of FL algorithms.
This book targets researchers and advanced level students in computer science and electrical engineering. Professionals working in signal processing and machine learning will also buy this book.