Network Support for High-Performance Distributed Machine Learning
نویسندگان
چکیده
The traditional approach to distributed machine learning is adapt algorithms the network, e.g., reducing updates curb overhead. Networks based on intelligent edge, instead, make it possible follow opposite approach, i.e., define logical network topology around task perform, so as meet desired performance. In this paper, we propose a system model that captures such aspects in context of supervised learning, accounting for both nodes (that perform computations) and information provide data). We then formulate problem selecting (i) which should cooperate complete task, (ii) number epochs run, order minimize cost while meeting target prediction error execution time. After proving important properties above problem, devise an algorithm, named DoubleClimb, can find $1+1/| \mathcal {I}|$ -competitive solution (with notation="LaTeX">$\mathcal {I}$ being set nodes), with cubic worst-case complexity. Our performance evaluation, leveraging real-world considering classification regression tasks, also shows DoubleClimb closely matches optimum, outperforming state-of-the-art alternatives.
منابع مشابه
Optimizing Network Performance in Distributed Machine Learning
To cope with the ever growing availability of training data, there have been several proposals to scale machine learning computation beyond a single server and distribute it across a cluster. While this enables reducing the training time, the observed speed up is often limited by network bottlenecks. To address this, we design MLNET, a host-based communication layer that aims to improve the net...
متن کاملLitz: An Elastic Framework for High-Performance Distributed Machine Learning
Machine Learning (ML) is becoming an increasingly popular application in the cloud and data-centers, inspiring a growing number of distributed frameworks optimized for it. These frameworks leverage the specific properties of ML algorithms to achieve orders of magnitude performance improvements over generic data processing frameworks like Hadoop or Spark. However, they also tend to be static, un...
متن کاملDistributed Extreme Learning Machine for Nonlinear Learning over a Network
Distributed data collection and analysis over a network are ubiquitous, especially over a wireless sensor network (WSN). To our knowledge, the data model used in most of the distributed algorithms is linear. However, in real applications, the linearity of systems is not always guaranteed. In nonlinear cases, the single hidden layer feedforward neural network (SLFN) with radial basis function (R...
متن کاملDistributed Extreme Learning Machine for Nonlinear Learning over Network
Distributed data collection and analysis over a network are ubiquitous, especially over a wireless sensor network (WSN). To our knowledge, the data model used in most of the distributed algorithms is linear. However, in real applications, the linearity of systems is not always guaranteed. In nonlinear cases, the single hidden layer feedforward neural network (SLFN) with radial basis function (R...
متن کاملUsing Support Vector Machines for Distributed Machine Learning
In this thesis we investigate the potential use of support vector machines (SVMs) for distributed machine learning. An SVM is an algorithm out of the machine learning field, which can be used for classification, regression, and other important tasks. The novel approach in this thesis is to apply the SVMs as co-active learning units while respecting the distributed setting of the problem. We con...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE ACM Transactions on Networking
سال: 2023
ISSN: ['1063-6692', '1558-2566']
DOI: https://doi.org/10.1109/tnet.2022.3189077