We consider learning in decentralized heterogeneous networks: agents seek to minimize a convex functional that aggregates data across the network, while only having access to their local data streams. We focus on the ...
详细信息
We consider learning in decentralized heterogeneous networks: agents seek to minimize a convex functional that aggregates data across the network, while only having access to their local data streams. We focus on the case where agents seek to estimate a regression function that belongs to a reproducing kernel Hilbert space (RKHS). To incentivize coordination while respecting network heterogeneity, we impose nonlinear proximity constraints. The resulting constrained stochastic optimization problem is solved using the functional variant of stochasticprimal-dual (Arrow-Hurwicz) method which yields a decentralized algorithm. In order to avoid the model complexity from growing linearly over time, we project the primal iterates onto subspaces greedily constructed from kernel evaluations of agents' local observations. The resulting scheme, dubbed Heterogeneous Adaptive Learning with Kernels (HALK), allows us, for the first time, to characterize the precise trade-off between the optimality gap, constraint violation, and the model complexity. In particular, the proposed algorithm can be tuned to achieve zero constraint violation, an optimality gap of O(T-1/2 + alpha) after T iterations, where the number of elements retained in the dictionary is determined by 1/alpha. Simulations on a correlated spatio-temporal field estimation validate our theoretical results, which are corroborated in practice for networked oceanic sensing buoys estimating temperature and salinity from depth measurements.
暂无评论