where ฮjโ(ฮธjtโ,zjtโ)=Ojโ(ฮธjtโ,zjtโ)โฮธjtโ, k0โ=j. Here Pj(ฯ)โ denotes the set of all sequences (k1โ,โฆ,kฯโ) such that ksโโNout(1)โ(ksโ1โ) for s=1,โฆ,ฯ and H(ฮธksโt+sโ;zksโt+sโ) is the Hessian matrix of L with respect to ฮธ evaluated at ฮธksโt+sโ and data zksโt+sโ.
For the cases when ฯ=0 and ฯ=1, the relevant product expressions are defined as identity matrices, thereby ensuring that the r-hop DICE-E remains well-defined.
Key Insights from DICE
Our theory uncovers the intricate interplay of factors that shape data influence in decentralized learning:
1. Asymmetric Influence and Topological Importance: The influence of identical data is not uniform across the network. Instead, nodes with greater topological significance exert stronger influence.
2. The Role of Intermediate Nodes and Loss Landscape: Intermediate nodes actively contribute to an "influence chain". The local loss landscape of these models also actively shapes the influence as it propagates through the network.
3. Influence Cascades with Damped Decay: Data influence cascades with "damped decay" induced by mixing parameter W. This decay, which can be exponential with the number of hops, ensures that influence is "localized".
Citation
Cite Our Paper ๐
If you find our work insightful, we would greatly appreciate it if you could cite our paper.
@inproceedings{zhu2025dice,
title="{DICE: Data Influence Cascade in Decentralized Learning}",
author="Tongtian Zhu and Wenhao Li and Can Wang and Fengxiang He",
booktitle="The Thirteenth International Conference on Learning Representations",
year="2025",
url="[https://openreview.net/forum?id=2TIYkqieKw](https://openreview.net/forum?id=2TIYkqieKw)"
}