Abstract | ||
---|---|---|
Classical mathematical models of information sharing and updating in multi-agent networks use linear operators. In the paradigmatic DeGroot model, agents update their states with linear combinations of their neighbors' current states. In prior work, an accelerated averaging model employing the use of memory has been suggested to accelerate convergence to a consensus state for undirected networks. There, the DeGroot update on the current states is followed by a linear combination with the previous states. We propose a modification where the DeGroot update is applied to the current and previous states and is then followed by a linear combination step. We show that this simple modification applied to undirected networks permits convergence even for periodic networks. Further, it allows for faster convergence than the DeGroot and accelerated averaging models for suitable networks and model parameters. |
Year | DOI | Venue |
---|---|---|
2021 | 10.1109/CDC45484.2021.9683055 | CDC |
DocType | Citations | PageRank |
Conference | 0 | 0.34 |
References | Authors | |
0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Aditya Bhaskar | 1 | 0 | 0.34 |
Shriya Rangarajan | 2 | 0 | 0.34 |
Vikram Shree | 3 | 0 | 2.37 |
Mark Campbell | 4 | 2 | 1.40 |
Francesca Parise | 5 | 4 | 1.46 |