Arrigo Francesca, Higham Desmond J
University of Strathclyde, 16 Richmond St, Glasgow, G1 1XQ UK.
Appl Netw Sci. 2017;2(1):17. doi: 10.1007/s41109-017-0038-z. Epub 2017 Jun 24.
Time sliced networks describing human-human digital interactions are typically large and sparse. This is the case, for example, with pairwise connectivity describing social media, voice call or physical proximity, when measured over seconds, minutes or hours. However, if we wish to quantify and compare the overall time-dependent centrality of the network nodes, then we should account for the global flow of information through time. Because the time-dependent edge structure typically allows information to diffuse widely around the network, a natural summary of sparse but dynamic pairwise interactions will generally take the form of a large dense matrix. For this reason, computing nodal centralities for a time-dependent network can be extremely expensive in terms of both computation and storage; much more so than for a single, static network. In this work, we focus on the case of dynamic communicability, which leads to broadcast and receive centrality measures. We derive a new algorithm for computing time-dependent centrality that works with a sparsified version of the dynamic communicability matrix. In this way, the computation and storage requirements are reduced to those of a sparse, static network at each time point. The new algorithm is justified from first principles and then tested on a large scale data set. We find that even with very stringent sparsity requirements (retaining no more than ten times the number of nonzeros in the individual time slices), the algorithm accurately reproduces the list of highly central nodes given by the underlying full system. This allows us to capture centrality over time with a minimal level of storage and with a cost that scales only linearly with the number of time points. We also describe and test three variants of the proposed algorithm that require fewer parameters and achieve a further reduction in the computational cost.
描述人与人之间数字交互的时间切片网络通常规模庞大且稀疏。例如,当按秒、分钟或小时进行测量时,描述社交媒体、语音通话或身体接近程度的成对连通性就是这种情况。然而,如果我们希望量化和比较网络节点随时间变化的整体中心性,那么我们就应该考虑信息随时间的全局流动。由于随时间变化的边结构通常允许信息在网络中广泛扩散,稀疏但动态的成对交互的自然汇总通常会采用大型密集矩阵的形式。因此,计算随时间变化的网络的节点中心性在计算和存储方面都可能极其昂贵;比单个静态网络要昂贵得多。在这项工作中,我们关注动态可达性的情况,这会引出广播和接收中心性度量。我们推导了一种用于计算随时间变化的中心性的新算法,该算法适用于动态可达性矩阵的稀疏版本。通过这种方式,计算和存储需求在每个时间点都降低到了稀疏静态网络的水平。新算法从第一原理出发得到了论证,然后在一个大规模数据集上进行了测试。我们发现,即使在非常严格的稀疏性要求下(每个时间切片中保留的非零元素数量不超过十倍),该算法也能准确重现底层完整系统给出的高度中心节点列表。这使我们能够以最小的存储量和仅与时间点数呈线性增长的成本来捕捉随时间变化的中心性。我们还描述并测试了所提出算法的三种变体,它们需要的参数更少,并且进一步降低了计算成本。