针对计算大规模复杂网络时介数的空间和时间复杂度问题,根据网络数据的存储特点,设计了减少内存占用并能提高查找速度的数据结构.根据介数计算的特点,用Python语言设计了粗粒度并行算法,在多核心工作站机群实现了并行算法.实验结果表明:并行算法不仅能够适用于上亿条边规模的网络,而且能够获得线性加速比,使120个计算核心的加速比达到了71左右,为分析大规模复杂网络数据的特性提供了易操作的方案.
Focusing on the space and time complexity problems in computing betweenness centrality in large complex network, the data structure that can reduce the memory consumption and increase execution speed was brought forward based on the storage property of network data. A coarse grain parallel algorithm was designed using Python language according to the features of computing betweenness centrality. The parallel algorithm was realized in the cluster of multi-core workstations. The test results indicate that the algorithm can be applied in the analysis of large network with hundred millions of edges, and has linear speedup. The speedup can reach 71 when 120 cores are used, so as to provide an operational method for the analysis of large scale complex network data.