Logo des Repositoriums

Scaling up network centrality computations – A brief overview

Vorschaubild nicht verfügbar

Volltext URI


Text/Journal Article





ISSN der Zeitschrift



De Gruyter


Network science methodology is increasingly applied to a large variety of real-world phenomena, often leading to big network data sets. Thus, networks (or graphs) with millions or billions of edges are more and more common. To process and analyze these data, we need appropriate graph processing systems and fast algorithms. Yet, many analysis algorithms were pioneered on small networks when speed was not the highest concern. Developing an analysis toolkit for large-scale networks thus often requires faster variants, both from an algorithmic and an implementation perspective. In this paper we focus on computational aspects of vertex centrality measures. Such measures indicate the (relative) importance of a vertex based on the position of the vertex in the network. We describe several common (and some recent and thus less established) measures, optimization problems in their context as well as algorithms for an efficient solution of the raised problems. Our focus is on (not necessarily exact) performance-oriented algorithmic techniques that enable significantly faster processing than the previous state of the art – often allowing to process massive data sets quickly and without resorting to distributed graph processing systems.


Grinten, Alexander van der; Angriman, Eugenio; Meyerhenke, Henning (2020): Scaling up network centrality computations – A brief overview. it - Information Technology: Vol. 62, No. 3-4. DOI: 10.1515/itit-2019-0032. Berlin: De Gruyter. PISSN: 2196-7032. pp. 189-204