How it Works#

NetworkX has the ability to dispatch function calls to separately-installed third-party backends.

NetworkX backends let users experience improved performance and/or additional functionality without changing their NetworkX Python code. Examples include backends that provide algorithm acceleration using GPUs, parallel processing, graph database integration, and more.

While NetworkX is a pure-Python implementation, backends may be written to use other libraries and even specialized hardware. nx-cugraph is a NetworkX backend that uses RAPIDS cuGraph and NVIDIA GPUs to significantly improve NetworkX performance.

nxcg-execution-flow

Enabling nx-cugraph#

It is recommended to use networkx>=3.4 for optimal zero code change performance, but nx-cugraph will also work with networkx 3.2+.

NetworkX will use nx-cugraph as the backend if any of the following are used:

NX_CUGRAPH_AUTOCONFIG environment variable.#

The NX_CUGRAPH_AUTOCONFIG environment variable can be used to configure NetworkX for full zero code change acceleration using nx-cugraph. If a NetworkX function is called that nx-cugraph supports, NetworkX will redirect the function call to nx-cugraph automatically, or fall back to either another backend if enabled or the default NetworkX implementation. See the NetworkX documentation on backends for configuring NetworkX manually.

bash> NX_CUGRAPH_AUTOCONFIG=True python my_networkx_script.py

backend= keyword argument#

To explicitly specify a particular backend for an API, use the backend= keyword argument. This argument takes precedence over the NX_CUGRAPH_AUTOCONFIG environment variable. This requires anyone running code that uses the backend= keyword argument to have the specified backend installed.

Example:

nx.betweenness_centrality(cit_patents_graph, k=k, backend="cugraph")

Type-based dispatching#

NetworkX also supports automatically dispatching to backends associated with specific graph types. Like the backend= keyword argument example above, this requires the user to write code for a specific backend, and therefore requires the backend to be installed, but has the advantage of ensuring a particular behavior without the potential for runtime conversions.

To use type-based dispatching with nx-cugraph, the user must import the backend directly in their code to access the utilities provided to create a Graph instance specifically for the nx-cugraph backend.

Example:

import networkx as nx
import nx_cugraph as nxcg

G = nx.Graph()

# populate the graph
#  ...

nxcg_G = nxcg.from_networkx(G)             # conversion happens once here
nx.betweenness_centrality(nxcg_G, k=1000)  # nxcg Graph type causes cugraph backend
                                           # to be used, no conversion necessary

Command Line Example#


Create bc_demo.ipy and paste the code below.

import pandas as pd
import networkx as nx

url = "https://data.rapids.ai/cugraph/datasets/cit-Patents.csv"
df = pd.read_csv(url, sep=" ", names=["src", "dst"], dtype="int32")
G = nx.from_pandas_edgelist(df, source="src", target="dst")

%time result = nx.betweenness_centrality(G, k=10)

Run the command:

user@machine:/# ipython bc_demo.ipy

CPU times: user 7min 36s, sys: 5.22 s, total: 7min 41s
Wall time: 7min 41s

You will observe a run time of approximately 7 minutes…more or less depending on your CPU.

Run the command again, this time specifying cugraph as the NetworkX backend.

user@machine:/# NX_CUGRAPH_AUTOCONFIG=True ipython bc_demo.ipy

CPU times: user 4.14 s, sys: 1.13 s, total: 5.27 s
Wall time: 5.32 s

This run will be much faster, typically around 5 seconds depending on your GPU.

Note, the examples above were run using the following specs:

    NetworkX 3.4
    nx-cugraph 24.10
    CPU: Intel(R) Xeon(R) Gold 6128 CPU @ 3.40GHz 45GB RAM
    GPU: NVIDIA Quadro RTX 8000 80GB RAM


The latest list of algorithms supported by nx-cugraph can be found in GitHub, or in the Supported Algorithms Section.