Skip to content

Shortest Path Biased Backpressure Routing Using Link Features and Graph Neural Networks

License

Notifications You must be signed in to change notification settings

zhongyuanzhao/dutyBP

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Shortest Path Biased Backpressure Routing Using Link Features and Graph Neural Networks

This repository is for the source code for paper "Biased Backpressure Routing Using Link Features and Graph Neural Networks" accepted for publication in IEEE TMLCN (IEEEXplore, preprint), as well as for its preliminary conference publications

  • Zhongyuan Zhao, Gunjan Verma, Ananthram Swami, Santiago Segarra, "Enhanced Backpressure with Wireless Link Features," IEEE CAMSAP 2023, Herradura, Costa Rica, 2023, pp. 271-275, doi: 10.1109/CAMSAP58249.2023.10403470. Preprint, IEEEXplore, slides, intro
  • Zhongyuan Zhao, Bojan Radojičić, Gunjan Verma, Ananthram Swami, Santiago Segarra, "Delay-aware Backpressure Routing Using Graph Neural Networks," IEEE ICASSP 2023, Rhodes Island, Greece, 2023, pp. 1-5, doi: 10.1109/ICASSP49357.2023.10095267. Preprint, IEEEXplore, slides, poster, video, Intro

Additional test results supplemental to the journal paper can be found here.

Abstract

To reduce the latency of Backpressure (BP) routing in wireless multi-hop networks, we propose to enhance the existing shortest path-biased BP (SP-BP) and sojourn time-based backlog metrics, since they introduce no additional time step-wise signaling overhead to the basic BP. Rather than relying on hop-distance, we introduce a new edge-weighted shortest path bias built on the scheduling duty cycle of wireless links, which can be predicted by a graph convolutional neural network based on the topology and traffic of wireless networks. Additionally, we tackle three long-standing challenges associated with SP-BP: optimal bias scaling, efficient bias maintenance, and integration of delay awareness. Our proposed solutions inherit the throughput optimality of the basic BP, as well as its practical advantages of low complexity and fully distributed implementation. Our approaches rely on common link features and introduces only a one-time constant overhead to previous SP-BP schemes, or a one-time overhead linear in the network size to the basic BP. Numerical experiments show that our solutions can effectively address the major drawbacks of slow startup, random walk, and the last packet problem in basic BP, improving the end-to-end delay of existing low-overhead BP algorithms under various settings of network traffic, interference, and mobility.

Environment setup

It is recommended to use docker to setup the software environment for convenience. We use Ubuntu Linux 20.04 LTS for demonstration, you may need to use different commands for other OS.

Step 1: install and setup docker, as well as your account, if you haven't done so.

Step 2: Pull tensorflow 2 with jupyter docker image

docker pull tensorflow/tensorflow:2.9.1-gpu-jupyter

Step 3: Launch your docker container from the image

docker run --user $(id -u):$(id -g) -it --rm -v ~/dutyBP:/tf/biasBP -w /tf/biasBP --gpus all tensorflow/tensorflow:2.9.1-gpu-jupyter

Step 4: Install additional packages from the jupyter command terminal

pip3 install -r requirements_exact.txt

Step 5: In your command line, commit the changes made to the docker container

docker container ls

Find the name of your docker container, for example, festive_einstein, then commit the changes

docker commit festive_einstein tensorflow/tensorflow:2.9.1-gpu-jupyter

You don't have to re-launch the container to run the code, it is for the convenience of the next time you launch the container.

How to run the code

The following steps demonstrate how to run the code referring to the journal paper "Biased Backpressure Routing Using Link Features and Graph Neural Networks".

The scripts are ran from the root of this repository.

Step 1: data generation (optional)

You can skip this step if using the given data sets

bash bash/data_gen.sh

Step 2: GNN training

You can skip this step if you are only to test the trained model in ./model/

python3.9 src/gnn_training_sch.py --training_set=PPm8 --learning_rate=0.0001 --weight_decay=0.001 --T=200 --datapath=./data/data_poisson_train --num_layer=5 --opt=5

To train the MLP-based model, comment out line 7 and uncomment line 8 in gnn_training_sch.py

The training converges in 5 epochs (about 5 hours on a workstation with a specification of 32GB memory, 8 cores, and Geforce GTX 1070 GPU.)

Step 3: testing

The test scripts here is designed for parallel execution a workstation with 48 CPU cores and 256GB RAM, if your computer has fewer CPU cores or RAM, it is advised to set the max_jobs accordingly in the bash scripts to avoid deadlock.

You can reduce or enlarge the test scope by changing for radius in 0.0 1.0 ; do in the following bash scripts, the manuscript only show results for radius=1.0.

Experiment 1: optimal bias scaling in Fig. 4

bash bash/test_100_10_mpx.sh 

Experiment 2: Section VII-B-1) test on all streaming traffic across different network sizes in Fig. 5(a)

bash bash/test_100_10_regular.sh 

Experiment 3: Section VII-B-2) test on all bursty traffic across different network sizes in Figs. 5(b) & 5(c)

bash bash/test_100_10_burst.sh 

Experiment 4: Section VII-B-3) test on 50/50 mixed streaming and bursty traffic across different network sizes in Figs. 6

bash bash/test_100_10_mixed.sh  

For additional curves for MLP-based routing schemes

bash bash/test_100_10_mixed_mlp.sh  

Experiment 5: Section VII-C network throughput across various traffic loads with all streaming flows in Fig. 7

bash bash/test_100_10_lambda.sh 

Notice that different routing schemes are coded as two digits numbers, some are experimental schemes not appear in the paper. The definition of the routing scheme code is as follows:

lgd_basic = {
    0: 'BP',
    1: 'SP-Hop',
    4: 'VBR',
    5: r'SP-$\bar{r}/(xr)$',
    6: r'SP-$1/x$',
    7: r'EDR-$\bar{r}$',
    8: 'MBP-4',
    9: r'SP-$1/r$',
}
lgd_add = {
    0: '',
    1: '-ph',
    2: '-expQ',
    3: '-min-expQ',
    4: '-min',
    5: '-SJB',
    6: '-HOL',
}
lgds = {}
for i in lgd_basic.keys():
    root = lgd_basic[i]
    for j in lgd_add.keys():
        app = lgd_add[j]
        key = j*10 + i
        lgds[key] = root + app

Results are visualized in results_plot.ipynb and addtional results for MLP test in results_plot-MLP.ipynb