Skip to content

g-benton/PCC-RL

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

PCC-RL

Reinforcement learning resources for the Performance-oriented Congestion Control project.

Overview

This repo contains the gym environment required for training reinforcement learning models used in the PCC project along with the Python module required to run RL models in the PCC UDT codebase found at github.com/PCCProject/PCC-Uspace.

Training

To run training only, go to ./src/gym/, install any missing requirements for stable_solve.py and run that script. By default, this should replicate the model presented in A Reinforcement Learning Perspective on Internet Congestion Control, ICML 2019.

Testing Models

To test models in the real world (i.e., sending real packets into the Linux kernel and out onto a real or emulated network), download and install the PCC UDT code from github.com/PCCProject/PCC-Uspace. Follow the instructions in that repo for using congestion control algorithms with Python modules, and see ./src/gym/online/README.md for additional instructions regarding testing or training models in the real world.

About

Reinforcement learning resources for PCC.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 97.7%
  • Python 2.3%