Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
Jeff Ichnowski 565805cbb1 | 2 years ago | |
---|---|---|
rlqp-python @ 8c3e4de9b3 | 2 years ago | |
rlqp_train @ 0966749811 | 2 years ago | |
.gitmodules | 2 years ago | |
README.md | 2 years ago |
We demonstrate reinforcement learning can significantly accelerate first-order optimization, outperforming state-of-the-art solvers by up to 3x. RLQP avoids suboptimal heuristics within solvers by tuning the internal parameters of the ADMM algorithm. By decomposing the policy as a multi-agent partially observed problem, RLQP adapts to unseen problem classes and to larger problems than seen during training.
RLQP is composed of a few submodules, namely to (a) train the RL policy on a specific class of problems (source in rlqp_train/
) and (b) evaluate the policy on a test problem. Most users will want to start by using RLQP's policy to accelerate optimization of their problems.
To install the Python package to evaluate a pre-trained policy, run:
pip install git+https://github.com/berkeleyautomation/rlqp-python.git@55f378e496979bd00e84cea4583ac37bfaa571a9
This package contains a pre-trained model which should improve convergence beyond OSQP. The interface is identical to OSQP.
Please follow the instructions in the rlqp_train/
directory to setup and run training code. This code is still in preview mode as we work to release features like our TD3 policy.
@article{ichnowski2021rlqp,
title={Accelerating Quadratic Optimization with Reinforcement Learning},
author={Jeffrey Ichnowski, Paras Jain, Bartolomeo Stellato,
and Goran Banjac, Michael Luo, Francesco Borrelli
and Joseph E. Gonzalez, Ion Stoica, Ken Goldberg},
year={2021},
journal={arXiv preprint}
}
No Description
other
Dear OpenI User
Thank you for your continuous support to the Openl Qizhi Community AI Collaboration Platform. In order to protect your usage rights and ensure network security, we updated the Openl Qizhi Community AI Collaboration Platform Usage Agreement in January 2024. The updated agreement specifies that users are prohibited from using intranet penetration tools. After you click "Agree and continue", you can continue to use our services. Thank you for your cooperation and understanding.
For more agreement content, please refer to the《Openl Qizhi Community AI Collaboration Platform Usage Agreement》