๐ฆ ReHLine๏
ReHLine is designed to be a computationally efficient and practically useful software package for large-scale ERMs.
GitHub repo: https://github.com/softmin/ReHLine-python
Documentation: https://rehline.readthedocs.io
Open Source: MIT license
Paper: NeurIPS | 2023
ReHLine is designed to be a computationally efficient and practically useful software package for large-scale ERMs.
The proposed ReHLine solver has four appealing linear properties:
It applies to any convex piecewise linear-quadratic loss function, including the hinge loss, the check loss, the Huber loss, etc.
In addition, it supports linear equality and inequality constraints on the parameter vector.
The optimization algorithm has a provable linear convergence rate.
The per-iteration computational complexity is linear in the sample size.
๐จ Installation๏
Install rehline
using pip
pip install rehline
See more details in installation.
๐ฎ Formulation๏
ReHLine is designed to address the empirical regularized ReLU-ReHU minimization problem, named ReHLine optimization, of the following form:
where \(\mathbf{U} = (u_{li}),\mathbf{V} = (v_{li}) \in \mathbb{R}^{L \times n}\) and \(\mathbf{S} = (s_{hi}),\mathbf{T} = (t_{hi}),\mathbf{\tau} = (\tau_{hi}) \in \mathbb{R}^{H \times n}\) are the ReLU-ReHU loss parameters, and \((\mathbf{A},\mathbf{b})\) are the constraint parameters. This formulation has a wide range of applications spanning various fields, including statistics, machine learning, computational biology, and social studies. Some popular examples include SVMs with fairness constraints (FairSVM), elastic net regularized quantile regression (ElasticQR), and ridge regularized Huber minimization (RidgeHuber).

๐ Benchmark (powered by benchopt)๏
To generate benchmark results in our paper, please check ReHLine-benchmark.
Problem |
Results |
---|---|
Note: You may select the โlog-log scaleโ option in the left sidebar, as this will significantly improve the readability of the results.
๐งพ Overview of Results๏

Reference๏
If you use this code please star the repository and cite the following paper:
@article{daiqiu2023rehline,
title={ReHLine: Regularized Composite ReLU-ReHU Loss Minimization with Linear Computation and Linear Convergence},
author={Dai, Ben and Yixuan Qiu},
journal={Advances in Neural Information Processing Systems},
year={2023},
}