88看球

清华主页 EN
导航菜单 八八看球

ReHLine: Regularized Composite ReLU-ReHU Loss Minimization with Linear Computation and Linear Convergence

来源: 04-23

八八看球
八八看球
八八看球
八八看球

时间:Fri., 16:00-17:00, April 24, 2026

地点:C654, Shuangqing Complex Building A

组织者:吴宇楠

主讲人:戴奔

Statistical Seminar

Organizer:

吴宇楠

Speaker:

戴奔 助理教授香港中文大学统计与数据科学系

Time:

Fri., 16:00-17:00, April 24, 2026

Venue:

C654, Shuangqing Complex Building A

Title:

ReHLine: Regularized Composite ReLU-ReHU Loss Minimization with Linear Computation and Linear Convergence

Abstract:

Empirical risk minimization (ERM) is a crucial framework that offers a general approach to handling a broad range of machine learning tasks. In this paper, we propose a novel algorithm, called ReHLine, for minimizing a set of regularized ERMs with convex piecewise linear-quadratic loss functions and optional linear constraints. The proposed algorithm can effectively handle diverse combinations of loss functions, regularizations, and constraints, making it particularly well-suited for complex domain-specific problems. Examples of such problems include FairSVM, elastic net regularized quantile regression, Huber minimization, etc. In addition, ReHLine enjoys a provable linear convergence rate and exhibits a per-iteration computational complexity that scales linearly with the sample size. The algorithm is implemented with both Python and R interfaces, and its performance is benchmarked on various tasks and datasets. Our experimental results demonstrate that ReHLine significantly surpasses generic optimization solvers in terms of computational efficiency on large-scale datasets. Moreover, it also outperforms specialized solvers such as liblinear in SVM, hqreg in Huber minimization and lightning(SAGA, SAG, SDCA, SVRG) in smooth SVM, exhibiting exceptional flexibility and efficiency.

返回顶部
88看球相关的文章