Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
Shows that gradient descent with a large constant stepsize can accelerate convergence for $\ell_2$-regularized logistic regression with linearly separable data to $\widetilde{\mathcal{O}}(\sqrt{\kappa})$ steps, even though the objective evolves non-monotonically. This improves upon classical theory requiring small stepsizes for monotonic convergence.
Provides theoretical insights that can lead to faster and more efficient training of machine learning models, particularly for logistic regression tasks, potentially reducing computational costs.