Stochastic resetting mitigates latent gradient bias of SGD from label noise
Giving up and starting over may seem wasteful in many situations such as searching for a target or training deep neural networks (DNNs). Our study, though, demonstrates that resetting from a checkpoint can significantly improve generalization performance when training DNNs with noisy labels. In the...
Saved in:
| Main Authors: | Youngkyoung Bae, Yeongwoo Song, Hawoong Jeong |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IOP Publishing
2025-01-01
|
| Series: | Machine Learning: Science and Technology |
| Subjects: | |
| Online Access: | https://doi.org/10.1088/2632-2153/adbc46 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
A Novel Sine Step Size for Warm-Restart Stochastic Gradient Descent
by: Mahsa Soheil Shamaee, et al.
Published: (2024-12-01) -
Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion
by: Tiankai Li, et al.
Published: (2024-12-01) -
The Generalization Error Bound for A Stochastic Gradient Descent Family bia A Gaussian Approximation Method
by: Chen Hao, et al.
Published: (2025-06-01) -
Long Short Term Memory Using Stochastic Gradient Descent and Adam for Stock Prediction
by: Muhammad Athanabil Andi Fawazdhia, et al.
Published: (2023-11-01) -
SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision
by: Yang Liu, et al.
Published: (2025-01-01)