TY - JOUR
T1 - Cyclical Learning Rate Optimization on Deep Learning Model for Brain Tumor Segmentation
AU - Fajar, Aziz
AU - Sarno, Riyanarto
AU - Fatichah, Chastine
AU - Susilo, Rahadian Indarto
AU - Pangestu, Gusti
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2023
Y1 - 2023
N2 - In recent years, deep learning has found widespread applications in tasks such as segmentation and classification. Fine-tuning hyperparameters is crucial to improve performance, with the learning rate being a key parameter. Various methods, including adaptive learning rates, learning rate scheduling, and cyclical learning rates, have been used to optimize learning rates. Cyclical learning rates offer significant benefits with minimal computational cost, as seen in previous research. This study introduces a novel approach to tuning the cyclical learning rate, which incorporates the exponential moving average. These methods are applied to the BraTS 2021 dataset for segmentation tasks, resulting in superior performance compared to the previous approach. Our proposed method reduces the epochs required to reach convergence by 19 and 54 epochs for U-Net and Dense U-net, respectively. For Res U-net, the epoch needed to convergence is 10 epochs more. However, the proposed method produces lower loss values with 0.707, 0.657, and 0.665 compared to the previous method with 0.712, 0.685, and 0.725 for U-net, Res U-net, and Dense U-net, respectively.
AB - In recent years, deep learning has found widespread applications in tasks such as segmentation and classification. Fine-tuning hyperparameters is crucial to improve performance, with the learning rate being a key parameter. Various methods, including adaptive learning rates, learning rate scheduling, and cyclical learning rates, have been used to optimize learning rates. Cyclical learning rates offer significant benefits with minimal computational cost, as seen in previous research. This study introduces a novel approach to tuning the cyclical learning rate, which incorporates the exponential moving average. These methods are applied to the BraTS 2021 dataset for segmentation tasks, resulting in superior performance compared to the previous approach. Our proposed method reduces the epochs required to reach convergence by 19 and 54 epochs for U-Net and Dense U-net, respectively. For Res U-net, the epoch needed to convergence is 10 epochs more. However, the proposed method produces lower loss values with 0.707, 0.657, and 0.665 compared to the previous method with 0.712, 0.685, and 0.725 for U-net, Res U-net, and Dense U-net, respectively.
KW - 3D image data
KW - deep learning
KW - exponential moving average
KW - information and communication technology
KW - learning rate
KW - scientific research
UR - http://www.scopus.com/inward/record.url?scp=85176313658&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2023.3326475
DO - 10.1109/ACCESS.2023.3326475
M3 - Article
AN - SCOPUS:85176313658
SN - 2169-3536
VL - 11
SP - 119802
EP - 119810
JO - IEEE Access
JF - IEEE Access
ER -