Tech Hive
  • Blog
  • About Me
  • Index

Machine learning


Optimization

Machine learning


Optimization
Back to Home Models Optimization

Comparison of loss functions of YOLO, SSD, RetinaNet

Posted on May 16, 2020

1. Introduction In this post, I would like to compare the loss functions used in different one-shot object detection methods, YOLO, SSD, and RetinaNet. One-shot object detection methods train the model on more than thousands grids with different scale, but the number of objects in one image is much... [Read More]

Overview of optimizers for DNN: when and how to choose which optimizer

Posted on April 17, 2020

In this post, I would like to review the development of optimization methods for deep neural network (DNN) and share suggestions to use optimizers. What you can find: A brief review of the popular optimizers from the an intuitive perspective. The disadavantage of the popular adaptive optimizer, Adam.... [Read More]

Hyper-parameters tuning practices: learning rate, batch size, momentum, and weight decay

Posted on April 13, 2020

Tuning the hyper-parameters of a deep learning (DL) model by grid search or random search is computationally expensive and time consuming. This technical report gives several practical suggestions and steps to choose the optimal hyper-parameters. Some prior knowledge to fully understand this technical report: Overfitting/Underfitting Learning Rate (LR) Batch... [Read More]
  • GitHub
  • LinkedIn

Sheng FANG  •  2025

Theme by beautiful-jekyll