Tech Hive
  • Blog
  • About Me
  • Index

Machine learning


Tech notes

Machine learning


Tech notes
Back to Home Models Optimization

Comparison of loss functions of YOLO, SSD, RetinaNet

Posted on May 16, 2020

1. Introduction In this post, I would like to compare the loss functions used in different one-shot object detection methods, YOLO, SSD, and RetinaNet. One-shot object detection methods train the model on more than thousands grids with different scale, but the number of objects in one image is much... [Read More]

Tech notes of implementation of YOLO V3

Posted on April 29, 2020

1. Introduction To reproduce a deep neural network model, I have to get clear of three important elements: Network architeture Loss function Training practices Because YOLO V3 technical report[3] is very clear about the network and loss function, I talk about my experiences of these 2... [Read More]

Overview of optimizers for DNN: when and how to choose which optimizer

Posted on April 17, 2020

In this post, I would like to review the development of optimization methods for deep neural network (DNN) and share suggestions to use optimizers. What you can find: A brief review of the popular optimizers from the an intuitive perspective. The disadavantage of the popular adaptive optimizer, Adam.... [Read More]

Hyper-parameters tuning practices: learning rate, batch size, momentum, and weight decay

Posted on April 13, 2020

Tuning the hyper-parameters of a deep learning (DL) model by grid search or random search is computationally expensive and time consuming. This technical report gives several practical suggestions and steps to choose the optimal hyper-parameters. Some prior knowledge to fully understand this technical report: Overfitting/Underfitting Learning Rate (LR) Batch... [Read More]
  • GitHub
  • LinkedIn

Sheng FANG  •  2025

Theme by beautiful-jekyll