Category Research Paper

196. Feature Pyramid Network

Feature Pyramid Network Feature pyramids are a basic component for detecting objects on different scales. Before this paper, a lot of research has been avoiding these pyramid structures due to their high computational and memory costs. Feature Pyramid Network tackles…

194. ArcFace Loss

ArcFace Loss One of tha main challenges in feature learning using Deep Convolutional Neural Networks for large scale face recognition is designing the optimal loss function that enhance discriminative power. Before this paper, there were mainly 2 lines to train…

192. SiLU

SILU SiLU is proposed as an activation function for neural network function approximation in reinforcement learning, and DSiLU is the derivative function for SiLU. DSiLU is a steeper and “overshot” version of the sigmoid function and it is proposed as…

186. PVTv2

PVTv2 The previous PVT had mainly 3 limitations. When processing high-resolution images, the computational cost is still relatively high. Loses local continuity of the image because it treats the image as a sequence of non-overlapping patches. Inflexible for arbitrary image…

185. DETR

DETR Modern object detectors predict a set of bounding boxes and category labels for each object of interest by defining surrogate regression and classification problems on a large set of proposals. This means that their performances heavily rely on post-processing…

184. Pyramid Vision Transformers

Background When using Traditional CNN-backboned architecture models, due to the convolutional filter’s weights being fully fixed after training, they suffered to adapt to different inputs dynamically. Vision Transformers attempted to remove the convolution from the backbone, but since it is…

180. Polynomial Learning Rate

Polynomial Learning Rate For deep learning models, the learning rate is one of the most important hyper-parameters in any deep neural network optimization process. Polynomial Learning Rate is a proposed technique to apply learning rate decay and optimize such process.…

177. PIDNet

PIDNet Today I’ve learned about PIDNet, so I’d like to share it here. Previously, I learned about BiSeNet which had a two-branched architecture to solve high latency problems. However, this architecture suffers another problem called “overshoot” where the boundary of…