Skip to content

An introduction to generative adversarial network

License

Notifications You must be signed in to change notification settings

edigeze/GAN-course

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 

Repository files navigation

Important paper to read in deep learning.

⭐️⭐️⭐️: You must read

⭐️⭐️ : You should read

⭐️ : you may read if you have time

🔥 : hot paper, new paper a the state of the art

paper you must read in deep learning

Neural network

Neural Network⭐️⭐️⭐️

This paper introduce the neural network concept and these applications in NLP and computer vision

Y. LeCun, Y. Bengio, and G. Hinton. Deep learning, Nature, 2015.

Dropout⭐️⭐️

Dropout is a technique widely used for regularization in order to reduce overfitting in neural network

N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever and R. Salakhutdinov. Dropout: A simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research, 2014

Batch normalization⭐️⭐️

batch normalization is a technique used in neural network, We normalize input layer by adjusting and scaling activations. This technique allow to improve speed, performance and stability of neural network

S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. 2015.

Gradients Descent⭐️⭐️

This paper is a overview of the gradient descent based optimization algorithms for learning deep neural network models.

J. Zhang Gradient Descent based Optimization Algorithms for Deep Learning Models Training , 2019

Adam⭐️⭐️

Adam is a specific gradient descent algorithm widely use for the backpropagation of an neural network.

Kingma, Diederik P and Ba, Jimmy Lei. Adam: A method for stochastic optimization. 2014.

CNN

Image Classification

LeNet⭐️

One on the first convolutional neural network train on minsit classification. It popularize deep neural netwok

Y. LeCun et al. Gradient Based Learning Applied to Document Recognition, 1998.

AlexNet⭐️⭐️⭐️

This CNN fot image classification significantly outperformed all the prior competitors and won the challenge by reducing the top-5 error from 26% to 15.3%.

First Neural network to won imagenet chalenge

A. Krizhevsky, I. Sutskever, and G. Hinton. Imagenet classification with deep convolutional neural networks. In NIPS, 2012.

Inception⭐️⭐️

CNN for classification perform imageNet competion with a top-5 error rate of 6.67% (very close to human level performance) for V1

Inception v1 : C. Szegedy et al. Going deeper with convolutions. 2014.

Inception v2,3 :C. Szegedy et al. Rethinking the Inception Architecture for Computer Vision,2015

Inception v4 : C. Szegedy et al. Inception-v4, inception-resnet and the impact of residual connections on learning

VGG Net⭐️⭐️

CNN for classification, Similar to AlexNet, only 3x3 convolutions and increasing the nuber of filter and the number of hidden layer.

Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition, 2014

ResNet⭐️⭐️

CNN for image classification, introduce skip connection, batch normalisation. the increas the depth of 152 layers it achieves 3.57% error on th ImageNet challenge.

Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. IEEE, 2015.

DenseNet⭐️⭐️

This CNN used dense connection, fewer parameters than the others and we find a high accuracy.

G. Huang, Z. Liu, and K.Q. Weinberger. Densely connected convolutional networks.2016

Object Detection

R-CNN⭐️⭐️

CNN for object detection propose methode in three steps: extract region proposals, compute CNN features and Classify regions

R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. 2014.

Fast R-CNN ⭐️⭐️

Same that R-CNN but it build faster object detection algorithm

CNN for object detection, similar to R-CNN

R. Girshick. Fast R-CNN,2015

Faster R-CNN⭐️⭐️

CNN for object detection, this one don't use selective search to find out the region proposals. it use RPN Region Proposal Network give it less time consuming.

S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. 2015.

YOLO⭐️⭐️

All of the previous object detection algorithms use regions to localize the object within the image. The network does not look at the complete image.

J. Redmon, S. Divvala, R. Girshick, and A. Farhadi. You only look once: Unified, real-time object detection. In CVPR, 2016.

SSD⭐️⭐️

It is a CNN designed for object detection in real-time. SSD eliminate the need of the region proposal network to be faster. It applies a few improvements including multi-scale features and default boxes

W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.Y. Fu, and A. C. Berg. Ssd: Single shot multibox detector. 2016.

Mask R-CNN⭐️⭐️

CNN for object detection and segmentation,it extends Faster R-CNN by adding a branch for predicting an object mask in parallel with the existing branch for bounding box recognition.

K. He, G. Gkioxari, P. Dollar, and R. Girshick. Mask R-CNN. 2017.

Segmentation

U-Net⭐️⭐️

CNN for image segmentation initialy used for biomedical image, This CNN add a convolution layers and deconvolution layer that give it a U shape.

O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. 2015.

FCN⭐️⭐️

Constitute of two parts, Downsampling and Upsampling, this CNN it use in addition skip connection to segmentate the image.

J. Long, E. Shelhamer, and T. Darrell, Fully convolutional networks for semantic segmentation, 2016

Other Neural network

Variational autoencoders⭐️⭐️

D P. Kingma and M. Welling. Auto-Encoding Variational Bayes. CLR, 2013

AlphaGo⭐️

D. Silver et al, Mastering the game of Go with deep neural networks and tree search

AlphaGo Zero : D. Silver et al, Mastering the game of Go without human knowledge

Deep Reinforcement Learning⭐️⭐️

V. Mnih, et al, Human-level control through deep reinforcement learning

Seq To Seq⭐️⭐️

I. Sutskever, O. Vinyals and Q V. Le,. Sequence to Sequence Learning with Neural Networks, 2014

**Neural Machine Translation **⭐️⭐️

D. Bahdanau, K. Cho and Y. Bengio. Neural machine translation by jointly learning to align and translate, ICLR 2015

RNN Encoder-Decoder⭐️⭐️

K. Cho, B. van Merrienboer, C. Gulcehre, F. Bougares, H. Schwenk, and Y. Bengio, Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation, 2014

BERT⭐️

J. Cevlin et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,2019

LSTM⭐️ S. Hochreiter and J. Schmilhuber. Long Short-Term Memory

Visualization CNN⭐️

M D. Zeiler and R. Fergus. Visualizing and understanding convolutional networks, 2013

Show attend and tell⭐️

K. Xu et al. Show, attend and tell: Neural image caption generation with visual attention

Y. Kim. Convolutional Neural Networks for Sentence Classification, 2014

Delving Deep into Rectifiers⭐️

K. He et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

LRCN⭐️

J. Donahue, et al.Long-term Recurrent Convolutional Networks for Visual Recognition and Description,2016

Adversarial Autoencoders

A. Makhzani et al. Adversarial Autoencoders,2016

About

An introduction to generative adversarial network

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published