(there is a lot here, I will try to organize as I am familar with all papers listed below)
-
A Logical Calculus of the Ideas Immanent in Nervous Activity
-
The Logic Theory Machine; A Complex Information Processing System
-
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
-
Dropout: A Simple Way to Prevent Neural Networks from Overfitting
-
Large-scale Video Classification with Convolutional Neural Networks
-
Supervised Sequence Labelling with Recurrent Neural Networks
-
A Theoretically Grounded Application of Dropout in Recurrent Neural Networks
-
Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting
-
Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling
-
On the Properties of Neural Machine Translation: Encoder-Decoder Approaches
-
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation
-
Zero-bias autoencoders and the benefits of co-adapting features
-
Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs)
-
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
-
Rectifier Nonlinearities Improve Neural Network Acoustic Models
-
Efficient Estimation of Word Representations in Vector Space
-
On the importance of initialization and momentum in deep learning
-
Adaptive Subgradient Methods for Online Learning and Stochastic Optimization
-
Aggregated Residual Transformations for Deep Neural Networks
-
Very Deep Convolutional Networks for Large-Scale Image Recognition
-
Xception: Deep Learning with Depthwise Separable Convolutions
-
Learning Transferable Architectures for Scalable Image Recognition
-
MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications
-
Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning
-
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
-
Understanding the difficulty of training deep feedforward neural networks
-
Exact solutions to the nonlinear dynamics of learning in deep linear neural networks
-
Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks
-
Neural Machine Translation by Jointly Learning to Align and Translate
-
Effective Approaches to Attention-based Neural Machine Translation
-
Norbert Wiener
-
Marvin Minsky
-
Richard Wallace
https://medium.com/@jankrikkeChina/cybernetics-explains-what-ai-is-and-what-it-isnt-13b2baec6cca