Collection of papers for Stanford ML Group's summer 2018 reading group. Titles for each week are listed below, and more information on each paper can be found in the README file of each week's subdirectory. Discussions focused on numbered papers—bulleted papers were optional reading.
- Deep Residual Learning for Image Recognition
- Identity Mappings in Deep Residual Networks
- Wide Residual Networks
- Densely Connected Convolutional Networks
- Aggregated Residual Transformations for Deep Neural Networks
- Squeeze-and-Excitation Networks
- Can Spatiotemporal 3D CNNs Retrace the History of 2D CNNs and ImageNet?
- Batch Normalization: Accelerating Deep Network Training by Reducing
- Weight Normalization: A Simple Reparameterization to Accelerate Training
- Layer Normalization
- Instance Normalization: The Missing Ingredient for Fast Stylization
- Group Normalization
- Recurrent Batch Normalization
- Understanding the difficulty of training deep feedforward neural networks
- Delving Deep into Rectifiers: Surpassing Human-Level Performance on
- Rich feature hierarchies for accurate object detection and semantic
- Fast R-CNN
- Faster R-CNN: Towards Real-Time Object Detection with Region Proposal
- Mask R-CNN
- Spatial Transformer Networks
- R-FCN: Object Detection via Region-based Fully Convolutional Networks
- Feature Pyramid Networks for Object Detection
- Representation Learning: A Review and New Perspectives
- Why Does Unsupervised Pre-training Help Deep Learning?
- Taskonomy: Disentangling Task Transfer Learning
- How transferable are features in deep neural networks?
- A Survey on Transfer Learning
- Understanding deep learning requires rethinking generalization
- Intriguing properties of neural networks
- Fully Convolutional Networks for Semantic Segmentation
- U-Net: Convolutional Networks for Biomedical Image Segmentation
- Feature Pyramid Networks for Object Detection
- DeepLab: Semantic Image Segmentation with Deep Convolutional Nets,
- Encoder-Decoder with Atrous Separable Convolution for Semantic Image
- Focal Loss for Dense Object Detection
- Xception: Deep Learning with Depthwise Separable Convolutions
- Rethinking Atrous Convolution for Semantic Image Segmentation
- Intriguing properties of neural networks
- Explaining and Harnessing Adversarial Examples
- Distilling the Knowledge in a Neural Network
- Distillation as a Defense to Adversarial Perturbations against Deep
- Practical Black-Box Attacks against Machine Learning
- Adversarial examples in the physical world
- Auto-Encoding Variational Bayes
- Generative Adversarial Networks
- NICE: Non-linear Independent Components Estimation
- Pixel Recurrent Neural Networks
- Conditional Image Generation with PixelCNN Decoders
- Glow: Generative Flow with Invertible 1x1 Convolutions
- Improving Variational Inference with Inverse Autoregressive Flow
- Unsupervised Representation Learning with Deep Convolutional Generative
- Image-to-Image Translation with Conditional Adversarial Networks
- Self-Attention Generative Adversarial Networks
- Neural Architecture Search with Reinforcement Learning
- Population Based Training of Neural Networks
- Learning Transferable Architectures for Scalable Image Recognition
- Efficient Neural Architecture Search via Parameter Sharing
- Neural Architecture Optimization
- Efficient Neural Architecture Search with Network Morphism
- Searching for Activation Functions
- Neural Optimizer Search with Reinforcement Learning
- MnasNet: Platform-Aware Neural Architecture Search for Mobile
- Dropout as a Bayesian Approximation: Representing Model Uncertainty in
- Simple and Scalable Predictive Uncertainty Estimation using Deep
- What Uncertainties Do We Need in Bayesian Deep Learning for Computer
- Learning Confidence for Out-of-Distribution Detection in Neural Networks
- Leveraging uncertainty information from deep neural networks for disease detection
- A Probabilistic U-Net for Segmentation of Ambiguous Images
- The Importance of Knowing What We Don't Know
- Leveraging Uncertainty Estimates for Predicting Segmentation Quality
- Augmented RNNs (Blogpost)
- Neural Machine Translation by Jointly Learning to Align and Translate
- Show, Attend and Tell: Neural Image Caption Generation with Visual
- Effective Approaches to Attention-based Neural Machine Translation
- Attention Is All You Need
- QANet: Combining Local Convolution with Global Self-Attention for
- VideoLSTM Convolves, Attends and Flows for Action Recognition
- Non-local Neural Networks
- Self-Attention Generative Adversarial Networks
- Deep Compression: Compressing Deep Neural Networks with Pruning, Trained
- Quantized Neural Networks: Training Neural Networks with Low Precision
- SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB
- MobileNets: Efficient Convolutional Neural Networks for Mobile Vision
- ShuffleNet: An Extremely Efficient Convolutional Neural Network for