Object Detection

Method backbone test size VOC2007 VOC2010 VOC2012 ILSVRC 2013 MSCOCO 2015 Speed
OverFeat           24.3%    
R-CNN AlexNet   58.5% 53.7% 53.3% 31.4%    
R-CNN VGG16   66.0%          
SPP_net ZF-5   54.2%     31.84%    
DeepID-Net     64.1%     50.3%    
NoC 73.3%   68.8%          
Fast-RCNN VGG16   70.0% 68.8% 68.4%   19.7%(@[0.5-0.95]), 35.9%(@0.5)  
MR-CNN 78.2%   73.9%          
Faster-RCNN VGG16   78.8%   75.9%   21.9%(@[0.5-0.95]), 42.7%(@0.5) 198ms
Faster-RCNN ResNet101   85.6%   83.8%   37.4%(@[0.5-0.95]), 59.0%(@0.5)  
YOLO     63.4%   57.9%     45 fps
YOLO VGG-16     66.4%         21 fps
YOLOv2   448x448 78.6%   73.4%   21.6%(@[0.5-0.95]), 44.0%(@0.5) 40 fps
SSD VGG16 300x300 77.2%   75.8%   25.1%(@[0.5-0.95]), 43.1%(@0.5) 46 fps
SSD VGG16 512x512 79.8%   78.5%   28.8%(@[0.5-0.95]), 48.5%(@0.5) 19 fps
SSD ResNet101 300x300         28.0%(@[0.5-0.95]) 16 fps
SSD ResNet101 512x512         31.2%(@[0.5-0.95]) 8 fps
DSSD ResNet101 300x300         28.0%(@[0.5-0.95]) 8 fps
DSSD ResNet101 500x500         33.2%(@[0.5-0.95]) 6 fps
ION     79.2%   76.4%      
CRAFT     75.7%   71.3% 48.5%    
OHEM     78.9%   76.3%   25.5%(@[0.5-0.95]), 45.9%(@0.5)  
R-FCN ResNet50   77.4%         0.12sec(K40), 0.09sec(TitianX)
R-FCN ResNet101   79.5%         0.17sec(K40), 0.12sec(TitianX)
R-FCN(ms train) ResNet101   83.6%   82.0%   31.5%(@[0.5-0.95]), 53.2%(@0.5)  
PVANet 9.0     84.9%   84.2%     750ms(CPU), 46ms(TitianX)
RetinaNet ResNet101-FPN              
Light-Head R-CNN Xception* 800/1200         31.5%@[0.5:0.95] 95 fps
Light-Head R-CNN Xception* 700/1100         30.7%@[0.5:0.95] 102 fps

Published: 09 Oct 2015

Object Counting

Object Counting

Published: 09 Oct 2015

Natural Language Processing

Tutorials

Practical Neural Networks for NLP

Structured Neural Networks for NLP: From Idea to Code

Understanding Deep Learning Models in NLP

http://nlp.yvespeirsman.be/blog/understanding-deeplearning-models-nlp/

Deep learning for natural language processing, Part 1

https://softwaremill.com/deep-learning-for-nlp/

Neural Models

Unifying Visual-Semantic Embeddings with Multimodal Neural Language Models

Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks

Visualizing and Understanding Neural Models in NLP

Character-Aware Neural Language Models

Skip-Thought Vectors

A Primer on Neural Network Models for Natural Language Processing

Character-aware Neural Language Models

Neural Variational Inference for Text Processing

Sequence to Sequence Learning

Generating Text with Deep Reinforcement Learning

MUSIO: A Deep Learning based Chatbot Getting Smarter

Translation

Learning phrase representations using rnn encoder-decoder for statistical machine translation

Neural Machine Translation by Jointly Learning to Align and Translate

Multi-Source Neural Translation

Multi-Way, Multilingual Neural Machine Translation with a Shared Attention Mechanism

Modeling Coverage for Neural Machine Translation

A Character-level Decoder without Explicit Segmentation for Neural Machine Translation

NEMATUS: Attention-based encoder-decoder model for neural machine translation

Variational Neural Machine Translation

Neural Network Translation Models for Grammatical Error Correction

Linguistic Input Features Improve Neural Machine Translation

Sequence-Level Knowledge Distillation

Neural Machine Translation: Breaking the Performance Plateau

Tips on Building Neural Machine Translation Systems

Semi-Supervised Learning for Neural Machine Translation

EUREKA-MangoNMT: A C++ toolkit for neural machine translation for CPU

Deep Character-Level Neural Machine Translation

Neural Machine Translation Implementations

Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation

Learning to Translate in Real-time with Neural Machine Translation

Is Neural Machine Translation Ready for Deployment? A Case Study on 30 Translation Directions

Fully Character-Level Neural Machine Translation without Explicit Segmentation

Navigational Instruction Generation as Inverse Reinforcement Learning with Neural Machine Translation

Neural Machine Translation in Linear Time

Neural Machine Translation with Reconstruction

A Convolutional Encoder Model for Neural Machine Translation

Toward Multilingual Neural Machine Translation with Universal Encoder and Decoder

MXNMT: MXNet based Neural Machine Translation

Doubly-Attentive Decoder for Multi-modal Neural Machine Translation

Massive Exploration of Neural Machine Translation Architectures

Depthwise Separable Convolutions for Neural Machine Translation

Deep Architectures for Neural Machine Translation

Marian: Fast Neural Machine Translation in C++

Sockeye

Summarization

Extraction of Salient Sentences from Labelled Documents

A Neural Attention Model for Abstractive Sentence Summarization

A Convolutional Attention Network for Extreme Summarization of Source Code

Abstractive Text Summarization Using Sequence-to-Sequence RNNs and Beyond

textsum: Text summarization with TensorFlow

How to Run Text Summarization with TensorFlow

Reading Comprehension

Text Comprehension with the Attention Sum Reader Network

Text Understanding with the Attention Sum Reader Network

A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task

Consensus Attention-based Neural Networks for Chinese Reading Comprehension

Separating Answers from Queries for Neural Reading Comprehension

Attention-over-Attention Neural Networks for Reading Comprehension

Teaching Machines to Read and Comprehend CNN News and Children Books using Torch

Reasoning with Memory Augmented Neural Networks for Language Comprehension

Bidirectional Attention Flow: Bidirectional Attention Flow for Machine Comprehension

NewsQA: A Machine Comprehension Dataset

Gated-Attention Readers for Text Comprehension

Get To The Point: Summarization with Pointer-Generator Networks

Language Understanding

Recurrent Neural Networks with External Memory for Language Understanding

Neural Semantic Encoders

Neural Tree Indexers for Text Understanding

Better Text Understanding Through Image-To-Text Transfer

Text Classification

Convolutional Neural Networks for Sentence Classification

Recurrent Convolutional Neural Networks for Text Classification

Character-level Convolutional Networks for Text Classification

A C-LSTM Neural Network for Text Classification

Rationale-Augmented Convolutional Neural Networks for Text Classification

Text classification using DIGITS and Torch7

Recurrent Neural Network for Text Classification with Multi-Task Learning

Deep Multi-Task Learning with Shared Memory

Virtual Adversarial Training for Semi-Supervised Text Classification

Adversarial Training Methods for Semi-Supervised Text Classification

Sentence Convolution Code in Torch: Text classification using a convolutional neural network

Bag of Tricks for Efficient Text Classification

Actionable and Political Text Classification using Word Embeddings and LSTM

Implementing a CNN for Text Classification in TensorFlow

fancy-cnn: Multiparadigm Sequential Convolutional Neural Networks for text classification

Convolutional Neural Networks for Text Categorization: Shallow Word-level vs. Deep Character-level

Tweet Classification using RNN and CNN

Hierarchical Attention Networks for Document Classification

AC-BLSTM: Asymmetric Convolutional Bidirectional LSTM Networks for Text Classification

Generative and Discriminative Text Classification with Recurrent Neural Networks

Adversarial Multi-task Learning for Text Classification

Deep Text Classification Can be Fooled

Deep neural network framework for multi-label text classification

Multi-Task Label Embedding for Text Classification

Text Clustering

Self-Taught Convolutional Neural Networks for Short Text Clustering

Alignment

Aligning Books and Movies: Towards Story-like Visual Explanations by Watching Movies and Reading Books

Dialog

Visual Dialog

Papers, code and data from FAIR for various memory-augmented nets with application to text understanding and dialogue.

Neural Emoji Recommendation in Dialogue Systems

Memory Networks

Neural Turing Machines

Memory Networks

End-To-End Memory Networks

Reinforcement Learning Neural Turing Machines - Revised


Learning to Transduce with Unbounded Memory

How to Code and Understand DeepMind’s Neural Stack Machine


Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

Ask Me Even More: Dynamic Memory Tensor Networks (Extended Model)

Structured Memory for Neural Turing Machines

Dynamic Memory Networks for Visual and Textual Question Answering

Neural GPUs Learn Algorithms

Hierarchical Memory Networks

Convolutional Residual Memory Networks

NTM-Lasagne: A Library for Neural Turing Machines in Lasagne

Evolving Neural Turing Machines for Reward-based Learning

Hierarchical Memory Networks for Answer Selection on Unknown Words

Gated End-to-End Memory Networks

Can Active Memory Replace Attention?

A Taxonomy for Neural Memory Networks

Papers

Globally Normalized Transition-Based Neural Networks

A Decomposable Attention Model for Natural Language Inference

Improving Recurrent Neural Networks For Sequence Labelling

Recurrent Memory Networks for Language Modeling

Tweet2Vec: Learning Tweet Embeddings Using Character-level CNN-LSTM Encoder-Decoder

Learning text representation using recurrent convolutional neural network with highway layers

Ask the GRU: Multi-task Learning for Deep Text Recommendations

From phonemes to images: levels of representation in a recurrent neural model of visually-grounded language learning

Visualizing Linguistic Shift

A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks

Deep Learning applied to NLP

https://arxiv.org/abs/1703.03091

Attention Is All You Need

Recent Trends in Deep Learning Based Natural Language Processing

HotFlip: White-Box Adversarial Examples for NLP

No Metrics Are Perfect: Adversarial Reward Learning for Visual Storytelling

Interesting Applications

Data-driven HR - Résumé Analysis Based on Natural Language Processing and Machine Learning

sk_p: a neural program corrector for MOOCs

Neural Generation of Regular Expressions from Natural Language with Minimal Domain Knowledge

emoji2vec: Learning Emoji Representations from their Description

Inside-Outside and Forward-Backward Algorithms Are Just Backprop (Tutorial Paper)

Cruciform: Solving Crosswords with Natural Language Processing

Smart Reply: Automated Response Suggestion for Email

Deep Learning for RegEx

Learning Python Code Suggestion with a Sparse Pointer Network

End-to-End Prediction of Buffer Overruns from Raw Source Code via Neural Memory Networks

https://arxiv.org/abs/1703.02458

Convolutional Sequence to Sequence Learning

DeepFix: Fixing Common C Language Errors by Deep Learning

Hierarchically-Attentive RNN for Album Summarization and Storytelling

Project

TheanoLM - An Extensible Toolkit for Neural Network Language Modeling

NLP-Caffe: natural language processing with Caffe

DL4NLP: Deep Learning for Natural Language Processing

Combining CNN and RNN for spoken language identification

Character-Aware Neural Language Models: LSTM language model with CNN over characters in TensorFlow

Neural Relation Extraction with Selective Attention over Instances

deep-simplification: Text simplification using RNNs

lamtram: A toolkit for language and translation modeling using neural networks

Lango: Language Lego

Sequence-to-Sequence Learning with Attentional Neural Networks

harvardnlp code

Seq2seq: Sequence to Sequence Learning with Keras

debug seq2seq

Recurrent & convolutional neural network modules

Datasets

Datasets for Natural Language Processing

Blogs

How to read: Character level deep learning

Heavy Metal and Natural Language Processing

Sequence To Sequence Attention Models In PyCNN

https://talbaumel.github.io/Neural+Attention+Mechanism.html

Source Code Classification Using Deep Learning

http://blog.aylien.com/source-code-classification-using-deep-learning/

My Process for Learning Natural Language Processing with Deep Learning

https://medium.com/@MichaelTeifel/my-process-for-learning-natural-language-processing-with-deep-learning-bd0a64a36086

Convolutional Methods for Text

https://medium.com/@TalPerry/convolutional-methods-for-text-d5260fd5675f

Word2Vec

Word2Vec Tutorial - The Skip-Gram Model

http://mccormickml.com/2016/04/19/word2vec-tutorial-the-skip-gram-model/

Word2Vec Tutorial Part 2 - Negative Sampling

http://mccormickml.com/2017/01/11/word2vec-tutorial-part-2-negative-sampling/

Word2Vec Resources

http://mccormickml.com/2016/04/27/word2vec-resources/

Demos

AskImage.org - Deep Learning for Answering Questions about Images

Talks / Videos

Navigating Natural Language Using Reinforcement Learning

Resources

So, you need to understand language data? Open-source NLP software can help!

Curated list of resources on building bots

Notes for deep learning on NLP

https://medium.com/@frank_chung/notes-for-deep-learning-on-nlp-94ddfcb45723#.iouo0v7m7

Published: 09 Oct 2015

Neural Architecture Search

Papers

Published: 09 Oct 2015

LiDAR 3D Object Detection

Papers

Published: 09 Oct 2015

Acceleration and Model Compression

Papers

Published: 09 Oct 2015

Graph Convolutional Networks

Learning Convolutional Neural Networks for Graphs

Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering

Semi-Supervised Classification with Graph Convolutional Networks

Graph Based Convolutional Neural Network

How powerful are Graph Convolutions? (review of Kipf & Welling, 2016)

http://www.inference.vc/how-powerful-are-graph-convolutions-review-of-kipf-welling-2016-2/

Graph Convolutional Networks

DeepGraph: Graph Structure Predicts Network Growth

Deep Learning with Sets and Point Clouds

Deep Learning on Graphs

Robust Spatial Filtering with Graph Convolutional Neural Networks

https://arxiv.org/abs/1703.00792

Modeling Relational Data with Graph Convolutional Networks

https://arxiv.org/abs/1703.06103

Distance Metric Learning using Graph Convolutional Networks: Application to Functional Brain Networks

Deep Learning on Graphs with Graph Convolutional Networks

Deep Learning on Graphs with Keras

Learning Graph While Training: An Evolving Graph Convolutional Neural Network

https://arxiv.org/abs/1708.04675

Graph Attention Networks

Residual Gated Graph ConvNets

https://arxiv.org/abs/1711.07553

Probabilistic and Regularized Graph Convolutional Networks

Videos as Space-Time Region Graphs

https://arxiv.org/abs/1806.01810

Relational inductive biases, deep learning, and graph networks

Can GCNs Go as Deep as CNNs?

GMNN: Graph Markov Neural Networks

DeepGCNs: Making GCNs Go as Deep as CNNs

Rethinking pooling in graph neural networks

Published: 09 Oct 2015

Generative Adversarial Networks

Generative Adversarial Networks

Generative Adversarial Nets

Adversarial Feature Learning

Generative Adversarial Networks

Adversarial Examples and Adversarial Training

How to Train a GAN? Tips and tricks to make GANs work

Unsupervised and Semi-supervised Learning with Categorical Generative Adversarial Networks

Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks

InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets

Learning Interpretable Latent Representations with InfoGAN: A tutorial on implementing InfoGAN in Tensorflow

Coupled Generative Adversarial Networks

Energy-based Generative Adversarial Network

SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient

Connecting Generative Adversarial Networks and Actor-Critic Methods

Generative Adversarial Nets from a Density Ratio Estimation Perspective

Unrolled Generative Adversarial Networks

Generative Adversarial Networks as Variational Training of Energy Based Models

Multi-class Generative Adversarial Networks with the L2 Loss Function

Least Squares Generative Adversarial Networks

Inverting The Generator Of A Generative Adversarial Networ

ml4a-invisible-cities

Semi-Supervised Learning with Context-Conditional Generative Adversarial Networks

Associative Adversarial Networks

Temporal Generative Adversarial Nets

Handwriting Profiling using Generative Adversarial Networks

  • intro: Accepted at The Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17 Student Abstract and Poster Program)
  • arxiv: https://arxiv.org/abs/1611.08789

C-RNN-GAN: Continuous recurrent neural networks with adversarial training

Ensembles of Generative Adversarial Networks

Improved generator objectives for GANs

Stacked Generative Adversarial Networks

Unsupervised Pixel-Level Domain Adaptation with Generative Adversarial Networks

AdaGAN: Boosting Generative Models

Towards Principled Methods for Training Generative Adversarial Networks

Wasserstein GAN

Improved Training of Wasserstein GANs

On the effect of Batch Normalization and Weight Normalization in Generative Adversarial Networks

On the Effects of Batch and Weight Normalization in Generative Adversarial Networks

Controllable Generative Adversarial Network

Generative Adversarial Networks: An Overview

  • intro: Imperial College London & Victoria University of Wellington & University of Montreal & Cortexica Vision Systems Ltd
  • intro: IEEE Signal Processing Magazine Special Issue on Deep Learning for Visual Understanding
  • arxiv: https://arxiv.org/abs/1710.07035

CyCADA: Cycle-Consistent Adversarial Domain Adaptation

https://arxiv.org/abs/1711.03213

Spectral Normalization for Generative Adversarial Networks

https://openreview.net/forum?id=B1QRgziT-

Are GANs Created Equal? A Large-Scale Study

GAGAN: Geometry-Aware Generative Adverserial Networks

https://arxiv.org/abs/1712.00684

CycleGAN: a Master of Steganography

PacGAN: The power of two samples in generative adversarial networks

ComboGAN: Unrestrained Scalability for Image Domain Translation

Decoupled Learning for Conditional Adversarial Networks

https://arxiv.org/abs/1801.06790

No Modes left behind: Capturing the data distribution effectively using GANs

Improving GAN Training via Binarized Representation Entropy (BRE) Regularization

On GANs and GMMs

https://arxiv.org/abs/1805.12462

The Unusual Effectiveness of Averaging in GAN Training

https://arxiv.org/abs/1806.04498

Understanding the Effectiveness of Lipschitz Constraint in Training of GANs via Gradient Analysis

https://arxiv.org/abs/1807.00751

The GAN Landscape: Losses, Architectures, Regularization, and Normalization

Which Training Methods for GANs do actually Converge?

Convergence Problems with Generative Adversarial Networks (GANs)

Bayesian CycleGAN via Marginalizing Latent Sampling

https://arxiv.org/abs/1811.07465

GAN Dissection: Visualizing and Understanding Generative Adversarial Networks

https://arxiv.org/abs/1811.10597

Do GAN Loss Functions Really Matter?

https://arxiv.org/abs/1811.09567

Image-to-Image Translation

Pix2Pix

Image-to-Image Translation with Conditional Adversarial Networks

Remastering Classic Films in Tensorflow with Pix2Pix

Image-to-Image Translation in Tensorflow

webcam pix2pix

https://github.com/memo/webcam-pix2pix-tensorflow


Unsupervised Image-to-Image Translation with Generative Adversarial Networks

Unsupervised Image-to-Image Translation Networks

Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks

CycleGAN and pix2pix in PyTorch

Perceptual Adversarial Networks for Image-to-Image Transformation

https://arxiv.org/abs/1706.09138

XGAN: Unsupervised Image-to-Image Translation for many-to-many Mappings

In2I : Unsupervised Multi-Image-to-Image Translation Using Generative Adversarial Networks

https://arxiv.org/abs/1711.09334

StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation

Discriminative Region Proposal Adversarial Networks for High-Quality Image-to-Image Translation

https://arxiv.org/abs/1711.09554

Toward Multimodal Image-to-Image Translation

Face Translation between Images and Videos using Identity-aware CycleGAN

https://arxiv.org/abs/1712.00971

Unsupervised Multi-Domain Image Translation with Domain-Specific Encoders/Decoders

https://arxiv.org/abs/1712.02050

High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs

On the Effectiveness of Least Squares Generative Adversarial Networks

https://arxiv.org/abs/1712.06391

GANs for Limited Labeled Data

Defending Against Adversarial Examples

Conditional Image-to-Image Translation

XOGAN: One-to-Many Unsupervised Image-to-Image Translation

https://arxiv.org/abs/1805.07277

Unsupervised Attention-guided Image to Image Translation

https://arxiv.org/abs/1806.02311

Exemplar Guided Unsupervised Image-to-Image Translation

https://arxiv.org/abs/1805.11145

Improving Shape Deformation in Unsupervised Image-to-Image Translation

https://arxiv.org/abs/1808.04325

Video-to-Video Synthesis

Segmentation Guided Image-to-Image Translation with Adversarial Networks

https://arxiv.org/abs/1901.01569

Projects

Generative Adversarial Networks with Keras

Generative Adversarial Network Demo for Fresh Machine Learning #2

TextGAN: A generative adversarial network for text generation, written in TensorFlow.

cleverhans v0.1: an adversarial machine learning library

Deep Convolutional Variational Autoencoder w/ Adversarial Network

A versatile GAN(generative adversarial network) implementation. Focused on scalability and ease-of-use.

AdaGAN: Boosting Generative Models

TensorFlow-GAN (TFGAN)

Blogs

Generative Adversial Networks Explained

Generative Adversarial Autoencoders in Theano

An introduction to Generative Adversarial Networks (with code in TensorFlow)

Difficulties training a Generative Adversarial Network

Are Energy-Based GANs any more energy-based than normal GANs?

http://www.inference.vc/are-energy-based-gans-actually-energy-based/

Generative Adversarial Networks Explained with a Classic Spongebob Squarepants Episode: Plus a Tensorflow tutorial for implementing your own GAN

Deep Learning Research Review Week 1: Generative Adversarial Nets

Stability of Generative Adversarial Networks

Instance Noise: A trick for stabilising GAN training

Generating Fine Art in 300 Lines of Code

Talks / Videos

Generative Adversarial Network visualization

Resources

The GAN Zoo

AdversarialNetsPapers: The classical Papers about adversial nets

GAN Timeline

Published: 09 Oct 2015