本文共 15241 字,大约阅读时间需要 50 分钟。
Awesome Torch
This blog from:
A curated list of awesome Torch tutorials, projects and communities.
Table of Contents
Tutorials
- CVPR15 Tutorial
- for IPAM Summer School on Deep Learning.
Model Zoo
Codes and related articles. (#)
means authors of code and paper are different.
Recurrent Networks
-
- Wojciech Zaremba, Tomas Mikolov, Armand Joulin, Rob Fergus, Learning Simple Algorithms from Examples, arXiv:1511.07275
-
- Tomas Mikolov, Armand Joulin, Sumit Chopra, Michael Mathieu, Marc'Aurelio Ranzato, Learning Longer Memory in Recurrent Neural Networks, arXiv:1406.1078
-
- Kai Sheng Tai, Richard Socher, Christopher D. Manning, Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks, ACL 2015
-
- Yoon Kim, Yacine Jernite, David Sontag, Alexander M. Rush, Character-Aware Neural Language Models, AAAI 2016,
- ,
- Andrej Karpathy, Justin Johnson, Li Fei-Fei, Visualizing and Understanding Recurrent Networks, ICLR 2016,
-
- Wojciech Zaremba, Ilya Sutskever, Oriol Vinyal, Recurrent Neural Network Regularization, arXiv:1409.2329
-
- Wojciech Zaremba, Ilya Sutskever, Learning to Execute, arXiv:1410.4615
-
- (#) Oriol Vinyals, Alexander Toshev, Samy Bengio, Dumitru Erhan, Show and Tell: A Neural Image Caption Generator, CVPR 2015
-
- (#) Ilya Sutskever, Oriol Vinyals, Quoc V. Le, Sequence to Sequence Learning with Neural Networks, NIPS 2014
-
- (#) Ilya Sutskever, Oriol Vinyals, Quoc V. Le, Sequence to Sequence Learning with Neural Networks, NIPS 2014
-
- (#) Ilya Sutskever, Oriol Vinyals, Quoc V. Le, Sequence to Sequence Learning with Neural Networks, NIPS 2014
-
- (#) Nal Kalchbrenner, Ivo Danihelka, Alex Graves, Grid Long Short-Term Memory, arXiv:1507.01526,
-
- (#) Volodymyr Mnih, Nicolas Heess, Alex Graves, Koray Kavukcuoglu, Recurrent Models of Visual Attention, NIPS 2014
-
- (#) Karol Gregor, Ivo Danihelka, Alex Graves, Danilo Jimenez Rezende, Daan Wierstra, DRAW: A Recurrent Neural Network For Image Generation, arXiv:1502.04623
-
- (#) Aaron van den Oord, Nal Kalchbrenner, Koray Kavukcuoglu, Pixel Recurrent Neural Networks, arXiv:1601.06759
-
- Aishwarya Agrawal, Jiasen Lu, Stanislaw Antol, Margaret Mitchell, C. Lawrence Zitnick, Dhruv Batra, Devi Parikh, VQA: Visual Question Answering, arXiv:1505.00468,
-
- Dario Amodei, Rishita Anubhai, Eric Battenberg, Carl Case, Jared Casper, Bryan Catanzaro, Jingdong Chen, Mike Chrzanowski, Adam Coates, Greg Diamos, Erich Elsen, Jesse Engel, Linxi Fan, Christopher Fougner, Tony Han, Awni Hannun, Billy Jun, Patrick LeGresley, Libby Lin, Sharan Narang, Andrew Ng, Sherjil Ozair, Ryan Prenger, Jonathan Raiman, Sanjeev Satheesh, David Seetapun, Shubho Sengupta, Yi Wang, Zhiqian Wang, Chong Wang, Bo Xiao, Dani Yogatama, Jun Zhan, Zhenyao Zhu, Deep Speech 2: End-to-End Speech Recognition in English and Mandarin, arXiv:1512.02595,
-
- Justin Johnson, Andrej Karpathy, Li Fei-Fei, DenseCap: Fully Convolutional Localization Networks for Dense Captioning, CVPR 2016,
-
- (#) Minh-Thang Luong, Hieu Pham, Christopher D. Manning, Effective Approaches to Attention-based Neural Machine Translation, EMNLP 2015,
-
- (#) Tim Cooijmans, Nicolas Ballas, César Laurent, Çağlar Gülçehre, Aaron Courville, Recurrent Batch Normalization, arXiv:1603.09025,
-
- Colton Gyulay, Michael Farrell, * End-to-End Generative Dialogue*,
-
- Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem and Juan Carlos Niebles, Activitynet: A large-scale video benchmark for human activity understanding, CVPR 2015,
-
- Tomas Mikolov, Armand Joulin, Sumit Chopra, Michael Mathieu, Marc'Aurelio Ranzato, Learning Longer Memory in Recurrent Neural Networks, arXiv:1412.7753,
-
- Jiasen Lu, Jianwei Yang, Dhruv Batra, Devi Parikh, Hierarchical Question-Image Co-Attention for Visual Question Answering, arXiv:1606.00061,
-
- Viorica Patraucean, Ankur Handa, Roberto Cipolla, Spatio-temporal video autoencoder with differentiable memory, ICLR 2016 Workshop,
Convolutional Networks
-
- Xiang Zhang, Junbo Zhao, Yann LeCun, Character-level Convolutional Networks for Text Classification, NIPS 2015
-
- (#) Alec Radford, Luke Metz, Soumith Chintala, Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks, arXiv:1511.06434v1
-
- (#) Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, Andrew Rabinovich, Going Deeper with Convolutions, CVPR 2015
-
- (#) Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jonathon Shlens, Zbigniew Wojna, Rethinking the Inception Architecture for Computer Vision, arXiv:1512.00567,
-
- (#) Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning, arXiv:1602.07261,
-
- (#) Florian Schroff, Dmitry Kalenichenko, James Philbin, FaceNet: A Unified Embedding for Face Recognition and Clustering, CVPR 2015
- ,
- (#) Leon A. Gatys, Alexander S. Ecker, Matthias Bethge, A Neural Algorithm of Artistic Style, arXiv:1508.06576
-
- (#) Chao Dong, Chen Change Loy, Kaiming He, Xiaoou Tang, Image Super-Resolution Using Deep Convolutional Networks, arXiv:1501.00092
-
- (#) Pierre Sermanet, David Eigen, Xiang Zhang, Michael Mathieu, Rob Fergus, Yann LeCun, OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks, arXiv:1312.6229
-
- (#) K. Simonyan, A. Zisserman, Very Deep Convolutional Networks for Large-Scale Image Recognition, arXiv:1409.1556
-
- (#) Alex J. Champandard Semantic Style Transfer and Turning Two-Bit Doodles into Fine Artworks, arXiv:1603.01768
-
- Dmitry Ulyanov, Vadim Lebedev, Andrea Vedaldi, Victor Lempitsky Texture Networks: Feed-forward Synthesis of Textures and Stylized Images, arXiv:1603.03417
-
- Manuel Ruder, Alexey Dosovitskiy, Thomas Brox Artistic style transfer for videos, arXiv:1604.08610
-
- (#) Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun, Deep Residual Learning for Image Recognition, arXiv:1512.03385,
-
- Gao Huang, Yu Sun, Zhuang Liu, Daniel Sedra, Kilian Weinberger, Deep Networks with Stochastic Depth, arXiv:1603.09382,
-
- (#) Yoon Kim, Convolutional Neural Networks for Sentence Classification, arXiv:1408.5882,
-
- Chuan Li, Michael Wand, Precomputed Real-Time Texture Synthesis with Markovian Generative Adversarial Networks, arXiv:1604.04382,
-
- Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun, Identity Mappings in Deep Residual Networks, arXiv:1603.05027,
-
- (#) Fisher Yu, Vladlen Koltun, Multi-Scale Context Aggregation by Dilated Convolutions,
-
- Chuan Li, Michael Wand, Combining Markov Random Fields and Convolutional Neural Networks for Image Synthesis, arXiv:1601.04589,
-
- Alejandro Newell, Kaiyu Yang, Jia Deng, Stacked Hourglass Networks for Human Pose Estimation, arXiv:1603.06937,
-
- Sergey Zagoruyko, Nikos Komodakis, Wide Residual Networks, Wide Residual Networks, arXiv:1605.07146,
-
- Jianwei Yang, Devi Parikh, Dhruv Batra, Joint Unsupervised Learning of Deep Representations and Image Clusters, CVPR 2016,
-
- (#) Ross Girshick, Fast R-CNN, ICCV 2015,
-
- Scott Reed, Zeynep Akata, Honglak Lee, Bernt Schiele, Learning Deep Representations of Fine-grained Visual Descriptions, CVPR 2016,
-
- Scott Reed, Zeynep Akata, Xinchen Yan, Lajanugen Logeswaran, Bernt Schiele, Honglak Lee, Generative Adversarial Text to Image Synthesis, ICML 2016,
-
- Yuandong Tian, Yan Zhu, Better Computer Go Player with Neural Network and Long-term Prediction, ICLR 2016,
-
- Qi, Charles R and Su, Hao and Niessner, Matthias and Dai, Angela and Yan, Mengyuan and Guibas, Leonidas J, Volumetric and Multi-View CNNs for Object Classification on 3D Data, arXiv preprint arXiv:1604.03265, 2016, ,
-
- Pedro O. Pinheiro, Ronan Collobert, Piotr Dollar, Learning to Segment Object Candidates, arXiv:1506.06204,
- Pedro O. Pinheiro, Tsung-Yi Lin, Ronan Collobert, Piotr Dollàr, Learning to Refine Object Segments, arXiv:1603.08695,
-
- imagenet-multiGPU.torch + fb.resnet.torch in torchnet
-
- Edgar Simo-Serra, Hiroshi Ishikawa, Fashion Style in 128 Floats: Joint Ranking and Classification using Weak Data for Feature Extraction, CVPR 2016,
-
- Adam Paszke, Abhishek Chaurasia, Sangpil Kim, Eugenio Culurciello, ENet: A Deep Neural Network Architecture for Real-Time Semantic Segmentation, arXiv:1606.02147,
Reinforcement Learning
- ,
- Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, Martin Riedmiller, Andreas K. Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, Demis Hassabis, Human-Level Control through Deep Reinforcement Learning, Nature,
-
- (#) Ivan Sorokin, Alexey Seleznev, Mikhail Pavlov, Aleksandr Fedorov, Anastasiia Ignateva, Deep Attention Recurrent Q-Network, NIPS 2015,
-
- (#) Marc G. Bellemare, Georg Ostrovski, Arthur Guez, Philip S. Thomas, Rémi Munos, Increasing the Action Gap: New Operators for Reinforcement Learning, arXiv:1512.04860,
-
- Jie Fu, Zichuan Lin, Miao Liu, Nicholas Leonard, Jiashi Feng, Tat-Seng Chua, Deep Q-Networks for Accelerating the Training of Deep Neural Networks, arXiv:1606.01467,
-
- Emilio Parisotto, Jimmy Lei Ba, Ruslan Salakhutdinov, Actor-Mimic: Deep Multitask and Transfer Reinforcement Learning, ICLR 2016,
-
- Sainbayar Sukhbaatar, Arthur Szlam, Gabriel Synnaeve, Soumith Chintala, Rob Fergus, MazeBase: A Sandbox for Learning from Games, arXiv:1511.07401,
-
- This project contains code to train a model that automatically plays the first level of Super Mario World using only raw pixels as the input (no hand-engineered features).The used technique is deep Q-learning, as described in the (), combined with a .
-
- Tejas D. Kulkarni, Ardavan Saeedi, Simanta Gautam, Samuel J. Gershman, Deep Successor Reinforcement Learning, arXiv:1606.02396,
-
- ViZDoom allows developing AI bots that play Doom using only the visual information (the screen buffer). It is primarily intended for research in machine visual learning, and deep reinforcement learning, in particular.
-
- Marc'Aurelio Ranzato, Sumit Chopra, Michael Auli, Wojciech Zaremba, Sequence Level Training with Recurrent Neural Networks, ICLR 2016,
-
- Implementation of a simple example of Q learning in Torch.
-
- This package is a Reinforcement Learning package written in LUA for Torch.
-
- He He, Jordan Boyd-Graber, Kevin Kwok, Hal Daumé III, Opponent Modeling in Deep Reinforcement Learning, ICML 2016,
### ETC
-
- Alexander M. Rush, Sumit Chopra, Jason Weston, A Neural Attention Model for Abstractive Summarization, EMNLP 2015
-
- Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston, Rob Fergus, End-To-End Memory Networks, arXiv:1503.08895,
-
- Alex Graves, Greg Wayne, Ivo Danihelka, Neural Turing Machines, arXiv:1410.5401
-
- Emily Denton, Soumith Chintala, Arthur Szlam, Rob Fergus, Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks, arXiv:1506.05751
-
- Ke Tran, Arianna Bisazza, Christof Monz, Word Translation Prediction for Morphologically Rich Languages with Bilingual Neural Networks, EMNLP 2014
-
- (#) Tianqi Chen, Ian Goodfellow, Jonathon Shlens, Net2Net: Accelerating Learning via Knowledge Transfer, arXiv:1511.05641
-
- (#) Po-Sen Huang, Xiaodong He, Jianfeng Gao, Li Deng, Alex Acero, Larry Heck, Learning Deep Structured Semantic Models for Web Search using Clickthrough Data, CIKM 2013
-
- (#) Alexander Novikov, Dmitry Podoprikhin, Anton Osokin, Dmitry Vetrov, Tensorizing Neural Networks, NIPS 2015
-
- (#) Elad Hoffer, Nir Ailon, Deep metric learning using Triplet network, arXiv:1412.6622
-
- (#) Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean, Efficient Estimation of Word Representations in Vector Space, ICLR 2013
-
- (#) Florian Schroff, Dmitry Kalenichenko, James Philbin, FaceNet: A Unified Embedding for Face Recognition and Clustering, CVPR 2015
-
- Satoshi Iizuka, Edgar Simo-Serra, Hiroshi Ishikawa, Let there be Color!: Joint End-to-end Learning of Global and Local Image Priors for Automatic Image Colorization with Simultaneous Classification, SIGGRAPH 2016,
-
- Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, Alexei A. Efros, Context Encoders: Feature Learning by Inpainting, CVPR 2016,
-
- Max Jaderberg, Karen Simonyan, Andrew Zisserman, Koray Kavukcuoglu, Spatial Transformer Networks, arXiv:1506.02025,
-
- (#) Jie Fu, Hongyin Luo, Jiashi Feng, Kian Hsiang Low, Tat-Seng Chua, DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks, arXiv:1601.00917,
-
- Rudy Bunel, Alban Desmaison, Pushmeet Kohli, Philip H.S. Torr, M. Pawan Kumar, Adaptive Neural Compilation, arXiv:1605.07969,
-
- (#) Armand Joulin, Edouard Grave, Piotr Bojanowski, Tomas Mikolov, Bag of Tricks for Efficient Text Classification, arXiv:1607.01759,
-
- Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston, Rob Fergus, End-To-End Memory Networks, arXiv:1503.08895,
-
- Diederik P Kingma, Max Welling, Auto-Encoding Variational Bayes, arXiv:1312.6114,
-
- (#) Akira Fukui, Dong Huk Park, Daylen Yang, Anna Rohrbach, Trevor Darrell, Marcus Rohrbach, Multimodal Compact Bilinear Pooling for Visual Question Answering and Visual Grounding,
-
- Implementation of some object detection frameworks in torch. (Fast-RCNN, threaded RCNN, etc.)
-
- Roozbeh Mottaghi, Hessam Bagherinezhad, Mohammad Rastegari, Ali Farhadi, Newtonian Image Understanding: Unfolding the Dynamics of Objects in Static Images, CVPR 2016,
Libraries
Model related
- nn : an easy and modular way to build and train simple or complex neural networks
- dpnn : extensions to the nn lib, more modules
- nnx : extension to the nn lib, experimental neural network modules and criterions
- nninit : weight initialisation schemes
- rnn : Recurrent Neural Network library
- optim : A numeric optimization package for Torch
- dp : a deep learning library designed for streamlining research and development
- nngraph : provides graphical computation for nn library
- nnlr : Add layer-wise learning rate schemes to Torch
- optnet: Memory optimizations for torch neural networks.
- autograd : Autograd automatically differentiates native Torch code.
- torchnet: framework for torch which provides a set of abstractions aiming at encouraging code re-use as well as encouraging modular programming
GPU related
- distro-cl: An OpenCL distribution for Torch
- cutorch : A CUDA backend for Torch
- cudnn : Torch FFI bindings for NVIDIA CuDNN
- fbcunn : Facebook's extensions to torch/cunn
IDE related
- iTorch : IPython kernel for Torch with visualization and plotting
- Lua Development Tools (LDT) : based on Eclipse
- zbs-torch : A lightweight Lua-based IDE for Lua with code completion, syntax highlighting, live coding, remote debugger, and code analyzer
ETC
- fblualib : Facebook libraries and utilities for Lua
- loadcaffe : Load Caffe networks in Torch
- Purdue e-lab lib : A collection of snippets and libraries
- torch-android : Torch for Android
- torch-models : Implementation of state-of-art models in Torch.
- lutorpy : Lutorpy is a libray built for deep learning with torch in python.
- CoreNLP.lua : Lua client for Stanford CoreNLP.
- Torchlib: Data structures and libraries for Torch.
- THFFmpeg: Torch bindings for FFmpeg (reading videos only)
- tunnel: Data Driven Framework for Distributed Computing in Torch 7,
- pytorch: Python wrappers for torch and lua,
- lutorpy: Use torch in python for deep learning.,
- torch-pcl: Point Cloud Library (PCL) bindings for Torch,
- Moses: A Lua utility-belt library for functional programming. It complements the built-in Lua table library, making easier operations on arrays, lists, collections.
Links
转载地址:http://oouhl.baihongyu.com/