Deep Learning (deep learning) study notes finishing series (8)

connect

 

X. Summary and Outlook

1) Deep learning summary

      Deep learning is about algorithms that automatically learn multi-layered (complex) representations of the underlying (implicit) distribution of the data to be modeled. In other words, deep learning algorithms automatically extract low-level or high-level features needed for classification. High-level features, one means that the feature can depend on other features hierarchically (hierarchically), for example: for machine vision, deep learning algorithms learn a low-level representation of it from the original image, such as edge detectors, wavelet filters, etc. , and then build expressions on the basis of these low-level expressions, such as linear or non-linear combinations of these low-level expressions, and then repeat this process, and finally get a high-level expression.

       Deep learning can obtain features that better represent data. At the same time, because the model has many layers and parameters, and the capacity is sufficient, the model has the ability to represent large-scale data, so it is not obvious for features such as images and speech (requires manual design and many There is no intuitive physical meaning), and it can achieve better results on large-scale training data. In addition, from the perspective of pattern recognition features and classifiers, the deep learning framework combines features and classifiers into one framework, and uses data to learn features, which reduces the huge workload of manual feature design in use (this is the current industry. Therefore, not only the effect can be better, but also there are many conveniences to use. Therefore, it is a set of frameworks worthy of attention, and everyone who does ML should pay attention to understand.

       Of course, deep learning itself is not perfect, nor is it a powerful tool for solving any ML problems in the world, and should not be magnified to an omnipotent level.

2) Deep learning future

       There is still a lot of work to be done in deep learning. The current focus is to borrow some methods that can be used in deep learning from the field of machine learning, especially in the field of dimensionality reduction. For example, one of the current work is sparse coding, which uses compressed sensing theory to reduce the dimension of high-dimensional data, so that a vector with very few elements can accurately represent the original high-dimensional signal. Another example is semi-supervised popularity learning, by measuring the similarity of training samples and projecting this similarity in high-dimensional data into a low-dimensional space. Another more inspiring direction is evolutionary programming approaches, which allow conceptual adaptive learning and core architecture changes by minimizing engineering energy.

Deep learning still has many core problems to be solved:

(1) For a given framework, for how many dimensional inputs does it perform well (maybe millions of dimensions in the case of images)?

(2) Which architecture is effective for capturing short-term or long-term temporal dependencies?

(3) How to fuse multiple perceptual information for a given deep learning architecture?

(4) What are the correct mechanisms to enhance a given deep learning architecture to improve its robustness and invariance to distortion and data loss?

(5) Are there other more effective and theoretically based deep model learning algorithms in terms of models?

       Exploring new feature extraction models is worthy of further study. In addition, effective parallel training algorithms are also a direction worthy of research. Current min-batch-based stochastic gradient optimization algorithms are difficult to train in parallel on multiple computers. The usual approach is to use a graphics processing unit to speed up the learning process. However, a single machine GPU is not suitable for large-scale data recognition or similar task datasets. In terms of deep learning application expansion, how to make full use of deep learning to enhance the performance of traditional learning algorithms is still the focus of current research in various fields.

 

11. References and Deep Learning learning resources ( continuously updated... )

       The first is the Weibo of the big cow in the field of machine learning: @玉凯_西 Erqi migrant workers; @teacher wood; @梁斌penny; @张东_machine learning; @dengkan; @big datapidong;

(1)Deep Learning

http://deeplearning.net/

(2)Deep Learning Methods for Vision

http://cs.nyu.edu/~fergus/tutorials/deep_learning_cvpr12/

(3)Neural Network for Recognition of Handwritten Digits[Project]

http://www.codeproject.com/Articles/16650/Neural-Network-for-Recognition-of-Handwritten-Digi

(4)Training a deep autoencoder or a classifier on MNIST digits

http://www.cs.toronto.edu/~hinton/MatlabForSciencePaper.html

(5)Ersatz:deep neural networks in the cloud

http://www.ersatz1.com/

(6)Deep Learning

http://www.cs.nyu.edu/~yann/research/deep/

(7)Invited talk "A Tutorial on Deep Learning" by Dr. Kai Yu (余凯)

http://vipl.ict.ac.cn/News/academic-report-tutorial-deep-learning-dr-kai-yu

(8)CNN - Convolutional neural network class

http://www.mathworks.cn/matlabcentral/fileexchange/24291

(9)Yann LeCun's Publications

http://yann.lecun.com/exdb/publis/index.html#lecun-98

(10) LeNet-5, convolutional neural networks

http://yann.lecun.com/exdb/lenet/index.html

(11) Deep Learning 大牛Geoffrey E. Hinton's HomePage

http://www.cs.toronto.edu/~hinton/

(12)Sparse coding simulation software[Project]

http://redwood.berkeley.edu/bruno/sparsenet/

(13)Andrew Ng's homepage

http://robotics.stanford.edu/~ang/

(14)stanford deep learning tutorial

http://deeplearning.stanford.edu/wiki/index.php/UFLDL_Tutorial

(15)「深度神经网络」(deep neural network)具体是怎样工作的

http://www.zhihu.com/question/19833708?group_id=15019075#1657279

(16)A shallow understanding on deep learning

http://blog.sina.com.cn/s/blog_6ae183910101dw2z.html

(17)Bengio's Learning Deep Architectures for AI

 http://www.iro.umontreal.ca/~bengioy/papers/ftml_book.pdf

(18)andrew ng's talk video:

http://techtalks.tv/talks/machine-learning-and-ai-via-brain-simulations/57862/

(19)cvpr 2012 tutorial:

http://cs.nyu.edu/~fergus/tutorials/deep_learning_cvpr12/tutorial_p2_nnets_ranzato_short.pdf

(20)Andrew ng清华报告听后感

http://blog.sina.com.cn/s/blog_593af2a70101bqyo.html

(21)Kai Yu:CVPR12 Tutorial on Deep Learning Sparse Coding

(22)Honglak Lee:Deep Learning Methods for Vision

(23)Andrew Ng :Machine Learning and AI via Brain simulations

(24)Deep Learning 【2,3】

http://blog.sina.com.cn/s/blog_46d0a3930101gs5h.html

(25)deep learning这件小事……

http://blog.sina.com.cn/s/blog_67fcf49e0101etab.html

(26)Yoshua Bengio, U. Montreal:Learning Deep Architectures

(27)Kai Yu:A Tutorial on Deep Learning

(28)Marc'Aurelio Ranzato:NEURAL NETS FOR VISION

(29)Unsupervised feature learning and deep learning

http://blog.csdn.net/abcjennifer/article/details/7804962

(30)机器学习前沿热点–Deep Learning

http://elevencitys.com/?p=1854

(31)机器学习——深度学习(Deep Learning)

http://blog.csdn.net/abcjennifer/article/details/7826917

(32)卷积神经网络

http://wenku.baidu.com/view/cd16fb8302d276a200292e22.html

(33)浅谈Deep Learning的基本思想和方法

http://blog.csdn.net/xianlingmao/article/details/8478562

(34)深度神经网络

http://blog.csdn.net/txdb/article/details/6766373

(35)Google的猫脸识别:人工智能的新突破

http://www.36kr.com/p/122132.html

(36)余凯,深度学习-机器学习的新浪潮,Technical News程序天下事

http://blog.csdn.net/datoubo/article/details/8577366

(37)Geoffrey Hinton:UCLTutorial on: Deep Belief Nets

(38)Learning Deep Boltzmann Machines

http://web.mit.edu/~rsalakhu/www/DBM.html

(39)Efficient Sparse Coding Algorithm

http://blog.sina.com.cn/s/blog_62af19190100gux1.html

(40)Itamar Arel, Derek C. Rose, and Thomas P. Karnowski: Deep Machine Learning—A New Frontier in Artificial Intelligence Research

(41)Francis Quintal Lauzon:An introduction to deep learning

(42)Tutorial on Deep Learning and Applications

(43)Boltzmann神经网络模型与学习算法

http://wenku.baidu.com/view/490dcf748e9951e79b892785.html

(44)Deep Learning 和 Knowledge Graph 引爆大数据革命

http://blog.sina.com.cn/s/blog_46d0a3930101fswl.html

(45)……

Guess you like

Origin http://43.154.161.224:23101/article/api/json?id=324932562&siteId=291194637