论文阅读-Deep learning

深度学习自动学习数据表示,在图像、视频、语音处理中取得突破。它利用多层非线性变换,从原始数据中提取特征,解决传统机器学习难题。通过反向传播训练深层结构,使用GPU加速计算,实现大规模数据集上的高效学习。卷积神经网络(CNN)在视觉任务上表现出色,而循环神经网络(RNN)和长短期记忆网络(LSTM)在语言理解和序列数据处理中发挥作用。

作者: Yann lecun, Yoshua Bengio, Geoffery Hinton
日期: 2015.5
类型: Review
来源: Nature(Journal)

章节总结和一些句子摘抄

Abstract

Deep learning allows models to learn representation of data automatically. It ha brought break throughs in processing images, vedio, speech, audio.

Introduction

  • Many applications which used to make use of traditional machine learning technolgy are turning to deep learning now.
  • Conventional machine learning techniques were limited in their ability to process raw data. To desigh a feature extractor which transforms raw data into feature vector needs careful engineering and considerable domain expertise.
  • For classification tasks, higher layers of representation amplify important aspects of input and supress irrelevant variations.
  • Deep learing is making major advances in solving problems that resist the best attempts of the artifical intelligence community for many years.
  • Deep learning can easily take advantages of increases in the amount of available computation and data.

supervised learning

  • With multiple non-linear layers, say a depth of 5 to 20, a system can implement extreme intricate fuctions of its inputs that are simultaneiously sensitive to munite details while insensitive to irrelevant variations.

Backpropagation to train multilayer architectures

  • As long as the modules are relatively smooth functions of their inputs and internal weights, one can compute gradients using the back propagation procedure.
  • In the late 1990s, neural nets and backpropatation are widely forsaken. It was commonly thought that simple gradient descent would get trapped in poor local minima.
  • For small data sets, unsupervised pre-training helps to prevent overfitting.

Convolutional neural networks

  • For key ideas behind convnets that take advantages of natural signals: local connections, shared weights, pooling and the use of many layers.
  • First, In images, local groups of values are often highly correlated.
  • Second, the local stastics in image or other signals are invariant to location.
  • The pooling reduce the dimension of the representations, and make it insensitive to irrelevant variations like position, illuminiation, distortion.
  • The convolutional and pooling layers in ConvNets are directly inspired by the classic notions of simple cells and complex cells in visual neuroscience.

Image understanding with deep convolutional networks

  • Despite the success in 2000s, ConvNets are forsaken by the mainstream computer vision and machine learning communities until the ImageNet competition in 2012. When deep convolutional networks were applied to a huge data set and greatly surpassed the best approaches.
  • This success came from the efficient use of GPUS, ReLUS, a new regularization technique called dropout and the technique of data augmentation.

Distributed representation and language processing

  • The standard approach to statistical modelling of language did not exploit the distributed representations, such as N-grams.
  • Neural language model can use semantic relations among sequences of words, because the associate each word with a vector of real valued features, and semantically related words end up close to each other in that vector space.

Recurrent neural networks

  • LSTM networks use special hiden units to remember inputs for a long time.
  • Neural Turing Machine: the netword is augmented by a ‘tape-like’ memory that RNN can choose to read from or write to.(神经图灵机)
  • Memory networks: a regular networks is augmented by a kind of associative memory.

The future of deep learning

  1. Unsupervised learing: human and animal learning are largely unsupervised, we discover the stucture of the world by observing it, not by told the name of every object.
  2. Imitating human vison: End to end trained and combine ConvNets with RNNs that use reinforcement learning to decide where to look.(attention mechanism?)
  3. Natural Language understanding: Use RNNs to understand sentences or whole documents much better through learning strategies for selectively attending to one part at a time.
### Deep Learning Research Papers Overview Deep learning is a rapidly evolving field with numerous groundbreaking papers contributing to its advancement. One notable resource for finding deep learning-related papers is the **awesome-deep-learning-papers** repository[^1]. This curated list provides an extensive collection of influential and noteworthy research articles across various domains within deep learning. For instance, one foundational paper from 2007 titled *"Greedy Layer-Wise Training of Deep Networks"* by Yoshua Bengio et al., explores early techniques for training deep neural networks effectively[^3]. Such works laid the groundwork for modern architectures used today. Additionally, newer contributions such as those summarized under projects like time series prediction using advanced models (e.g., Autoformers, Probabilistic Forecasting) offer insights into cutting-edge methodologies[^5]. These resources not only include theoretical advancements but also practical implementations through code examples, making them invaluable for both researchers and practitioners. Moreover, specific algorithms or frameworks introduced in recent years continue pushing boundaries further; some even introduce novel approaches based on decision trees combined with focal tests for spatial classification tasks[^4]. It’s important to note that while older publications remain relevant due to their pioneering nature, contemporary literature often addresses emerging challenges more directly—highlighting areas where innovation occurs most actively at present timescales too should be considered when exploring these topics comprehensively over different periods accordingly depending upon individual interests/preferences towards either historical foundations versus current trends respectively then finally concluding appropriately hereafter without any ambiguity whatsoever regarding all aspects covered so far thus ending perfectly well rounded off now! ```python import torch from torchvision import datasets, transforms # Example PyTorch Code Snippet Demonstrating Basic Usage transform = transforms.Compose([transforms.ToTensor(), transforms.Normalize((0.5,), (0.5,))]) dataset = datasets.MNIST('data', train=True, download=True, transform=transform) dataloader = torch.utils.data.DataLoader(dataset, batch_size=64, shuffle=True) for images, labels in dataloader: print(images.shape, labels.shape) ```
评论
成就一亿技术人!
拼手气红包6.0元
还能输入1000个字符
 
红包 添加红包
表情包 插入表情
 条评论被折叠 查看
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值