Deep Learning
paperCredibility Rating
Gold standard. Rigorous peer review, high editorial standards, and strong institutional reputation.
Rating inherited from publication venue: Nature
A foundational Nature review article explaining deep learning fundamentals, architectures (CNNs, RNNs), and applications across domains; important for understanding the technical foundations of modern AI systems relevant to AI safety research.
Paper Details
Metadata
Summary
This Nature article provides a comprehensive overview of deep learning, explaining how computational models with multiple processing layers can learn hierarchical representations of data. The paper highlights that deep learning has dramatically advanced performance in speech recognition, visual object recognition, object detection, drug discovery, and genomics. It describes key techniques including backpropagation for training neural networks, convolutional neural networks (CNNs) for image and audio processing, and recurrent neural networks (RNNs) for sequential data like text and speech.
Cited by 1 page
| Page | Type | Quality |
|---|---|---|
| Geoffrey Hinton | Person | 42.0 |
Cached Content Preview
Deep learning | Nature
Skip to main content
Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain
the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in
Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles
and JavaScript.
Advertisement
Subjects
Computer science
Mathematics and computing
Abstract
Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.
Access through your institution
Buy or subscribe
This is a preview of subscription content, access via your institution
Access options
Access through your institution
Subscribe to this journal
Receive 51 print issues and online access
$199.00 per year
only $3.90 per issue
Learn more
Buy this article
Purchase on SpringerLink
Instant access to the full article PDF.
USD 39.95
Prices may be subject to local taxes which are calculated during checkout
Figure 1: Multilayer neural networks and backpropagation. Figure 2: Inside a convolutional network. Figure 3: From image to text. Figure 4: Visualizing the learned word vectors. Figure 5: A recurrent neural network and the unfolding in time of the computation involved in its forward computation.
Similar content being viewed by others
Novel applications of Convolutional Neural Networks in the age of Transformers
A
... (truncated, 29 KB total)e4c6d6e59da16fc4 | Stable ID: ZTk2MTAwN2