Skip to content
Longterm Wiki
Back

Learning representations by back-propagating errors

paper
2002·Nature(peer-reviewed)·nature.com/articles/323533a0

Authors

David E. Rumelhart·Geoffrey E. Hinton·Ronald J. Williams

Credibility Rating

5/5
Gold(5)

Gold standard. Rigorous peer review, high editorial standards, and strong institutional reputation.

Rating inherited from publication venue: Nature

Seminal 1986 Nature paper introducing backpropagation, a foundational algorithm enabling deep neural networks; relevant to AI safety as it established the learning mechanisms underlying modern deep learning systems whose alignment and interpretability are central safety concerns.

Paper Details

Citations
82
Year
2002
Methodology
book-chapter
Categories
Cognitive Modeling

Metadata

journal articleprimary source

Summary

This seminal 1986 Nature paper introduces backpropagation, a learning algorithm for multi-layer neural networks that iteratively adjusts connection weights to minimize the difference between actual and desired outputs. The key innovation is that hidden units in the network automatically learn to represent important features of the task domain through this error-driven learning process. This capability to discover useful internal representations distinguishes backpropagation from earlier methods like the perceptron, enabling neural networks to learn complex, non-linear relationships in data.

Cited by 1 page

PageTypeQuality
Geoffrey HintonPerson42.0

Cached Content Preview

HTTP 200Fetched Mar 15, 20267 KB
Learning representations by back-propagating errors | Nature 
 
 
 

 
 

 
 

 

 
 
 
 

 

 
 
 
 
 
 

 
 
 
 
 
 

 
 

 
 
 
 
 
 
 
 
 
 
 

 
 

 

 

 
 

 
 
 

 
 

 
 
 

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

 
 
 
 
 
 
 

 
 
 

 
 Skip to main content 

 
 
 
 Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain
 the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in
 Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles
 and JavaScript.

 
 

 

 

 
 
 

 
 
 Advertisement

 
 
 
 
 
 
 
 
 
 
 

 
 
 
 

 

 
 
 
 

 

 

 
 
 
 
 
 
 
 
 

 
 
 
 

 
 

 
 

 
 Abstract

 We describe a new learning procedure, back-propagation, for networks of neurone-like units. The procedure repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector. As a result of the weight adjustments, internal ‘hidden’ units which are not part of the input or output come to represent important features of the task domain, and the regularities in the task are captured by the interactions of these units. The ability to create useful new features distinguishes back-propagation from earlier, simpler methods such as the perceptron-convergence procedure 1 .

 

 
 
 
 
 
 
 
 
 
 Access through your institution 
 
 
 
 
 
 
 
 Buy or subscribe 
 
 
 
 
 
 

 
 
 

 
 
 
 
 
 
 This is a preview of subscription content, access via your institution 

 
 
 

 

 Access options

 

 
 
 
 
 
 
 
 Access through your institution 
 
 
 
 
 
 

 

 
 
 
 
 Subscribe to this journal

 
 Receive 51 print issues and online access
 

 
 
 $199.00 per year

 only $3.90 per issue

 
 
 
 
 Learn more 
 
 
 
 Buy this article

 Purchase on SpringerLink 
 Instant access to the full article PDF. 
 USD 39.95 

 Prices may be subject to local taxes which are calculated during checkout

 
 

 
 
 
 

 
 
 

 
 
 
 
 Similar content being viewed by others

 
 
 
 
 
 
 
 
 
 Inferring neural activity before plasticity as a foundation for learning beyond backpropagation
 
 

 
 Article 
 Open access 
 03 January 2024 
 
 
 
 
 
 
 
 
 
 
 
 
 Vectorized instructive signals in cortical dendrites
 
 

 
 Article 
 Open access 
 25 February 2026 
 
 
 
 
 
 
 
 
 
 
 
 
 Neural heterogeneity promotes robust learning
 
 

 
 Article 
 Open access 
 04 October 2021 
 
 
 
 
 
 
 
 
 
 
 
 
 

 
 References

 Rosenblatt, F. Principles of Neurodynamics (Spartan, Washington, DC, 1961).

 
 Google Scholar  
 

 Minsky, M. L. & Papert, S. Perceptrons (MIT, Cambridge, 1969).

 MATH  
 
 Google Scholar  
 

 Le Cun, Y. Proc. Cognitiva 85 , 599–604 (1985).

 
 Google Scholar  
 

 Rumelhart, D. E., Hinton, G. E. & Williams,

... (truncated, 7 KB total)
Resource ID: fbbf2cbc86a8b7b4 | Stable ID: NDQ0MmYxZj