About 20,800 results
Open links in new tab
  1. Backpropagation - Wikipedia

    In machine learning, backpropagation is a gradient computation method commonly used for training a neural network in computing parameter updates. It is an efficient application of the chain rule to …

  2. Backpropagation in Neural Network - GeeksforGeeks

    Feb 9, 2026 · Backpropagation, short for Backward Propagation of Errors, is a key algorithm used to train neural networks by minimizing the difference between predicted and actual outputs.

  3. 14 Backpropagation – Foundations of Computer Vision

    This is the whole trick of backpropagation: rather than computing each layer’s gradients independently, observe that they share many of the same terms, so we might as well calculate each shared term …

  4. What is backpropagation? - IBM

    Backpropagation is a machine learning technique essential to the optimization of artificial neural networks. It facilitates the use of gradient descent algorithms to update network weights, which is …

  5. Backpropagation, explained from first principles

    Mar 22, 2026 · Backpropagation is one of those terms that gets thrown around so much in AI that people assume everyone already understands it. But most explanations stop at “the network adjusts its …

  6. Understanding Backpropagation in Deep Learning

    May 30, 2025 · Backpropagation, often referred to as “backward propagation of errors,” is the cornerstone of training deep neural networks. It is a supervised learning algorithm that optimizes the …

  7. Backpropagation | Brilliant Math & Science Wiki

    Backpropagation, short for "backward propagation of errors," is an algorithm for supervised learning of artificial neural networks using gradient descent. Given an artificial neural network and an error …

  8. 7.2 Backpropagation - Principles of Data Science | OpenStax

    Backpropagation is a supervised learning algorithm, meaning that it trains on data that has already been classified (see What Is Machine Learning? for more about supervised learning in general).

  9. Backpropagation — How Neural Networks Learn, Step by Step -

    Backpropagation solves this by using the chain rule to efficiently compute all gradients in a single backward pass. It starts at the output layer and propagates gradients backward — each layer’s …

  10. The Chain Rule in Deep Learning: What Backpropagation Is Really ...

    A plain-English guide to the chain rule in deep learning. Learn how indirect dependence works, why gradients add over multiple paths, and how Jacobian-style matrices make backpropagation efficient.