From fbf67c881a59c21b5b73d1047fae58689257fcc6 Mon Sep 17 00:00:00 2001 From: Charles Iliya Krempeaux Date: Sat, 16 Dec 2023 23:04:45 -0800 Subject: [PATCH] correction --- 2014/12/29/deep-learning-in-a-nutshell/index.xhtml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/2014/12/29/deep-learning-in-a-nutshell/index.xhtml b/2014/12/29/deep-learning-in-a-nutshell/index.xhtml index 1395bcb..40976c2 100644 --- a/2014/12/29/deep-learning-in-a-nutshell/index.xhtml +++ b/2014/12/29/deep-learning-in-a-nutshell/index.xhtml @@ -95,7 +95,7 @@
The neuron we want to train for the Dining Hall Problem
-
+

Let's say we have a bunch of training examples. Then we can calculate what the neural network will output on the training example using the simple formula in the diagram. We want to train the neuron so that we pick the optimal weights possible - the weights that minimize the errors we make on the training examples. In this case, let's say we want to minimize the square error over all of the training examples that we encounter. More formally, if we know that is the true answer for the training example and is the value computed by the neural network, we want to minimize the value of the error function :

Now at this point you might be thinking, wait up... Why do we need to bother ourselves with this error function nonsense when we have a bunch of variables (weights) and we have a set of equations (one for each training example)? Couldn't we just solve this problem by setting up a system of linear system of equations? That would automaically give us an error of zero assuming that we have a consistent set of training examples, right?