Skip to content

Commit

Permalink
Update 05-evaluate-predict-cnn.md
Browse files Browse the repository at this point in the history
Deleted info on neurons. Agree that there is plenty in ep03. Reworded intro to hyper parameters with metaphor of radio dials for tuning.
  • Loading branch information
isaac-jennings authored Oct 12, 2023
1 parent 5b4ebf8 commit afc2116
Showing 1 changed file with 3 additions and 10 deletions.
13 changes: 3 additions & 10 deletions episodes/05-evaluate-predict-cnn.md
Original file line number Diff line number Diff line change
Expand Up @@ -254,24 +254,17 @@ NOTE your output will vary!
:::::::::::::::::::::::::::::::::
::::::::::::::::::::::::::::::::::::::::::::::::


### Step 9. Tune hyperparameters

Recall the following from Episode 1:

#### What are hyperparameters?

Hyperparameters are all the parameters set by the person configuring the machine learning instead of those learned by the algorithm itself. These hyperparameters can include the learning rate, the number of layers in the network, the number of neurons per layer, and many more. Hyperparameter tuning refers to the process of systematically searching for the best combination of hyperparameters that will optimize the model's performance.

##### Wait, what is a neuron again?

The neurons referred to in this context are an artifical, mathematical representation of the most basic building blocks of an artificial neural network. They are modelled using inspiration from the biological design of neurons in our brains.

Each neuron could be considered its own, local mathematical model. With one or more input vectors (x_0 through x_m) and weights (w_k0 and w_km), where weights are a quantitative representation of the importance of a variable. Input values are multiplied by weights and summed, before being passed through an **activation function**. The result of the activiation function calculation is compared against a **threshold**. If the output exceeds this threshold, a neuron is then **activated**. This means the output is passed to the next node in the network, repeating until complete. It is this process which lends to describing neural networks as **feedforward networks**.
Hyperparameters are the parameters set by the person configuring the model instead of those learned by the algorithm itself. Like the dials on a radio which are *tuned* to the best frequency, hyperparameters can be *tuned* to the best combination for a given model and context.

#### Back to hyperparameters.
These hyperparameters can include the learning rate, the number of layers in the network, the number of neurons per layer, and many more. The tuning process is systematic searching for the best combination of hyperparameters that will optimize the model's performance.

Hyperparameters are all parameters set by the person configuring the machine learning instead of those learned by the algorithm itself. It might be necessary to adjust these and re-run the training many times before we are happy with the result.
In some cases, it might be necessary to adjust these and re-run the training many times before we are happy with the result.

Table 1. List of some of the hyperparameters to tune and when.

Expand Down

0 comments on commit afc2116

Please sign in to comment.