17.5 Problems
Problem 1. Let y ∈ℝn be an arbitrary vector. The general version of the famous mean-squared error is defined by

Compute its gradient and implement it using the MultivariateFunction base class!
Problem 2. Let f : ℝ2 →ℝ be the function defined by

Does f have a local extremum in x = (0,0)?
Problem 3. Use the previously implemented gradient_descent function to find the minimum of

Experiment with various learning rates and initial values!
Problem 4. In the problem section of Chapter 13, we saw the improved version of gradient descent, called gradient descent with momentum. We can do the same in multiple variables: define

where d0 = 0 and x0 is arbitrary. Implement it!
Join our community on Discord
Read this book alongside other users, Machine Learning experts, and the author himself. Ask questions, provide solutions to other readers, chat with the author via Ask Me Anything sessions, and much more. Scan the QR code or visit the...