Skip to content

Commit 610cdc0

Browse files
committed
update week 37
1 parent 69d0387 commit 610cdc0

59 files changed

Lines changed: 4054 additions & 2043 deletions

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

doc/pub/week37/html/._week37-bs000.html

Lines changed: 51 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -123,6 +123,28 @@
123123
2,
124124
None,
125125
'second-moment-of-the-gradient'),
126+
('Challenge: Choosing a Fixed Learning Rate',
127+
2,
128+
None,
129+
'challenge-choosing-a-fixed-learning-rate'),
130+
('Motivation for Adaptive Step Sizes',
131+
2,
132+
None,
133+
'motivation-for-adaptive-step-sizes'),
134+
('Derivation of the AdaGrad Algorithm',
135+
2,
136+
None,
137+
'derivation-of-the-adagrad-algorithm'),
138+
('AdaGrad Update Rule Derivation',
139+
2,
140+
None,
141+
'adagrad-update-rule-derivation'),
142+
('AdaGrad Properties', 2, None, 'adagrad-properties'),
143+
('RMSProp: Adaptive Learning Rates',
144+
2,
145+
None,
146+
'rmsprop-adaptive-learning-rates'),
147+
('Adam Optimizer', 2, None, 'adam-optimizer'),
126148
('RMS prop', 2, None, 'rms-prop'),
127149
('"ADAM optimizer":"https://arxiv.org/abs/1412.6980"',
128150
2,
@@ -247,7 +269,7 @@
247269
<!-- navigation toc: --> <li><a href="._week37-bs012.html#program-example-for-gradient-descent-with-ridge-regression" style="font-size: 80%;">Program example for gradient descent with Ridge Regression</a></li>
248270
<!-- navigation toc: --> <li><a href="._week37-bs013.html#using-gradient-descent-methods-limitations" style="font-size: 80%;">Using gradient descent methods, limitations</a></li>
249271
<!-- navigation toc: --> <li><a href="._week37-bs014.html#improving-gradient-descent-with-momentum" style="font-size: 80%;">Improving gradient descent with momentum</a></li>
250-
<!-- navigation toc: --> <li><a href="._week37-bs043.html#same-code-but-now-with-momentum-gradient-descent" style="font-size: 80%;">Same code but now with momentum gradient descent</a></li>
272+
<!-- navigation toc: --> <li><a href="._week37-bs050.html#same-code-but-now-with-momentum-gradient-descent" style="font-size: 80%;">Same code but now with momentum gradient descent</a></li>
251273
<!-- navigation toc: --> <li><a href="._week37-bs016.html#overview-video-on-stochastic-gradient-descent-sgd" style="font-size: 80%;">Overview video on Stochastic Gradient Descent (SGD)</a></li>
252274
<!-- navigation toc: --> <li><a href="._week37-bs017.html#batches-and-mini-batches" style="font-size: 80%;">Batches and mini-batches</a></li>
253275
<!-- navigation toc: --> <li><a href="._week37-bs018.html#pros-and-cons" style="font-size: 80%;">Pros and cons</a></li>
@@ -268,26 +290,33 @@
268290
<!-- navigation toc: --> <li><a href="._week37-bs033.html#more-on-momentum-based-approaches" style="font-size: 80%;">More on momentum based approaches</a></li>
269291
<!-- navigation toc: --> <li><a href="._week37-bs034.html#momentum-parameter" style="font-size: 80%;">Momentum parameter</a></li>
270292
<!-- navigation toc: --> <li><a href="._week37-bs035.html#second-moment-of-the-gradient" style="font-size: 80%;">Second moment of the gradient</a></li>
271-
<!-- navigation toc: --> <li><a href="._week37-bs036.html#rms-prop" style="font-size: 80%;">RMS prop</a></li>
272-
<!-- navigation toc: --> <li><a href="._week37-bs037.html#adam-optimizer-https-arxiv-org-abs-1412-6980" style="font-size: 80%;">"ADAM optimizer":"https://arxiv.org/abs/1412.6980"</a></li>
273-
<!-- navigation toc: --> <li><a href="._week37-bs038.html#algorithms-and-codes-for-adagrad-rmsprop-and-adam" style="font-size: 80%;">Algorithms and codes for Adagrad, RMSprop and Adam</a></li>
274-
<!-- navigation toc: --> <li><a href="._week37-bs038.html#adagrad-algorithm-taken-from-goodfellow-et-al-https-www-deeplearningbook-org-contents-optimization-html" style="font-size: 80%;">AdaGrad algorithm, taken from "Goodfellow et al":"https://www.deeplearningbook.org/contents/optimization.html"</a></li>
275-
<!-- navigation toc: --> <li><a href="._week37-bs038.html#rmsprop-algorithm-taken-from-goodfellow-et-al-https-www-deeplearningbook-org-contents-optimization-html" style="font-size: 80%;">RMSProp algorithm, taken from "Goodfellow et al":"https://www.deeplearningbook.org/contents/optimization.html"</a></li>
276-
<!-- navigation toc: --> <li><a href="._week37-bs038.html#adam-algorithm-taken-from-goodfellow-et-al-https-www-deeplearningbook-org-contents-optimization-html" style="font-size: 80%;">ADAM algorithm, taken from "Goodfellow et al":"https://www.deeplearningbook.org/contents/optimization.html"</a></li>
277-
<!-- navigation toc: --> <li><a href="._week37-bs039.html#practical-tips" style="font-size: 80%;">Practical tips</a></li>
278-
<!-- navigation toc: --> <li><a href="._week37-bs040.html#sneaking-in-automatic-differentiation-using-autograd" style="font-size: 80%;">Sneaking in automatic differentiation using Autograd</a></li>
279-
<!-- navigation toc: --> <li><a href="._week37-bs043.html#same-code-but-now-with-momentum-gradient-descent" style="font-size: 80%;">Same code but now with momentum gradient descent</a></li>
280-
<!-- navigation toc: --> <li><a href="._week37-bs042.html#including-stochastic-gradient-descent-with-autograd" style="font-size: 80%;">Including Stochastic Gradient Descent with Autograd</a></li>
281-
<!-- navigation toc: --> <li><a href="._week37-bs043.html#same-code-but-now-with-momentum-gradient-descent" style="font-size: 80%;">Same code but now with momentum gradient descent</a></li>
282-
<!-- navigation toc: --> <li><a href="._week37-bs044.html#but-none-of-these-can-compete-with-newton-s-method" style="font-size: 80%;">But none of these can compete with Newton's method</a></li>
283-
<!-- navigation toc: --> <li><a href="._week37-bs045.html#similar-second-order-function-now-problem-but-now-with-adagrad" style="font-size: 80%;">Similar (second order function now) problem but now with AdaGrad</a></li>
284-
<!-- navigation toc: --> <li><a href="._week37-bs046.html#rmsprop-for-adaptive-learning-rate-with-stochastic-gradient-descent" style="font-size: 80%;">RMSprop for adaptive learning rate with Stochastic Gradient Descent</a></li>
285-
<!-- navigation toc: --> <li><a href="._week37-bs047.html#and-finally-adam-https-arxiv-org-pdf-1412-6980-pdf" style="font-size: 80%;">And finally "ADAM":"https://arxiv.org/pdf/1412.6980.pdf"</a></li>
286-
<!-- navigation toc: --> <li><a href="._week37-bs048.html#material-for-the-lab-sessions" style="font-size: 80%;">Material for the lab sessions</a></li>
287-
<!-- navigation toc: --> <li><a href="._week37-bs049.html#reminder-on-different-scaling-methods" style="font-size: 80%;">Reminder on different scaling methods</a></li>
288-
<!-- navigation toc: --> <li><a href="._week37-bs050.html#functionality-in-scikit-learn" style="font-size: 80%;">Functionality in Scikit-Learn</a></li>
289-
<!-- navigation toc: --> <li><a href="._week37-bs051.html#more-preprocessing" style="font-size: 80%;">More preprocessing</a></li>
290-
<!-- navigation toc: --> <li><a href="._week37-bs052.html#frequently-used-scaling-functions" style="font-size: 80%;">Frequently used scaling functions</a></li>
293+
<!-- navigation toc: --> <li><a href="._week37-bs036.html#challenge-choosing-a-fixed-learning-rate" style="font-size: 80%;">Challenge: Choosing a Fixed Learning Rate</a></li>
294+
<!-- navigation toc: --> <li><a href="._week37-bs037.html#motivation-for-adaptive-step-sizes" style="font-size: 80%;">Motivation for Adaptive Step Sizes</a></li>
295+
<!-- navigation toc: --> <li><a href="._week37-bs038.html#derivation-of-the-adagrad-algorithm" style="font-size: 80%;">Derivation of the AdaGrad Algorithm</a></li>
296+
<!-- navigation toc: --> <li><a href="._week37-bs039.html#adagrad-update-rule-derivation" style="font-size: 80%;">AdaGrad Update Rule Derivation</a></li>
297+
<!-- navigation toc: --> <li><a href="._week37-bs040.html#adagrad-properties" style="font-size: 80%;">AdaGrad Properties</a></li>
298+
<!-- navigation toc: --> <li><a href="._week37-bs041.html#rmsprop-adaptive-learning-rates" style="font-size: 80%;">RMSProp: Adaptive Learning Rates</a></li>
299+
<!-- navigation toc: --> <li><a href="._week37-bs042.html#adam-optimizer" style="font-size: 80%;">Adam Optimizer</a></li>
300+
<!-- navigation toc: --> <li><a href="._week37-bs043.html#rms-prop" style="font-size: 80%;">RMS prop</a></li>
301+
<!-- navigation toc: --> <li><a href="._week37-bs044.html#adam-optimizer-https-arxiv-org-abs-1412-6980" style="font-size: 80%;">"ADAM optimizer":"https://arxiv.org/abs/1412.6980"</a></li>
302+
<!-- navigation toc: --> <li><a href="._week37-bs045.html#algorithms-and-codes-for-adagrad-rmsprop-and-adam" style="font-size: 80%;">Algorithms and codes for Adagrad, RMSprop and Adam</a></li>
303+
<!-- navigation toc: --> <li><a href="._week37-bs045.html#adagrad-algorithm-taken-from-goodfellow-et-al-https-www-deeplearningbook-org-contents-optimization-html" style="font-size: 80%;">AdaGrad algorithm, taken from "Goodfellow et al":"https://www.deeplearningbook.org/contents/optimization.html"</a></li>
304+
<!-- navigation toc: --> <li><a href="._week37-bs045.html#rmsprop-algorithm-taken-from-goodfellow-et-al-https-www-deeplearningbook-org-contents-optimization-html" style="font-size: 80%;">RMSProp algorithm, taken from "Goodfellow et al":"https://www.deeplearningbook.org/contents/optimization.html"</a></li>
305+
<!-- navigation toc: --> <li><a href="._week37-bs045.html#adam-algorithm-taken-from-goodfellow-et-al-https-www-deeplearningbook-org-contents-optimization-html" style="font-size: 80%;">ADAM algorithm, taken from "Goodfellow et al":"https://www.deeplearningbook.org/contents/optimization.html"</a></li>
306+
<!-- navigation toc: --> <li><a href="._week37-bs046.html#practical-tips" style="font-size: 80%;">Practical tips</a></li>
307+
<!-- navigation toc: --> <li><a href="._week37-bs047.html#sneaking-in-automatic-differentiation-using-autograd" style="font-size: 80%;">Sneaking in automatic differentiation using Autograd</a></li>
308+
<!-- navigation toc: --> <li><a href="._week37-bs050.html#same-code-but-now-with-momentum-gradient-descent" style="font-size: 80%;">Same code but now with momentum gradient descent</a></li>
309+
<!-- navigation toc: --> <li><a href="._week37-bs049.html#including-stochastic-gradient-descent-with-autograd" style="font-size: 80%;">Including Stochastic Gradient Descent with Autograd</a></li>
310+
<!-- navigation toc: --> <li><a href="._week37-bs050.html#same-code-but-now-with-momentum-gradient-descent" style="font-size: 80%;">Same code but now with momentum gradient descent</a></li>
311+
<!-- navigation toc: --> <li><a href="._week37-bs051.html#but-none-of-these-can-compete-with-newton-s-method" style="font-size: 80%;">But none of these can compete with Newton's method</a></li>
312+
<!-- navigation toc: --> <li><a href="._week37-bs052.html#similar-second-order-function-now-problem-but-now-with-adagrad" style="font-size: 80%;">Similar (second order function now) problem but now with AdaGrad</a></li>
313+
<!-- navigation toc: --> <li><a href="._week37-bs053.html#rmsprop-for-adaptive-learning-rate-with-stochastic-gradient-descent" style="font-size: 80%;">RMSprop for adaptive learning rate with Stochastic Gradient Descent</a></li>
314+
<!-- navigation toc: --> <li><a href="._week37-bs054.html#and-finally-adam-https-arxiv-org-pdf-1412-6980-pdf" style="font-size: 80%;">And finally "ADAM":"https://arxiv.org/pdf/1412.6980.pdf"</a></li>
315+
<!-- navigation toc: --> <li><a href="._week37-bs055.html#material-for-the-lab-sessions" style="font-size: 80%;">Material for the lab sessions</a></li>
316+
<!-- navigation toc: --> <li><a href="._week37-bs056.html#reminder-on-different-scaling-methods" style="font-size: 80%;">Reminder on different scaling methods</a></li>
317+
<!-- navigation toc: --> <li><a href="._week37-bs057.html#functionality-in-scikit-learn" style="font-size: 80%;">Functionality in Scikit-Learn</a></li>
318+
<!-- navigation toc: --> <li><a href="._week37-bs058.html#more-preprocessing" style="font-size: 80%;">More preprocessing</a></li>
319+
<!-- navigation toc: --> <li><a href="._week37-bs059.html#frequently-used-scaling-functions" style="font-size: 80%;">Frequently used scaling functions</a></li>
291320

292321
</ul>
293322
</li>
@@ -341,7 +370,7 @@ <h4>September 8-12, 2025</h4>
341370
<li><a href="._week37-bs008.html">9</a></li>
342371
<li><a href="._week37-bs009.html">10</a></li>
343372
<li><a href="">...</a></li>
344-
<li><a href="._week37-bs052.html">53</a></li>
373+
<li><a href="._week37-bs059.html">60</a></li>
345374
<li><a href="._week37-bs001.html">&raquo;</a></li>
346375
</ul>
347376
<!-- ------------------- end of main content --------------- -->

0 commit comments

Comments
 (0)