Intelligence.Log

2022-10-04

Extracted: 1 items. Sources: YouTube.
YT

We dive into some of the internals of MLPs with multiple layers and scrutinize the statistics of the forward pass activations, backward pass gradients...

👁 489.0k Views|Andrej Karpathy
"This video examines the statistical challenges in training deep neural networks, focusing on how improperly scaled activations and gradients can cause instability. It introduces Batch Normalization as a key technique to stabilize training by normalizing layer inputs."
-- END OF LOG --
[STATS] 1 items · Filter applied
Powered by Horizon + DeepSeek