Skip to main content

Normalization in ML Explained

1 min

What I worked on

Tried to understand why normalization requires integrating over all possible outcomes in probability. Looked at examples like dice PDFs and compared spiking neural networks and energy-based models.

What I noticed

  • Integration ensures total probability sums to 1
  • PDF for dice would be discrete and flat
  • Spiking networks and energy-based models have very different dynamics

”Aha” Moment

na

What still feels messy

Still unclear how integration maps directly to ML normalization steps.

Next step

Review how energy-based models use normalization constants (partition functions).