Computational Model Library

Displaying 2 of 2 results optimization algorithms clear search

Gradient Descent Simulation

Ilyes Azouani | Published Wednesday, March 18, 2026

This model visualizes gradient descent optimization - the fundamental algorithm used to train neural networks and other machine learning models. Agents represent different optimization algorithms searching for the minimum of a loss landscape (the “error surface” that ML models try to minimize during training).

The model demonstrates how different optimizer types (SGD, Momentum with different parameters) behave on various loss landscapes, from simple bowls to the notoriously difficult Rosenbrock “banana valley” function. This helps build intuition about why certain optimization algorithms work better than others for different problem geometries.

HOW IT WORKS

PowerGen-ABM is an optimisation model for power plant expansions from 2010 to 2025 with Indonesian electricity systems as the case study. PowerGen-ABM integrates three approaches: techno-economic analysis (TEA), linear programming (LP), and input-output analysis (IOA) and environmental analysis. TEA is based on the revenue requirement (RR) formula by UCDavis (2016), and the environmental analysis accounts for resource consumption (i.e., steel, concrete, aluminium, and energy) and carbon dioxide equivalent (CO2e) emissions during the construction and operational stages of power plants.

This website uses cookies and Google Analytics to help us track user engagement and improve our site. If you'd like to know more information about what data we collect and why, please see our data privacy policy. If you continue to use this site, you consent to our use of cookies.
Accept