Coursify
Create New CourseGalleryContact

Artificial Intelligence

Unit 1
Optimization
Introduction to OptimizationGradient DescentStochastic Gradient DescentAdam Optimization
Unit 1 • Chapter 3

Stochastic Gradient Descent

Video Summary

Stochastic Gradient Descent (SGD) is a simple yet effective algorithm for optimizing a function with many variables. It works by iteratively taking steps in the direction of the negative gradient of the function, where the gradient is estimated using a small subset of the data. This makes SGD very efficient, as it does not require storing the entire dataset in memory. However, SGD can be prone to overfitting, so it is often used in conjunction with other techniques such as regularization.

Knowledge Check

Stochastic gradient descent is an iterative optimization algorithm that uses what to find the minimum of a function?

Stochastic gradient descent is a type of what?

Stochastic gradient descent is used to solve what type of problems?