Yahoo India Web Search

Search results

  1. Jan 24, 2024 · Gradient Descent is a fundamental optimization algorithm in machine learning used to minimize the cost or loss function during model training. It iteratively adjusts model parameters by moving in the direction of the steepest decrease in the cost function.

  2. May 22, 2021 · Gradient descent (GD) is an iterative first-order optimisation algorithm, used to find a local minimum/maximum of a given function. This method is commonly used in machine learning (ML) and deep learning (DL) to minimise a cost/loss function (e.g. in a linear regression).

  3. Gradient descent is a method for unconstrained mathematical optimization. It is a first-order iterative algorithm for finding a local minimum of a differentiable multivariate function .

  4. May 22, 2020 · Gradient Descent is an optimizing algorithm used in Machine/ Deep Learning algorithms. Gradient Descent with Momentum and Nesterov Accelerated Gradient Descent are advanced versions of Gradient Descent. Stochastic GD, Batch GD, Mini-Batch GD is also discussed in this article.

  5. Gradient descent is an algorithm that numerically estimates where a function outputs its lowest values. That means it finds local minima, but not by setting ∇ f = 0 ‍ like we've seen before. Instead of finding minima by manipulating symbols, gradient descent approximates the solution with numbers.

  6. Aug 12, 2019 · Gradient descent is an optimization algorithm used to find the values of parameters (coefficients) of a function (f) that minimizes a cost function (cost). Gradient descent is best used when the parameters cannot be calculated analytically (e.g. using linear algebra) and must be searched for by an optimization algorithm.

  7. Gradient descent is an optimization algorithm which is commonly-used to train machine learning models and neural networks. It trains machine learning models by minimizing errors between predicted and actual results.

  1. People also search for