Artificial Intelligence
Machine Learning
Subjective
Oct 13, 2025
Explain gradient descent and its variants.
Detailed Explanation
Gradient descent optimizes model parameters by iteratively moving in the direction of steepest descent of the loss function.\n\n• Batch GD: Uses entire dataset, stable but slow\n• Stochastic GD: Uses single sample, fast but noisy\n• Mini-batch GD: Compromise between batch and stochastic\n• Advanced: Adam, RMSprop, AdaGrad with adaptive learning rates\n\nExample: For neural networks, use Adam optimizer with learning rate 0.001, batch size 32. Monitor loss curves, adjust learning rate schedule, and use gradient clipping to prevent exploding gradients.
Discussion (0)
No comments yet. Be the first to share your thoughts!
Share Your Thoughts