# A basic question on Stochastic gradient descent

 0 Consider a stochastic gradient iteration: \(\theta_$k+1$ = \theta_$k$ - \gamma_k F(\theta_k))\ where $F$ is a noisy estimate of the gradient $\nabla f$ Now, a book says that it converges in the following sense : $f(\theta_k)$ converges and $\nabla f(\theta_k)$ converges to zero and then it says that it is the strongest possible result for gradient related stochastic approximation. What is the meaning of it ? Why does not it shows the convergence of the iterates ? asked 15 Aug '14, 09:27 sosha 100●1●3●7 accept rate: 0%
Be the first one to answer this question!
 toggle preview community wiki

By Email:

Markdown Basics

• *italic* or _italic_
• **bold** or __bold__
• image?![alt text](/path/img.jpg "Title")
• numbered list: 1. Foo 2. Bar
• to add a line break simply add two spaces to where you would like the new line to be.
• basic HTML tags are also supported

Tags: