Is plain torch.optim.SGD
doing the same as gradient descent?
#1146
Unanswered
shenhai-ran
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I am wondering about the implementation of
torch.optim.SGD
.If I don't use any batch size or dataloader, as the code snippet below from Chatper 1.
In this case, if I understand it right, there is no stochastic, but calculating native gradient descent for the whole dataset, and the
epoch
is basically the counter of steps in the optimization.Do I get it right?
Beta Was this translation helpful? Give feedback.
All reactions