Question

Why do we randomly shuffle the training examples before using SGD optimization?

Why do we randomly shuffle the training examples before using SGD optimization?

0 0
Add a comment Improve this question Transcribed image text
Answer #1

Hey here is answer to your query.

Main reason to shuffle training data before using SGD is to get unbiased result for the real gradient. Like mostly data is in ordered manner so if we provide that data to SGD then there is high probability that it will bias the gradient which will eventually lead to poor convergence . which will not be good for our model as then overall accuracy of model will be effected in negative way.

In case of any doubt please comment. Happy Learning :)

Add a comment
Know the answer?
Add Answer to:
Why do we randomly shuffle the training examples before using SGD optimization?
Your Answer:

Post as a guest

Your Name:

What's your source?

Earn Coins

Coins can be redeemed for fabulous gifts.

Not the answer you're looking for? Ask your own homework help question. Our experts will answer your question WITHIN MINUTES for Free.
Similar Homework Help Questions
ADVERTISEMENT
Free Homework Help App
Download From Google Play
Scan Your Homework
to Get Instant Free Answers
Need Online Homework Help?
Ask a Question
Get Answers For Free
Most questions answered within 3 hours.
ADVERTISEMENT
ADVERTISEMENT
ADVERTISEMENT