Replies: 2 comments
-
Hey man, this is a common problem in machine learning where if we overfit our training set, we see in testing that often performance will decrease as our algorithm will learn the training data to well. There are many ways to address this across different algorithms, I suggest googling "overfitting" and whatever algorithm you are selecting. But as I said this is a very well explored, fundamental issue in ML/RL and its something you will have to take the time to learn about thoroughly. |
Beta Was this translation helpful? Give feedback.
-
obviously not. you should stop training to avoid overfitting. |
Beta Was this translation helpful? Give feedback.
-
The graph below shows Q, actor loss and reward as training process. The model is SAC.
As far I know, the algorithm objective is to maximise Q value, so I confuse that why Q value decreases along the training.
Also I notice that the more training episode, loss gets smaller but the result gets worsen in testing. If that's the case how many episode should I choose to train?
Thanks
Jason
Beta Was this translation helpful? Give feedback.
All reactions