-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fundamentally different problem #1
Comments
according to the problem statement from https://gym.openai.com/envs/MountainCar-v0/ "A car is on a one-dimensional track, positioned between two "mountains". The goal is to drive up the mountain on the right; however, the car's engine is not strong enough to scale the mountain in a single pass. Therefore, the only way to succeed is to drive back and forth to build up momentum." so our goal is to make the car cross the flag.... and that is our objective and i can use whatever approach to achieve that objective ... in my case i have seen that the way i have chosen the rewards give the best result .... you can choose your own reward system to push the car in the correct direction... i don't think that it is wrong as it depends upon the implementor ..... |
Modifying the reward creates a new environment that is much easier to learn in. Yes you can do that to achieve the best result, but it is no longer comparable to solutions that use the vanilla environment. This is similar to openai/gym#499 (comment) |
if i cannot change the reward then how can i improve the results from another person who is using the same algorithm (ex DQN) |
Is this not a flawed approach since you have fundamentally changed the problem by altering the reward? To compare to other solutions the environment should be the same (including rewards given)
The text was updated successfully, but these errors were encountered: