Update for the latest library enviromnet #61
Closed
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Hello, first of all, thank you for providing this great project.
As I was referring to your repository to write the code, I found that it has not been updated for quite a long time. I have modified some parts of the code to make it suitable for the latest gym library environment (changes include Acer, Actor_critic, DQN, PPO, PPO-LSTM, REINFORCE, and VTRACE algorithms).
Some corrections have been made for the tuple-type values returned from the updated Open AI Gym's dependency, and for the conversion speed due to data type transformations.
(e.g. UserWarning: Creating a tensor from a list of numpy.ndarrays is extremely slow. Please consider converting the list to a single numpy.ndarray with numpy.array() before converting to a tensor.)
Most of the algorithms' speed has been improved by more than 10%, and we have made minimal changes to maximize efficiency, keeping in line with the spirit of minimalRL.
Thank you again for this great project!
p.s. Add some detailed patch note
#59 is solved now.