tf_rl | Refinforcement learning framework
kandi X-RAY | tf_rl Summary
kandi X-RAY | tf_rl Summary
Refinforcement learning framework
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of tf_rl
tf_rl Key Features
tf_rl Examples and Code Snippets
Community Discussions
Trending Discussions on tf_rl
QUESTION
I am trying to write my own DQN algorithm in Python, using Tensorflow following the paper(Mnih et al., 2015). In train_DQN
function, I have defined the training procedure, and DQN_CartPole
is for defining the function approximation(simple 3-layered Neural Network). For loss function, Huber loss or MSE is implemented followed by the gradient clipping(between -1 and 1). Then, I have implemented soft-update method instead of hard-update of the target network by copying the weights in the main network.
I am trying it on the CartPole environment(OpenAI gym), but the rewards does not improve as it does in other people's algorithms, such as keras-rl. Any help will be appreciated.
If possible, could you have a look at the source code?
- DQN model: https://github.com/Rowing0914/TF_RL/blob/master/agents/DQN_model.py
- Training Script: https://github.com/Rowing0914/TF_RL/blob/master/agents/DQN_train.py
- Reddit post: https://www.reddit.com/r/reinforcementlearning/comments/ba7o55/question_dqn_algorithm_does_not_work_well_on/?utm_source=share&utm_medium=web2x
ANSWER
Answered 2019-Apr-06 at 19:33Briefly looking over, it seems that the dones
variable is a binary vector where 1
denotes done, and 0
denotes not-done.
You then use dones
here:
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install tf_rl
You can use tf_rl like any standard Python library. You will need to make sure that you have a development environment consisting of a Python distribution including header files, a compiler, pip, and git installed. Make sure that your pip, setuptools, and wheel are up to date. When using pip it is generally recommended to install packages in a virtual environment to avoid changes to the system.
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page