Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The training step of CQL-SAC. #1

Open
DooyoungH opened this issue Sep 29, 2021 · 1 comment
Open

The training step of CQL-SAC. #1

DooyoungH opened this issue Sep 29, 2021 · 1 comment

Comments

@DooyoungH
Copy link

I am studying by referring to your CQL code.
But, I think Line 68 not be proper to Offline RL when I run the train.py of CQL-SAC.
Line 68 : buffer.add(state, action, reward, next_state, done)

Isn't this line an off-policy model by putting data that interacts with the agent and the environment into a buffer?

I thank you for your hard works.

@BY571
Copy link
Owner

BY571 commented Sep 29, 2021

Yes, indeed this is only for the RL setting for an SL setting or BatchRL setting you might have to adapt that

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants