Comments (5)
Hello,
This is normal. If you do hyperparameter tuning, you should set policy='MlpPolicy'
otherwise you will get the mentioned error, as the CustomSACPolicy
is already custom in term of number of layers, would be nice to change CustomSACPolicy
to MlpPolicy
but with policy_kwargs="dict(layers=[256,256])"
from rl-baselines-zoo.
Ok thanks for your very quick reply.
from rl-baselines-zoo.
Just a doubt, is it OK then to tune the hyperparameters with policy='MlpPolicy'
and then to train the model with CustomSACPolicy
? Does it not defeat the purpose of tuning in the first place? i.e. would hyperparameters optimised with one policy be also optimal for another policy?
from rl-baselines-zoo.
Does it not defeat the purpose of tuning in the first place? i.e. would hyperparameters optimised with one policy be also optimal for another policy?
If in your hyperparameter optimization you allow architecture search:
rl-baselines-zoo/utils/hyperparams_opt.py
Lines 245 to 250 in 645ea17
then it does make sense to have policy='MlpPolicy'
.
However, if you fix the architecture (by commenting the lines above), then you can use CustomSACPolicy
(or in a equivalent way, MlpPolicy
+ policy_kwargs="dict(layers=[256,256])"
)
from rl-baselines-zoo.
Ok thanks a lot for your help, I'm closing this issue now.
from rl-baselines-zoo.
Related Issues (20)
- HER success_rate output [question]
- What is the best initial values for the parameters in the .yml? [question] HOT 4
- How is the trial value calculated for RL jobs? HOT 1
- update the baseline for new envs? [feature request] HOT 1
- Multithreaded training with SubprocVecEnv() not working HOT 11
- [feature request] Install Zoo as a package HOT 6
- Irreproducible zoo trials HOT 5
- BadZipFile when running PPO2. HOT 5
- Can we use it with Tenserflow 2.3.1 ? HOT 5
- Why during the evaluation VecNormalize stats arenot loaded from the current training stats? HOT 1
- [Question] Why do envs look different? HOT 3
- [Question] Why is ortho_init=False during hyperparameter searches? HOT 2
- [feature request] Hyperparameter optimization for pretrained agent
- Hyperparameter tuning using Custom env with RL-zoo and SB2 HOT 3
- SB3 v1.1 Breaking changes HOT 1
- [question] Transfer hyperparameters from optuna HOT 2
- [feature request]Add Mujoco-Gym since mujoco is free now HOT 2
- Missing (or can't find) Hyperparameters HOT 2
- Reproducible zoo result HOT 3
- I am getting ImportError in atari environment
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from rl-baselines-zoo.