Input normalization using a reinforcement learning DQN agent

8 vues (au cours des 30 derniers jours)
Margarita Cabrera
Margarita Cabrera le 26 Mar 2021
Commenté : H. M. le 3 Déc 2022
Hi all
I have built a DQN agent to solv a custom reinforcement problem.
Following mathworks examples I have no used any kind of normalization applied to the critic input.
In fact, as far as I could check, all examples of RL that use a DNN to create an actor or a critic especify 'Normalization', 'none' at the input layers of the Actor and Critic.
My question is, is it possible to use a normalization as for instance "zscore" at the input layers of a critic or of an actor when these are based on a DNN?
I'have tried to applied zscore normalization, but then, the agent does not work.
thanks

Réponse acceptée

Emmanouil Tzorakoleftherakis
Hello,
Normalization through the input layers is not supported for RL training. As a workaround, you can scale the observations rewards on the environment side.
  2 commentaires
Margarita Cabrera
Margarita Cabrera le 27 Mar 2021
Ok, thanks
H. M.
H. M. le 3 Déc 2022
@Emmanouil Tzorakoleftherakis
Could you explain more, the way you mentioned about normalization. I want to do it, but I can't figure it out.
Regards

Connectez-vous pour commenter.

Plus de réponses (0)

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by