Effacer les filtres
Effacer les filtres

When an agent makes more than one action sometimes these actions have opposite effects rather than coordinating effects

11 vues (au cours des 30 derniers jours)
I am now doing a deep reinforcement learning experiment based on multiple agents, each agent can emit 3 action signals, but why do these action signals always appear opposite effects instead of synergistic effects
Sincerely look forward to an answer

Réponse acceptée

Subhajyoti
Subhajyoti le 22 Août 2024 à 10:50
Hi Huixin,
These kinds of challenges are common in reinforcement learning applications. You can address the issues you face with your RL Agent using some of the following ways:
  • Reward Function: It must reflect the desired cooperative behaviour. You can try incorporating shared rewards or team-based objectives to align individual agent goals with the overall system performance.
  • Training Stability: A stable ‘Average Reward Curve’ indicates consistent learning, while high variance might indicate instability or conflicting actions among agents. Ideally, the variance reduces as training progresses.
  • Hyperparameter Tuning: Often, tuning the hyperparameters like the Learning Rate (LR), Discount Factors (DF), etc. can significantly improve the performance of the RL Agents.
Amongst other things, you can also tweak the architecture of the network by increasing the depth with additional layers to improve the learning capability of your agent.
Also, during training, you can vary the ‘Epsilon-Greedy Parameter’. Begin with a high epsilon value (close to 1) to encourage exploration and allow agents to discover diverse strategies and state-action pairs. Gradually decay epsilon towards 0 in later phases to shift focus on exploiting the learned policy, ensuring agents make the most of their training experiences.
You may go through the following MathWorks documentation links to know more about Training RL Agents and training options.
I hope this helps

Plus de réponses (0)

Produits


Version

R2022b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by