Réponse apportée
Epsilon greedy algorithm and environment reset do not work during DQN agent training
Hello, Here are some comments: 1.The reset function should not produce the same output. You should first doublecheck the reset...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Mix of static and dynamic actions for a Reinforcement Learning episode
Hello, I am not sure the approach you mention would work, since even if you constrain the constant action, the agent will still...

environ 5 ans il y a | 0

Réponse apportée
Problems in using Reinforcement Learning Agent
Hello, I am assuming you have seen this example already? Seems similar. I don't see the script where you set up DDPG but there ...

environ 5 ans il y a | 0

Réponse apportée
How to simulate saved agents?
Hello, Righ-click on the 'Agents' folder from within MATLAB and add it to path (or use addpath). Then load Agent1.mat xpr = s...

environ 5 ans il y a | 0

Réponse apportée
Simulating environment while Training rlAgent
I can interpret your question in 3 ways so I will put my thoughts here and hopefully they will be sufficient. 1) Depending on t...

environ 5 ans il y a | 0

Réponse apportée
How do I specify multiple, heterogeneous actions for the rl.env.MATLABEnvironment of the Reinforcement learning toolbox or another way, if there is one?
Hello, I probably don't understand your objective but the two actions you mention above (distance and speed) are still scalars....

environ 5 ans il y a | 1

| A accepté

Réponse apportée
How can I create a simulation environment with reinforcement learning?
rlFunctionEnv can be used to create an environment where the dynamics are in a MATLAB function. You could also create an environ...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Reinforcement Learning Grid World multi-figures
Hello, I wouldn't worry about the spikes as long as the average reward has converged. Could be the agent exploring something. ...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Reinforcement Learning Toolbox Example Ball Balancing
Hello, This example was created as part of a tutorial for IROS 2020. The example files are here, in folder #3. Hope this helps...

environ 5 ans il y a | 1

| A accepté

Réponse apportée
How can I use the different Target Smooth Factor in actor and critic network? (Reinforcement Learning Toolbox)
Hello, This is not currently possible, but I have let the development team know and they will look into it. Thanks for bringin...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Episode Q0 increases exponentially
Hello, Please take a look at this answer for some suggestions. Normalizing observations, rewards, and actions can also help avo...

environ 5 ans il y a | 0

Réponse apportée
How GAE calculates in Reinforement Learning Toolbox(PPO)?
Hello, Thank you for catching this typo - it should be Gt = Dt+V. I have let the documentation team know.

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Get observation of final episode of RL agent
Do you want to save the observations in the last time step of the final episode? Or all the observations shown in the final epis...

environ 5 ans il y a | 1

| A accepté

Réponse apportée
Problem using scalingLayer for shifting actor outputs to desired range
If you remove ',...' from the 'tanh' row the error goes away. The way you have it now, you are adding the scaling layer in the s...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Prediction of NOx emissions by using the cylinderpressure curves of an internal combustion engine
This sounds like a supervised learning problem with time dependencies in which case I would recommend working with LSTMs and Dee...

environ 5 ans il y a | 0

Réponse apportée
RL toolbox train on continuous simulation with delay between episodes
Hi Joe, I believe the setup you mention may be possible but it will require some work.Essentially, you need to set up training ...

environ 5 ans il y a | 0

Réponse apportée
Hybrid reinforcement learning and traditional control environment
Hello, You can put the RL Agent block in an enabled subsystem and use the desired condition to indicate when to use RL and when...

environ 5 ans il y a | 0

Réponse apportée
Reinforcement Learning : MaxSumWordLength is 65535 bits and a minimum word length of 65536 bits is necessary so that this sum or difference can be computed with no loss of precision - ( 'rl.simulink.blocks.AgentWrapper' )
It's a bit hard to find the cause of the error without a reproduction model, but based on the error you are seeing, I would chec...

environ 5 ans il y a | 1

Réponse apportée
RL Toolbox: DQN epsilon greedy exploration with epsilon=1 does not act random
Hello, Maybe I misread the question, but you are saying "when starting the Simulation and watching the output of the episodes.....

environ 5 ans il y a | 0

| A accepté

Réponse apportée
How to get the history action value of reinforcement learning agent
Not exactly sure what you mean. During training the RL algorithms are already doing inference. You can use getAction and getValu...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
How to train a Reinforcement Learning agent from 0.02s of simulation
I believe you can put the RL Agent block in an enabled subsystem and set the enable time to be 0.02 seconds. Hope that helps

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Why my RL training abruptly stopping before the total EpisodeCount?
Please take a look at this doc page. While you are selecting "episodecount" as the termination criterion, you don't set the stop...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
Saving simulation data during training process of RL agents
Have you tried logging the data with Simulation Data Inspector? Make sure to pick only signals you actually need since depending...

environ 5 ans il y a | 0

Réponse apportée
How to use RNN+DDPG together?
The ability to create LSTM policies with DDPG is available starting in R2021a. Hope that helps

environ 5 ans il y a | 1

| A accepté

Réponse apportée
Customized Action Selection in RL DQN
Hello, I believe this is not possible yet. A potential workaround (although not state dependent) would be to emulate a pdf by p...

environ 5 ans il y a | 0

Réponse apportée
How to save and use the pre-trained DQN agent in the reinforcement learning tool box
Hello, Take a look at this example, and specifically the code snippet below: if doTraining % Train the agent. tr...

environ 5 ans il y a | 1

Réponse apportée
Save data in Deep RL in the Simulink environment for all episodes
Hello, You can always select the signals you want to log and view them later in Simulation Data Inspector. Same goes for the re...

environ 5 ans il y a | 1

| A accepté

Réponse apportée
How to compute the gradient of deep Actor network in DRL (with regard to all of its parameters)?
In the link you provide above, the gradients are calculated with the "gradient" function that uses automatic differentiation. So...

environ 5 ans il y a | 0

Réponse apportée
MPC: step response starts with unwanted negative swing when using previewing
It appears that the optimization thinks that moving in the opposite direction first is "optimal". You can change that by adding ...

environ 5 ans il y a | 0

| A accepté

Réponse apportée
RL: Continuous action space, but within a desired range
Hello, There are two ways to enforce this: 1) Using the upper and lower limits in rlNumericSpec when you are creating the acti...

environ 5 ans il y a | 0

Charger plus