Community Profile

photo

Emmanouil Tzorakoleftherakis


Last seen: en dag ago

MathWorks

74 total contributions since 2018

Emmanouil Tzorakoleftherakis's Badges

  • Knowledgeable Level 3
  • 6 Month Streak
  • Revival Level 2
  • First Answer

View details...

Contributions in
View by

Answered
Why is the DDPG episode rewards never change during the whole training process?
Looks like the scale between Q0 and episode reward is very different. Try unchecking "Show Episode Q0" to see of the episode rew...

7 dagar ago | 0

| accepted

Answered
Reinforcement Learning -- Rocket Lander
Hi Averill, Here is a version that converges in ~18-20k episodes - thank you for pointing out that this example was not converg...

13 dagar ago | 0

| accepted

Answered
QTable reset when using train
If you stop training, you should be able to continue from where you left off. I called 'train' on the basic grid world example a...

14 dagar ago | 0

| accepted

Answered
How to test critic network in DDPG agent?
You can use getValue for inference on critic networks.

14 dagar ago | 0

| accepted

Answered
Questions about Reinforcement Learning
Hi Averill, Can you please let me know which examples do not converge? I will share this information with the development team....

21 dagar ago | 0

Answered
Error while designing MPC controller where output of vehicle body block is using as measured output in MPC
Seems the error is self-explanatory - have you checked the dimensions of the signal you are connecting to the 'mo' port of the M...

ungefär en månad ago | 0

| accepted

Answered
Simulink MPC Controller Constraints
Hi James, By "controlled variables" I am assuming you are referring to measured outputs. I believe that with linear MPC you can...

ungefär en månad ago | 0

| accepted

Answered
Can I Disable the Kalman Filter for MPC?
The following example shows how to disable the built-in estimator: https://www.mathworks.com/help/mpc/ug/custom-state-estimatio...

ungefär en månad ago | 0

Answered
How to apply the range control with Model Predictive Control in Simulink?
It seems to me that the following two examples correspond to the two scenarios you are describing: https://www.mathworks.com/he...

ungefär en månad ago | 0

Answered
How to make MPC controller act before disturbance hits?
Hi Nathan, I am assuming you have seen this example here? If the controller acted earlier, would it not violate the rotation pe...

ungefär en månad ago | 0

Answered
Help me understand the Architecture of DQN cor Cartpole problem in RL
Hi Michael, There are various architectures you can use when setting up the Q-network. In the example you mentioned and most ex...

ungefär en månad ago | 0

Answered
Episode simulation doesn't run while training DDPG
Hi Alice, This example has not been set up to update the visualization during training. If you follow add a MATLAB Function blo...

ungefär en månad ago | 0

| accepted

Answered
Unable to run 'rlwatertank' example in R2020a
Hi Nima, This is the plot I got when running the watertank example in 20a: While this is not exactly the same as the one show...

ungefär en månad ago | 0

Answered
In the process of reinforcement learning training, the training was terminated erroneously because the data dimensions did not match.
Are you using GPU to train? If yes, there is an indentified bug in R2020a which will be fixed when the first update for this rel...

ungefär 2 månader ago | 0

| accepted

Answered
Deep Deterministic Policy Gradient Agents (DDPG at Reinforcement Learning), actor output is oscilating a few times then got stuck on the minimum.
Hi Samir, After reviewing your model, if you check the actions the agent outputs, they blow up to infinity. That should not b...

2 månader ago | 0

Answered
Reinforcement learning for adaptive cruise control
Hello, Try the following load('SimulinkACCDDPG.mat','agent') actor = getActor(agent); weights = getLearnableParameterValue...

2 månader ago | 0

| accepted

Answered
Policy Gradient with Baseline Reward Oscillation (MATLAB Reinforcement Learning Toolbox)
Hello, Some suggestions: 1) For a 13 to 2 mapping, maybe you need another set of FCl+Relu layers in your actor 2) Since you h...

3 månader ago | 0

Answered
Do we need a delay block in reinforcement learning simulink?
A delay block may be necessary to break algebraic loops or to rely on observations from previous time steps (see e.g. this examp...

3 månader ago | 1

Answered
Reinforcement Learning Toolbox: defining Size of layers and other parameters.
Hello, I would recomment looking at a similar example in thw documentation (e.g. this one). To your questions: Yes. Number of ...

3 månader ago | 0

Answered
training an inverted pendel with DDPG agent
Can you try saving to a different folder? It looks like a permissions issue.

3 månader ago | 0

| accepted

Answered
Custom Action Space DDPG Reinforcement Learning Agent
To my knowledge, you cannot implement a custom action space with rlNumericSpec, but what you could possibly do (since adding pen...

3 månader ago | 0

| accepted

Answered
Generate Cuda code from a pretrained rlDDPGAgent object for NVIDIA board
If you see here, tha tanhLayer support code generation with GPU Coder starting in R2019b.

3 månader ago | 0

| accepted

Answered
Binary Decision Variable in MPC
This should be doable with custom constraints in nonlinear MPC. You can create your own function that decides how the constraint...

3 månader ago | 0

Answered
Export the reinforcement learning result - to see the weights of critic network and actor network
Hello, You can see the values of the neural network weights using this function. Yes, you can apply DDPG and RL in general to ...

3 månader ago | 0

Answered
Deploy trained policy to simulink model
Hello, Looks like the dimensions cannot be determined automatically. If you double click the MATLAB Fcn block and then click "E...

3 månader ago | 0

| accepted

Answered
RL Toolbox: Combine Discrete and Continuous Observations
Does the environments output continuous and discrete observations? If yes, couldn't you use 'rlNumericSpec' for both? The discre...

4 månader ago | 0

Answered
RF - Create MATLAB Environment using Custom Functions - myResetFunction
Hi Zhen, I believe you are right - I have informed the documentation team about this.

4 månader ago | 0

| accepted

Answered
reinforcement learning using my own function
It looks like your "core" function qualifies as the actual policy (or value function for that matter). The environment would be ...

4 månader ago | 0

Answered
How can I find the template for the predefined environment: "CartPole-Discrete"
The predefined environments are coded in an object-oriented way, so you may not find all the info in one file. I would start wit...

4 månader ago | 0

| accepted

Answered
How to continue training a DQN agent in the reinforcement learning toolbox?
Hi James, It looks like the experience buffer is the culprit here. Have a look at this question for a suggestion. Pretty much y...

4 månader ago | 0

| accepted

Load more