How to export and use only the end product of a reinforcement learning algorithm ?

조회 수: 15 (최근 30일)
Hello
I have used reinforcement learning to train a TD3 agent. Now I want to use this agent and actually deploy it as a controller in a simulink model, then possibly on an embedded platform. From what I understand about reinforcement learning, the actor network is the actual end product which computes the control action. Therefore, I don't want to export everything else with it as a RL agent representation, just the neural net. Is there something I should be wary of when doing this ? Also what simulink block can I use for a deep NN controller ? I am not sure if the predict block is suitable here as the task is not classification and the output should be an action, rather than likelihood percentages.
Also, I have created a TD3 agent with LSTM layers on MATLAB R2021a. When I try to import the agent to R2020b, which it is incompatible with, the agent surprisingly does get imported but when I try to simulate it to validate the results I get very different ones from when I try to do it on R2021a. Are the LSTM layers inside of the agent not working properly in R2020b or are they completely incompatible (for simulating the agent only, not training) ? Would doing something I described above (importing only the neural net as a controller) allow me to use the network as a controller on older versions of MATLAB?
Thanks for the help.

채택된 답변

Arkadiy Turevskiy
Arkadiy Turevskiy 2021년 8월 6일
Hi there,
To deploy trained RL agent you need to:
  1. Extract trained policy from RL agent. For that you use generatePolicyFunction. As the doc explains, this function will create a function evaluatePolicy.m and agentData.mat file.
  2. To run inference on trained policy in Simulink, use MATLAB Function block in Simulink. In MATLAB Funciton block call evaluatePolicy
  3. You are done! You can now simulate your trained policy in Simulink. Starting with 21a we support ANSI C code gen for deep learning networks. So you can generate ANSI C code from your Simulink model that contains whetever algorithms you have plus trained RL policy represented by MATLAB Function Block.
  4. The generated code should be compatible with any microcontroller, or with a rapid prototyping setup, for example, with speedgoat machine.
To see the details on what layers currently support ANSI C (generic C), please refer to this doc page.
As for the second question, T3D support for LSTMs came in 21a in Reinforcement Learning Toolbox. So as you point out you would not be able to use this agent in 20b. However, you should be able to extract trained policy as described above and run inference on it in 20b. In 21a we added a Simulink block to Deep Learning Toolbox for simulating LSTMs, but I think you should be able to simulate in 20b using MATLAB Function block. If you can;t or reults don't make sense, please reach out to tech support.
  댓글 수: 3
Shah Fahad
Shah Fahad 2023년 11월 14일
Hello Arkadiy,
Is it possible to export the experiences of the DRL agent as a file such as excel file. So that they can be used in another hardware device such as RTDS.
Thank you.
Emmanouil Tzorakoleftherakis
Emmanouil Tzorakoleftherakis 2023년 11월 15일
Hi,
What you are asking should be possible, but I would need some additional info to make sure what the best way to proceed would be. For example:
1) It seems RTDS is used for HIL testing - is this what you are trying to accomplish?
2) What do you mean by 'export the experiences of the DRL agent?'? By experiences we usually refer to actions, observations, rewards. How are you going to use all of this? What would be the setup? I am asking because, e.g. for HIL test, you likely don't need all of that.
Also, since this is an old thread, can you maybe create a new questions with some additional details? It helps with discoverability.
Thanks!

댓글을 달려면 로그인하십시오.

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 Reinforcement Learning에 대해 자세히 알아보기

제품


릴리스

R2021a

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by