I want to convert the PI controller in this instance to a PID controller, so can my network be changed to something like this?

조회 수: 5 (최근 30일)
Respected all,
The following example shows the PI controller design using TD3 reinforcement learning algorithm. I want to convert the PI controller in this instance to a PID controller, so can my actornetwork be changed to something like the following?
actorNetwork = [
featureInputLayer(numObservations3,'Normalization','none','Name','state')
fullyConnectedPILayer(randi([-50,50],1,3), 'Action')];
  댓글 수: 1
轩
2024년 1월 4일
Hello, I am trying the same method in the current time, can you leave a contact information for conmunication ?
你好,可以留一个邮箱或微信交流吗?

댓글을 달려면 로그인하십시오.

채택된 답변

Emmanouil Tzorakoleftherakis
Emmanouil Tzorakoleftherakis 2023년 10월 23일
That should work - you are essentially adding one more weight to the PI layer for the D term. Potential issues may arise when you provide the error derivative as an observation (may be hard to approximate due to the trial and error nature of RL) but give it a try
  댓글 수: 2
嘻嘻
嘻嘻 2023년 10월 23일
I am very grateful for your answer, which is very helpful to me. I would like to ask you another question, I would like to know how the initial value of the network weight in the example is determined, and what impact will it have on the result? Thank you again!
Emmanouil Tzorakoleftherakis
Emmanouil Tzorakoleftherakis 2023년 10월 23일
It looks like it's just a couple of values close to zero. RL solves an optimization problem under the hood so the better the initial guess, the more likely to get a better solution

댓글을 달려면 로그인하십시오.

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 PID Controller Tuning에 대해 자세히 알아보기

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by