Deep Q Learning - define an adaptive critic learning rate?

조회 수: 2 (최근 30일)
Niklas Reinisch
Niklas Reinisch 2020년 7월 16일
편집: Magnify 2020년 7월 29일
Hello,
at the moment i use Deep Q Learning for process planning and i would like to use an adaptive critic learning rate to speed up the training.
Is there any direct way (or workaround) to use a learning rate that lowers over the training process, e.g. depending on the number of epochs/steps, in DQL?
Thanks in advance and best wishes
Niklas

채택된 답변

Emmanouil Tzorakoleftherakis
Emmanouil Tzorakoleftherakis 2020년 7월 16일
Hi Niklas,
I believe this is currently not supported. This is an interesting usecase though - I will inform the development team. Is there any particular model you have in mind that would work better? For example linear/exponential decay, etc.
  댓글 수: 7
Niklas Reinisch
Niklas Reinisch 2020년 7월 21일
Thanks for your advice, i think i got your point!
Magnify
Magnify 2020년 7월 29일
편집: Magnify 2020년 7월 29일
There is one more question why the frequency of Agent's action outport is 0.05s rather than 0.025s specified by the agent sample time in my script createDDPGAgent.m, moreover, there is no way to modify it. there is a picture about sample time display as follow:sample time displayI would appreciate it if you give some tips to me.

댓글을 달려면 로그인하십시오.

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 Training and Simulation에 대해 자세히 알아보기

제품


릴리스

R2020a

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by