Reinforcement Learning experience buffer length and parallelisation toolbox

조회 수: 2 (최근 30일)
When parallelisation is used when training a DDPG agent with the following settings:
trainOpts.UseParallel = true;
trainOpts.ParallelizationOptions.Mode = 'async';
trainOpts.ParallelizationOptions.StepsUntilDataIsSent = -1;
trainOpts.ParallelizationOptions.DataToSendFromWorkers = 'Experiences';
Does the the parallel simulations have their own experience buffer? This could take up more memory hence I am hoping that only one experience buffer is stored to update the critic network.
From the documentations, it seems like there will only be one experience buffer as the experiences are sent back to the host.

채택된 답변

Emmanouil Tzorakoleftherakis
Emmanouil Tzorakoleftherakis 2020년 12월 3일
편집: Emmanouil Tzorakoleftherakis 2020년 12월 3일
Hello,
There is one big experience buffer on the host, the size of which you determine as usual in your agent options. Each worker has a much smaller buffer to collect experiences until you reach "StepsUntilDataIsSent".

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 Training and Simulation에 대해 자세히 알아보기

제품


릴리스

R2020b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by