이 질문은 마감되었습니다. 편집하거나 답변을 올리려면 질문을 다시 여십시오.
I want to know that why simulation takes more time to show it's signal constallation when the signal to noise ratio goes higher? I think it's not the issue of server/pc. Is anybody there to say about this ?
조회 수: 1(최근 30일)
Hari Desanur 2017년 1월 5일
The model is set up to run forever - stop time is "inf". The two Error Rate Calculation (ERC) blocks are set up to stop simulation when 1000 errors are collected or 1e6 samples are processed, whichever happens first.
In the default set up (Eb/No = 0), the first ERC block collects 1000 errors in ~6600 samples and stops the simulation. When you increase the Eb/No value, the error rate reduces significantly. Therefore none of the ERC blocks hit 1000 errors even after processing 1e6 samples.
So, in the default case, the simulation stops after processing ~6600 samples, while in the later case the simulation continues until 1e6 samples are processed. Hence, the longer simulation time.