Rules of thumb on GPU usage?
조회 수: 3 (최근 30일)
이전 댓글 표시
I've converted several algorithms to using a gpu and I've always seen a tremendous improvement in execution time. For the first time, this particular trick has failed me. I'm seeing execution time lengthen when I use the GPU.
Is there a better way to determine the performance of a code snippet in a gpu than to alter the code and try it out?
Specifically when the target code may be executed on different classes of gpu's are there rules of thumb to predict the improvement/degradation that will result?
댓글 수: 0
답변 (1개)
Joss Knight
2015년 10월 12일
You ought to provide some examples so that we know the kind of thing you're getting at.
The main rule of thumb is that the GPU will generally perform well when your code is highly data-parallel. If you get a speed-up from vectorizing your code, you'll probably get a speed-up on the GPU. This means the same sort of operations are taking place in multiple places on a large dataset. If however, you have small pieces of data, a lot of disparate tasks, dependent operations, and loops, you probably don't have something that will parallelize well.
댓글 수: 2
Joss Knight
2015년 10월 14일
편집: Joss Knight
2015년 10월 14일
gpuArray supports logical indexing so I see no reason why you would need any data transfers (see the blog article I linked above for examples). Can you explain?
참고 항목
카테고리
Help Center 및 File Exchange에서 GPU Computing에 대해 자세히 알아보기
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!