답변 있음
GPU Array Max Dimensions/Size (i.e., int32 of ~2e9) Not Same as Other CUDA-Enabled Languages (i.e., Python 3.9 - Tensorflow 2.11)
In CUDA 12 NVIDIA introduced a limited set of 64-bit integer APIs in cublas and other libraries, allowing arrays with more eleme...

1년 초과 전 | 2

| 수락됨

답변 있음
Multilayer Perceptron with More Than One Output and Data Interpretation?
You can continue to use trainNetwork if you don't want to use dlnetwork. dlnetwork obviously provides much more flexibility as w...

1년 초과 전 | 0

답변 있음
Matlab continuously launching on GPU
Using the NVIDIA Control Panel, disable hardware acceleration for MATLAB. Alternatively, type |opengl software| on startup.

1년 초과 전 | 0

답변 있음
Using transformer neural network for classification task
You've passed |layers| instead of |lgraph| to |trainNetwork|.

1년 초과 전 | 0

| 수락됨

답변 있음
I get the illegal instruction error while running Matlab code with GPU
On the face of it it does seem like you've hit some kind of bug in CUDA 11. If you can upgrade your MATLAB to a more recent vers...

1년 초과 전 | 0

| 수락됨

답변 있음
GPU memory usage for Hadamard product
I can't reproduce this. You say you have 48Gb of GPU memory available...have you checked this? Try running gpuDevice and lookin...

1년 초과 전 | 0

답변 있음
Initializing LSTM which is imported using ONNX
This code is suspect % initialize input arrays obs = dlarray(rand(obs_size),"BS"); state = dlarray(rand(state_size),"SBS"); ...

1년 초과 전 | 0

답변 있음
trainnet gives training loss is NaN
Do your network weights contain NaNs? Try this nansInMyNetwork = ~(all(cellfun(@allfinite, net.Learnables.Value)) && all(cellfu...

1년 초과 전 | 0

답변 있음
how to use printf inside a CUDA kernel?
Just use it, and launch MATLAB from a terminal. On Linux, the output will appear in the terminal window. On Windows you will nee...

1년 초과 전 | 0

| 수락됨

답변 있음
mexcuda compiler error: "__global__" does not apply here
You cannot call a function declared __global__ from another function declared __global__. Declare the second function as __devic...

1년 초과 전 | 0

| 수락됨

답변 있음
CUSOLVER_STATUS_INTERNAL_ERROR with RTX 4080 Super in MATLAB 2022a C MEX Function
This is a known issue with CUDA 11.2, Ada cards, and cusolver. The solution is to upgrade MATLAB to R2023a or above, because it ...

1년 초과 전 | 1

| 수락됨

답변 있음
CUSOLVER_STATUS_INTERNAL_ERROR with RTX 4080 Super in MATLAB 2022a C MEX Function
It's probably necessary for you to provide your code, since any number of things could have gone wrong prior to intializing cuso...

1년 초과 전 | 0

답변 있음
why do I got such a low score when i use bench function in matlab2024a
Do check your computer's power settings, to make sure they are set to "best performance".

1년 초과 전 | 0

답변 있음
Matab is very slow in processing figures: might it be caused by NVidia Optimus?
Check your laptop's power settings and make sure they are in 'best performance' mode.

1년 초과 전 | 0

| 수락됨

답변 있음
Which is the difference between 'multi-gpu' and 'parallel-gpu' in 'trainingOption' function of the DeepLearning Toolbox?
The purpose of 'multi-gpu' is effectively to try to ensure you are using a local pool with numGpus workers, rather than needing ...

1년 초과 전 | 1

답변 있음
MATLAB does not detect NVIDIA T400 GPU despite updated drivers
The 516 driver is too old for R2024a, which is on CUDA 12.2. Download the recommended 552 driver for your T400 from https://www....

1년 초과 전 | 0

답변 있음
MATLAB GPU UBUNTU 20.04 compatibility CUDA driver
Try the 555 and 550 drivers recommended for your device available from https://www.nvidia.com/en-gb/drivers/, rather than relyin...

1년 초과 전 | 0

답변 있음
Memory leak occurring when repeatedly performing matrix left division on GPU
Thank you very much for reporting this. This appears to be a bug in MATLAB's gpuArray support for overdetermined solves. For now...

1년 초과 전 | 0

| 수락됨

답변 있음
How do I use multiple inputs in mexcuda?
You have passed in two inputs but you have a check if (nrhs!=1), so you are erroring because you don't have one input. You need ...

거의 2년 전 | 1

답변 있음
Why training stops showing error (out of memory) during implementing transfer learning with pre-trained network despite having a laptop of well configuration?
You have a very small GPU I'm afraid. You need to lower your |MiniBatchSize| until it works. Also, set your |ExecutionEnvironmen...

거의 2년 전 | 1

| 수락됨

답변 있음
Parallel Processing is better than buying another MATLAB licence?
There may be a misunderstanding here. Your typical license is to a user or a computer. That user, or that computer, can open as ...

거의 2년 전 | 0

답변 있음
I can`t solve out this problem, there is always Output argument "varargout{2}" (and possibly others) not assigned a value in the execution with "dlarray/dlgradient" function.
Your mistake is |dlfeval(@dlgradient,...)|. You need to put your code that computes the loss and the loss gradients into a funct...

거의 2년 전 | 0

| 수락됨

답변 있음
AlexNet (filter groups) cannot be well resolved inside dlfeval (dimension mismatch error)
Sorry about this. This is because you have an old version of alexnet (although yes, it shouldn't error). For now, please get you...

거의 2년 전 | 0

| 수락됨

답변 있음
arrayfun vs loops again
I wish it were a cleverer answer, but I'm afraid that it's simply that MATLAB has been heavily optimized for |for| loops over th...

거의 2년 전 | 3

| 수락됨

답변 있음
Extreme memory usage when using gather
As others have worked out, it looks like the issue is the indexed assignment into u: u(TERRAIN:end,m) = abs(gather(head(ffield,...

거의 2년 전 | 0

| 수락됨

답변 있음
How can i fix the error mismatch format when convert multiplication layer from LayerGraph to dlnetwork in prune and quantize network example
For one solution, replace the fully connected layers with convolution layers with filters the size of the input and num filters ...

거의 2년 전 | 0

답변 있음
How can i fix the error mismatch format when convert multiplication layer from LayerGraph to dlnetwork in prune and quantize network example
That's an annoying limitation which is hopefully fixed in current releases. Is there any particular reason why you can't use |mu...

거의 2년 전 | 0

답변 있음
Arrayfun/gpuArray CUDA kernel need to be able to remember previous steps
This is a bit too vague to answer. Without indexing, how can each subproblem retrieve its subset of the data? If you just mean i...

거의 2년 전 | 0

답변 있음
Loop through list of vectors (ie, the rows of a matrix), applying same "simple" function to each one, on GPU? Should I use arrayfun somehow?
If your calculation is truly sequential then by definition you cannot parallelize along the sequence. But it sounds like you def...

거의 2년 전 | 0

답변 있음
Parallel Computing in C when using Matlab Coder (parpool and gpuArray)
GPU Coder will generate CUDA code for you. It can even automatically vectorize |for| loops. For a multithreaded parallel for loo...

대략 2년 전 | 1

더 보기