You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.
when I use NaiveEngine in openmp multi-threading, I binded 2 executor on gpu(0) and gpu(1).
and do exe.forward parallelly. However I got some error when running the program:
I am getting the same error, when running some pretty benign code on a normal CPU instance (m4.xlarge, Ubuntu Deep Learning AMI). The code is binding several executors in sequence. The error occurs only with NaiveEngine. It does not occur on my Mac.
It seems like the internal memory pool tries to call cudaFree() on the same resource assigned to two or more MXNet instances when running Naive Engine. The memory pool is supposed to be thread-local singleton so each MXNet instance spawned by each thread does not contend with each other the same resource, but when running in Naive Engine it apparently is not the case. I get all sorts of errors like CUDA invalid pointer error and eventually cuBLAS failure
when I use NaiveEngine in openmp multi-threading, I binded 2 executor on gpu(0) and gpu(1).
and do exe.forward parallelly. However I got some error when running the program:
why the NaiveEngine has asynchronous operations?
The text was updated successfully, but these errors were encountered: