You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
System: 2x RTX 4090, fully working ( pytorch, etc )
I am trying to implement multi gpu inference ( by splitting my inputs up etc )
When i create my memory allocator for Ort Tensors like this:
var memoryInfo = new OrtMemoryInfo(OrtMemoryInfo.allocatorCUDA_PINNED,
OrtAllocatorType.DeviceAllocator,
deviceId,
OrtMemType.CpuInput | OrtMemType.CpuOutput);
var ortAllocator = new OrtAllocator(session, memoryInfo);
The session was of course created on the correct device as well:
Describe the issue
System: 2x RTX 4090, fully working ( pytorch, etc )
I am trying to implement multi gpu inference ( by splitting my inputs up etc )
When i create my memory allocator for Ort Tensors like this:
The session was of course created on the correct device as well:
This works fine for my first card ( deviceId = 0 ) but for deviceId = 1 i get an error:
[ErrorCode:InvalidArgument] No requested allocator available
Could this be an issue or is there some incantantion to make this work ?
To reproduce
´´´
var memoryInfo = new OrtMemoryInfo(OrtMemoryInfo.allocatorCUDA_PINNED,
OrtAllocatorType.DeviceAllocator,
deviceId,
OrtMemType.CpuInput | OrtMemType.CpuOutput);
var ortAllocator = new OrtAllocator(session, memoryInfo);
´´´
Urgency
No response
Platform
Windows
OS Version
Fedora 41
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.21.0
ONNX Runtime API
C#
Architecture
X64
Execution Provider
CUDA
Execution Provider Library Version
Cuda 12.8
The text was updated successfully, but these errors were encountered: