Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ONNX CUDA session not working in python backend #7719

Open
jsoto-gladia opened this issue Oct 18, 2024 · 3 comments
Open

ONNX CUDA session not working in python backend #7719

jsoto-gladia opened this issue Oct 18, 2024 · 3 comments

Comments

@jsoto-gladia
Copy link

jsoto-gladia commented Oct 18, 2024

Bug Description
The ONNX CUDA session is not working in the Python backend. When attempting to run inference using the ONNX model with CUDAExecutionProvider, the session fails to initialize or execute properly.

Triton Information
Triton version: 22.07
Using Triton container: [Yes]

To Reproduce
Steps to reproduce the behavior:
https://github.com/jsoto-gladia/onnx-in-python-backend

When I use CPUExecutionProvider, everything works fine
When I use CUDAExecutionProvider, i get the following lines of error

I1018 20:11:24.764594 1 python_be.cc:2248] TRITONBACKEND_ModelInstanceFinalize: delete instance state
I1018 20:11:24.773726 1 python_be.cc:2087] TRITONBACKEND_ModelFinalize: delete model state
E1018 20:11:24.773855 1 model_lifecycle.cc:626] failed to load 'onnx_in_python_backend' version 1: Internal: Stub process 'onnx_in_python_backend_0' is not healthy.
I1018 20:11:24.773900 1 model_lifecycle.cc:755] failed to load 'onnx_in_python_backend'

Expected behavior
The ONNX model should initialize and execute properly using the CUDAExecutionProvider, leveraging GPU acceleration for inference.

@jsoto-gladia
Copy link
Author

inside the container i have logged:

triton-server-all-1  | cuda: 12.1
triton-server-all-1  | cudnn: 90100
triton-server-all-1  | onnxruntime: 1.19.2
triton-server-all-1  | torch: 2.4.1+cu121
triton-server-all-1  | providers: ['TensorrtExecutionProvider', 'CUDAExecutionProvider', 'CPUExecutionProvider']

@jsoto-gladia
Copy link
Author

on the machine when doing nvidia-smi, I get
NVIDIA-SMI 560.28.03 Driver Version: 560.28.03 CUDA Version: 12.6

@xiangkanghuang
Copy link

I'm facing the same problem, are there any solutions?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

No branches or pull requests

2 participants