Comments (2)
@jackylu0124 Support for onnxruntime-genai is currently work in progress - the python bindings should work within the python backend - but we haven't had a chance to test that ourselves yet.
That being said we are actively investigating support - can you share more about your use case / timeline needed for support?
from server.
@jackylu0124 Support for onnxruntime-genai is currently work in progress - the python bindings should work within the python backend - but we haven't had a chance to test that ourselves yet.
That being said we are actively investigating support - can you share more about your use case / timeline needed for support?
Hi @nnshah1 , thank you very much for your fast reply! By "the python bindings should work within the python backend", you meant that I can do things like import onnxruntime_genai
and write the custom inference logic in the Python backend, as opposed to having Triton Inference Server automatically manage all my .onnx model files (that use onnxruntime-genai) in the model repository for me (which is a feature currently in development), is my understanding correct?
My use case is mainly for serving LLM models, where some of which are in the form of ONNX models that depend on onnxruntime_genai
. I don't have a specific timeline, I am mainly interested in knowing whether this feature is on Triton Inferencer Server's development roadmap or not.
Also a follow-up question: regarding serving LLM, what would be the best backend for serving and achieving token streaming outside of using the TensorRT-LLM backend?
Thanks!
from server.
Related Issues (20)
- CUDA Failing to initialize in docker container HOT 3
- Add to the serve-side metrics on the input and output sizes HOT 1
- Pods Receiving Traffic Too Early When Scaling with HPA Causes 'Socket Closed' Errors on Triton Inference Server
- Custom backend using recommended.cc not generating correct output HOT 1
- docker image for triton 24.04 has incorrect CUDA version reported HOT 2
- Tritonserver hangs on launch with python backend HOT 1
- the method hang
- How to deploy Triton Inference Server Container (tritonserver:24.04-trtllm-python-py3) in K8S without launching Triton Server directly?
- Triton BLS model with dynamic batching does not execute expected batch size.
- [Bug] Model 'ensemble' receives inputs originated from different decoupled models
- No trtllm tag in ngc for 24.05 HOT 4
- No 24.05-trtllm-python-py3 in NGC Repo HOT 2
- YOLOv8n-poses is giving me a negative output error HOT 2
- Automatically unload (oldest) models when memory is full HOT 2
- Specific structure for ensemble model may causes deadlock
- Windows 10 docker build Error "Could not locate a complete Visual Studio instance" HOT 2
- A Confusion about prefetch HOT 2
- What is the correct way to run inference in parallel in Triton?
- Support histogram custom metric in Python backend HOT 2
- Backend support for .keras files?
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from server.