Comments (3)
I tried to print basic information about the model using : print(onnx.helper.printable_graph(onnx_model.graph)) and got this :graph model-onnx (
%float_input[FLOAT, ?x8]
) {
%label, %probabilities = LinearClassifierclasslabels_ints = [0, 1], coefficients = [-1.15231204032898, -3.61196088790894, 0.323476761579514, -0.234493508934975, -0.0497973449528217, -2.38965225219727, -1.12923967838287, -0.840260863304138, 1.15231204032898, 3.61196088790894, -0.323476761579514, 0.234493508934975, 0.0497973449528217, 2.38965225219727, 1.12923967838287, 0.840260863304138], intercepts = [4.81376791000366, -4.81376791000366], multi_class = 0, post_transform = 'LOGISTIC'
%output_label = Castto = 7
%output_probability = ZipMapclasslabels_int64s = [0, 1]
return %output_label, %output_probability
}.
So is there anything I am missing in the config file?
from server.
Hello @Manishthakur2503 thanks for reaching out, this looks like a model repository error.
Is your model repository accessible inside the container?
Is the structure of your model repository?
/
/
config.pbtxt
/
model.onnx
Before you do
docker run --rm --name tritonserver -p 8000:8000 -p 8001:8001 -p 8002:8002 -v /mnt/host/c/Projects/ML/NvidiaTriton/model_repository:/models nvcr.io/nvidia/tritonserver:23.01-py3 tritonserver --model-repository=/models
Can you check if /models is as expected in the container?
Look at this tutorial https://github.com/triton-inference-server/server/blob/main/docs/getting_started/quickstart.md
This too load a onnx model the way you are trying to.
Thanks,
Indrajit
from server.
Hi @indrajit96
Thanks for your response.
I checked the model repository structure, and it was correct. However, the issue was with the input dimensions in the config file, which was causing the error. I have fixed it now, and everything is working fine.
Thanks,
Manish Thakur
from server.
Related Issues (20)
- Triton ensemble not working as expected to support reshape HOT 1
- nvcr.io: i/o timeout - error 401 HOT 1
- How to use pb_utils in python backend to receive data from cudashm?
- why is only 1st 'batch' inferred?
- How does the stateful model maintain state among multiple pods? HOT 3
- Minimal Custom Backend Example Not Working
- Model 'tensorrt_llm' loading failed with error: key 'use_context_fmha_for_generation' not found HOT 3
- Dynamic batching with OpenVINO backend
- In BLS mode, does the data go to the CPU HOT 2
- Handling Unsupported Input and Ensuring GPU Processing in Triton Inference Server HOT 1
- Using output of a model as input for multiple other models in ensemble give unclear error HOT 2
- my label_filename don't read I have class0 for label HOT 1
- gRPC Segfaults in Triton 24.05 due to Request Cancellation HOT 2
- Better doc for different between timeout and client_timeout of grpc_client.infer
- [k8s-on-prem] Timeout issue with Traefik deployment replicas more than 1
- What's the difference when starting tritonserver with `mpirun --allow-run-as-root -n 1 /opt/tritonserver/bin/tritonserver` vs. `/opt/tritonserver/bin/tritonserver` directly?
- As the number of CPU cores decreases, the BLS mode processing time increases
- Prebuilt Triton Server 24.05-trtllm-python-py3 does not have correct TensorRT version HOT 5
- Proposal for a Generic Custom Backend Template
- Decoupled mode, dimensionality explosion
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from server.