You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Consider a system where a feature service fetches model metadata that has information on what feature to fetch and finally infer from the model. In order for me fetch this metadata regarding inputs and outputs I am trying to use the recently added Open inference protocol.
while trying to infer using grpcurl, it shows me the name and version of the model.
Take an input metadata file that can be exposed on both gRPC and REST metadata endpoints. One example is on the lines of seldon metadata ep that exposes this information.
The text was updated successfully, but these errors were encountered:
🐛 Describe the bug
Consider a system where a feature service fetches model metadata that has information on what feature to fetch and finally infer from the model. In order for me fetch this metadata regarding inputs and outputs I am trying to use the recently added Open inference protocol.
while trying to infer using grpcurl, it shows me the name and version of the model.
with simple curl, the output is REST API does not add anything model custom to it.
I was trying to understand where it sets this metadata so i can impute it accordingly. I could not find a way for it to set inputs and outputs.
Do you know of how the metadata is set if so in torchserve.
Error logs
n/a
Installation instructions
Dockerfile on top of latest torchserve image
Model Packaing
mnist model can be used, independent of model type.
config.properties
inference_address=http://0.0.0.0:8080
management_address=http://0.0.0.0:8081
metrics_address=http://0.0.0.0:8082
enable_metrics_api=true
model_metrics_auto_detect=true
metrics_mode=prometheus
number_of_netty_threads=32
job_queue_size=1000
enable_envvars_config=true
model_store=/home/model-server/model-store
workflow_store=/home/model-server/wf-store
load_models=all
Versions
Environment headers
Torchserve branch:
**Warning: torchserve not installed ..
**Warning: torch-model-archiver not installed ..
Python version: 3.11 (64-bit runtime)
Python executable: /home/hmeena/.pyenv/versions/airflow/bin/python
Versions of relevant python libraries:
requests==2.31.0
**Warning: torch not present ..
**Warning: torchtext not present ..
**Warning: torchvision not present ..
**Warning: torchaudio not present ..
Java Version:
OS: CentOS Linux release 7.5.1804 (Core)
GCC version: (GCC) 4.8.5 20150623 (Red Hat 4.8.5-39)
Clang version: 3.4.2 (tags/RELEASE_34/dot2-final)
CMake version: N/A
Environment:
library_path (LD_/DYLD_): :/search/dist/bin:/search/dist/bin
Repro instructions
Model from old issue i created can be used.
Possible Solution
Take an input metadata file that can be exposed on both gRPC and REST metadata endpoints. One example is on the lines of seldon metadata ep that exposes this information.
The text was updated successfully, but these errors were encountered: