-
-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
API details #560
Comments
👋 Hello @LiangA, thank you for raising an issue about Ultralytics HUB 🚀! Please visit our HUB Docs to learn more:
If this is a 🐛 Bug Report, please provide screenshots and steps to reproduce your problem to help us get started working on a fix. If this is a ❓ Question, please provide as much information as possible, including dataset, model, environment details etc. so that we might provide the most helpful response. We try to respond to all issues as promptly as possible. Thank you for your patience! |
@LiangA I am not sure I am understanding your question, are you asking about implementing your own custom API or using the API provided by HUB? We currently provide a shared inference option in both the free and pro plans with defined rate limits. They do not include any additional customization. We have plans to introduce a dedicated inference option that will allow some more customization. |
Hi Kalenmike, Looks like Hub doesn't provide this service now, am I understand you correctly? |
@LiangA You have access to the shared HUB API at the moment. All of your models have an endpoint that will return inference results. Its important to understand that as these are shared they are also stateless so its more of a solution for previewing and testing than for production. The shared inference API has higher latency as your model needs to be fetched and loaded on each request, this results in slower responses. The dedicated inference API which is still in the development pipeline deploys a scalable API that is preloaded with your model. This has lower latency as your model is ready to go on each request. We are finishing up our Cloud Training feature and it is being prioritized over the dedicated inference API. So the development work is backlogged for the moment. These are all cloud solution where Ultralytics handles the hardware and software implementations for you. You can of course implement this all yourself using the ultralytics python package and your own hardware/software solution. You can demo the shared API using a free HUB account by following these steps:
Free accounts have the following rate limits applied: |
Thanks for making this crystal clear. I'll keep an eye on the updates. |
You're welcome, @LiangA! If you have any more questions or need further assistance in the future, don't hesitate to reach out. Happy to help, and we appreciate your interest in Ultralytics HUB! Keep an eye on our updates for new features. 😊 Have a great day! |
Search before asking
Question
I would like to create an inference API, caring about the accessibility, reliability security, and, of course, cost. But I'm not seeing any detail information except that 10,000 calls/month in HUB Pro plan.
I'm also curious about what tools do I have? (I'm an AWS user, so terms like S3, EC2, IAM or VPC can help me understand)
If it's possible, I would like to know about the service level too.
Thanks!
Additional
No response
The text was updated successfully, but these errors were encountered: