Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Try OV-DINO, a more powerful open-vocabulary detector. #452

Open
wanghao9610 opened this issue Jul 30, 2024 · 2 comments
Open

Try OV-DINO, a more powerful open-vocabulary detector. #452

wanghao9610 opened this issue Jul 30, 2024 · 2 comments

Comments

@wanghao9610
Copy link

wanghao9610 commented Jul 30, 2024

Thanks for the awesome YOLO-World, I share our recent work 🦖OV-DINO: Unified Open-Vocabulary Detection with Language-Aware Selective Fusion.

  • OV-DINO is a novel unified open vocabulary detection approach that offers superior performance and effectiveness for practical real-world application.

  • OV-DINO entails a Unified Data Integration pipeline that integrates diverse data sources for end-to-end pre-training, and a Language-Aware Selective Fusion module to improve the vision-language understanding of the model.

  • OV-DINO shows significant performance improvement on COCO and LVIS benchmarks compared to previous methods, achieving relative improvements of +5.5% AP on COCO and +4.7% AP on LVIS compared to YOLO-World in zero-shot evaluation.

We have released the evaluation, fine-tuning, demo code in our project, feel free to try our model for your application.

Project: https://wanghao9610.github.io/OV-DINO

Paper: https://arxiv.org/abs/2407.07844

Code: https://github.com/wanghao9610/OV-DINO

Demo: http://47.115.200.157:7860

Welcome everyone to try our model and feel free to raise issue if you encounter any problem.

@wanghao9610 wanghao9610 changed the title Try OV-DINO, a more strong open-vocabulary detector. Try OV-DINO, a more powerful open-vocabulary detector. Jul 30, 2024
@ForestWang
Copy link

how about the inference time compared with yolo-world?

@wanghao9610
Copy link
Author

@ForestWang We haven't tested the inference time systematically, the inference time is already fast according to the demo without any deployment optimization.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants