Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

@jonathanxu81205 - add llava-critic-6 #239

Closed
wants to merge 1 commit into from
Closed
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
22 changes: 22 additions & 0 deletions assets/bytedance.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -49,3 +49,25 @@
prohibited_uses: unknown
monitoring: unknown
feedback: https://huggingface.co/ByteDance/SDXL-Lightning/discussions
- type: model
name: LLaVA-Critic
organization: ByteDance, University of Maryland, College Park
description: LLaVA-Critic is an open-source large multimodal model (LMM), developed as a generalist evaluator to assess performance across a variety of multimodal tasks. It is designed to provide evaluation scores that are comparable to or exceed those of GPT models and to provide reward signals for preference learning, thereby enhancing model alignment capabilities. It builds on a high-quality dataset for critic instruction-following, enabling it to provide quantitative judgment and reasoning for its evaluations.
created_date: 2024-10-06
url: https://arxiv.org/pdf/2410.02712
model_card: unknown
modality: text, image; text, evaluation scores (judgement)
analysis: The model's effectiveness was demonstrated in providing evaluation scores reliably, showing high correlation with commercial GPT models and outperforming other models in preference learning by offering enhanced AI-generated feedback.
size: unknown
dependencies: [GPT-4V, LLaVA-Instruction-150k, SVIT, ComVint, LLaVAR, LRV-Instruction, M3IT, LLaVA-Med, PCA-EVAL, VLFeedback]
training_emissions: unknown
training_time: unknown
training_hardware: unknown
quality_control: The model uses a well-curated critic instruction-following dataset, provides transparency and consistency with its evaluations, and ensures clarity and comprehensiveness in the evaluation process.
access: open
license: unknown
intended_uses: Designed to serve as a reliable evaluator in multimodal contexts, useful for conducting model evaluations, generating reward signals for preference learning, and enhancing alignment in large multimodal models.
prohibited_uses: unknown
monitoring: unknown
feedback: unknown

Loading