diff --git a/assets/alibaba.yaml b/assets/alibaba.yaml index 496e4dc9..2e900a97 100644 --- a/assets/alibaba.yaml +++ b/assets/alibaba.yaml @@ -87,7 +87,7 @@ - type: model name: Qwen 1.5 - organization: Qwen AI + organization: Qwen Team description: Qwen 1.5 is the next iteration in their Qwen series, consisting of Transformer-based large language models pretrained on a large volume of data, including web texts, books, codes, etc. @@ -95,8 +95,9 @@ url: https://qwenlm.github.io/blog/qwen1.5/ model_card: https://huggingface.co/Qwen/Qwen1.5-72B modality: text; text - analysis: Evaluated on MMLU, C-Eval, GSM8K, MATH, HumanEval, MBPP, BBH, CMMLU, - all standard English and Chinese benchmarks. + analysis: Base models are evaluated on MMLU, C-Eval, GSM8K, MATH, HumanEval, MBPP, + BBH, CMMLU, all standard English and Chinese benchmarks, and chat models are + evaluated on Chatbot Arena, AlpacaEval, MT-Bench, etc. size: 72B parameters (dense) dependencies: [] training_emissions: unknown @@ -111,3 +112,32 @@ prohibited_uses: '' monitoring: unknown feedback: https://huggingface.co/Qwen/Qwen1.5-72B/discussions + +- type: model + name: Qwen 1.5 MoE + organization: Qwen Team + description: Qwen 1.5 is the next iteration in their Qwen series, consisting of + Transformer-based large language models pretrained on a large volume of data, + including web texts, books, codes, etc. Qwen 1.5 MoE is the MoE model of the + Qwen 1.5 series. + created_date: 2024-03-28 + url: https://qwenlm.github.io/blog/qwen-moe/ + model_card: https://huggingface.co/Qwen/Qwen1.5-MoE-A2.7B + modality: text; text + analysis: Base models are evaluated on MMLU, C-Eval, GSM8K, MATH, HumanEval, MBPP, + BBH, CMMLU, all standard English and Chinese benchmarks, and chat models are + evaluated on Chatbot Arena, AlpacaEval, MT-Bench, etc. + size: 14B parameters with 2.7B parameters for activation (MoE) + dependencies: [] + training_emissions: unknown + training_time: unknown + training_hardware: unknown + quality_control: unknown + access: open + license: + explanation: Model license can be found at https://huggingface.co/Qwen/Qwen1.5-MoE-A2.7B/blob/main/LICENSE + value: custom + intended_uses: '' + prohibited_uses: '' + monitoring: unknown + feedback: https://huggingface.co/Qwen/Qwen1.5-MoE-A2.7B/discussions