-
Notifications
You must be signed in to change notification settings - Fork 35
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #224 from stanford-crfm/revert-222-revert-217-jona…
…than/090524-monthly-assets Revert "Revert "add notable summer assets""
- Loading branch information
Showing
24 changed files
with
1,022 additions
and
32 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -99,3 +99,39 @@ | |
prohibited_uses: '' | ||
monitoring: '' | ||
feedback: '' | ||
- type: model | ||
name: Pharia-1-LLM-7B | ||
organization: Aleph Alpha | ||
description: Pharia-1-LLM-7B is a model that falls within the Pharia-1-LLM model | ||
family. It is designed to deliver short, controlled responses that match the | ||
performance of leading open-source models around 7-8 billion parameters. The | ||
model is culturally and linguistically tuned for German, French, and Spanish | ||
languages. It is trained on carefully curated data in line with relevant EU | ||
and national regulations. The model shows improved token efficiency and is particularly | ||
effective in domain-specific applications, especially in the automotive and | ||
engineering industries. It can also be aligned to user preferences, making it | ||
appropriate for critical applications without the risk of shut-down behaviour. | ||
created_date: 2024-09-08 | ||
url: https://aleph-alpha.com/introducing-pharia-1-llm-transparent-and-compliant/#:~:text=Pharia%2D1%2DLLM%2D7B | ||
model_card: unknown | ||
modality: text; text | ||
analysis: Extensive evaluations were done with ablation experiments performed | ||
on pre-training benchmarks such as lambada, triviaqa, hellaswag, winogrande, | ||
webqs, arc, and boolq. Direct comparisons were also performed with applications | ||
like GPT and Llama 2. | ||
size: 7B parameters | ||
dependencies: [] | ||
training_emissions: Unknown | ||
training_time: Unknown | ||
training_hardware: Unknown | ||
quality_control: The model comes with additional safety guardrails via alignment | ||
methods to ensure safe usage. Training data is carefully curated to ensure compliance | ||
with EU and national regulations. | ||
access: open | ||
license: Aleph Open | ||
intended_uses: The model is intended for use in domain-specific applications, | ||
particularly in the automotive and engineering industries. It can also be tailored | ||
to user preferences. | ||
prohibited_uses: Unknown | ||
monitoring: Unknown | ||
feedback: Feedback can be sent to [email protected]. |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,34 @@ | ||
--- | ||
- type: model | ||
name: AstroPT | ||
organization: Aspia Space, Instituto de Astrofísica de Canarias (IAC), UniverseTBD, | ||
Astrophysics Research Institute, Liverpool John Moores University, Departamento | ||
Astrofísica, Universidad de la Laguna, Observatoire de Paris, LERMA, PSL University, | ||
and Universit´e Paris-Cit´e. | ||
description: AstroPT is an autoregressive pretrained transformer developed with | ||
astronomical use-cases in mind. The models have been pretrained on 8.6 million | ||
512x512 pixel grz-band galaxy postage stamp observations from the DESI Legacy | ||
Survey DR8. They have created a range of models with varying complexity, ranging | ||
from 1 million to 2.1 billion parameters. | ||
created_date: 2024-09-08 | ||
url: https://arxiv.org/pdf/2405.14930v1 | ||
model_card: unknown | ||
modality: image; image | ||
analysis: The models’ performance on downstream tasks was evaluated by linear | ||
probing. The models follow a similar saturating log-log scaling law to textual | ||
models, their performance improves with the increase in model size up to the | ||
saturation point of parameters. | ||
size: 2.1B parameters | ||
dependencies: [DESI Legacy Survey DR8] | ||
training_emissions: Unknown | ||
training_time: Unknown | ||
training_hardware: Unknown | ||
quality_control: The models’ performances were evaluated on downstream tasks as | ||
measured by linear probing. | ||
access: open | ||
license: MIT | ||
intended_uses: The models are intended for astronomical use-cases, particularly | ||
in handling and interpreting large observation data from astronomical sources. | ||
prohibited_uses: Unknown | ||
monitoring: Unknown | ||
feedback: Any problem with the model can be reported to Michael J. Smith at [email protected]. |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -2,12 +2,18 @@ | |
- type: model | ||
name: Sonic | ||
organization: Cartesia | ||
description: Sonic is a low-latency voice model that generates lifelike speech. Developed by Cartesia, it was designed to be an efficient real-time AI capable of processing any-sized contexts and running on any device. | ||
description: Sonic is a low-latency voice model that generates lifelike speech. | ||
Developed by Cartesia, it was designed to be an efficient real-time AI capable | ||
of processing any-sized contexts and running on any device. | ||
created_date: 2024-05-29 | ||
url: https://cartesia.ai/blog/sonic | ||
model_card: none | ||
modality: text; audio | ||
analysis: Extensive testing on Multilingual Librispeech dataset resulted in 20% lower validation perplexity. In downstream evaluations, this leads to a 2x lower word error rate and a 1 point higher quality score. Sonic also displays impressive performance metrics at inference, achieving lower latency (1.5x lower time-to-first-audio), faster inference speed (2x lower real-time factor), and higher throughput (4x). | ||
analysis: Extensive testing on Multilingual Librispeech dataset resulted in 20% | ||
lower validation perplexity. In downstream evaluations, this leads to a 2x lower | ||
word error rate and a 1 point higher quality score. Sonic also displays impressive | ||
performance metrics at inference, achieving lower latency (1.5x lower time-to-first-audio), | ||
faster inference speed (2x lower real-time factor), and higher throughput (4x). | ||
size: 2024-05-29 | ||
dependencies: [Multilingual Librispeech dataset] | ||
training_emissions: unknown | ||
|
@@ -16,7 +22,9 @@ | |
quality_control: '' | ||
access: limited | ||
license: unknown | ||
intended_uses: Sonic has potential applications across customer support, entertainment, and content creation and is a part of Cartesias broader mission to bring real-time multimodal intelligence to every device. | ||
intended_uses: Sonic has potential applications across customer support, entertainment, | ||
and content creation and is a part of Cartesias broader mission to bring real-time | ||
multimodal intelligence to every device. | ||
prohibited_uses: unknown | ||
monitoring: unknown | ||
feedback: Contact through the provided form or via email at [email protected]. |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.