Skip to content

Commit 492787b

Browse files
committed
add nationality
1 parent fd820a8 commit 492787b

File tree

175 files changed

+16106
-13925
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

175 files changed

+16106
-13925
lines changed

assets/01ai.yaml

+40-39
Original file line numberDiff line numberDiff line change
@@ -1,56 +1,57 @@
1-
---
2-
- type: model
3-
name: Yi
4-
organization: 01 AI
5-
description: The Yi series models are large language models trained from scratch
6-
by developers at 01 AI.
7-
created_date: 2023-11-02
8-
url: https://github.com/01-ai/Yi
9-
model_card: https://huggingface.co/01-ai/Yi-34B
10-
modality: text; text
1+
- access: open
112
analysis: Evaluated on standard language benchmarks, common sense reasoning, and
123
reading comprehension in comparison to SoTA LLMs.
13-
size: 34B parameters (dense)
4+
created_date: 2023-11-02
145
dependencies: []
15-
training_emissions: unknown
16-
training_time: unknown
17-
training_hardware: unknown
18-
quality_control: Model underwent supervised fine-tuning, leading to a greater
19-
diversity of responses.
20-
access: open
6+
description: The Yi series models are large language models trained from scratch
7+
by developers at 01 AI.
8+
feedback: https://huggingface.co/01-ai/Yi-34B/discussions
9+
intended_uses: ''
2110
license:
2211
explanation: Model license can be found at https://huggingface.co/01-ai/Yi-34B/blob/main/LICENSE.
2312
Code license is under Apache 2.0
2413
value: custom
25-
intended_uses: ''
26-
prohibited_uses: none
14+
modality: text; text
15+
model_card: https://huggingface.co/01-ai/Yi-34B
2716
monitoring: unknown
28-
feedback: https://huggingface.co/01-ai/Yi-34B/discussions
29-
- type: model
30-
name: Yi-VL
17+
name: Yi
18+
nationality: France
3119
organization: 01 AI
32-
description: The Yi Vision Language (Yi-VL) model is the open-source, multimodal
33-
version of the Yi Large Language Model (LLM) series, enabling content comprehension,
34-
recognition, and multi-round conversations about images.
35-
created_date: 2024-01-23
20+
prohibited_uses: none
21+
quality_control: Model underwent supervised fine-tuning, leading to a greater diversity
22+
of responses.
23+
size: 34B parameters (dense)
24+
training_emissions: unknown
25+
training_hardware: unknown
26+
training_time: unknown
27+
type: model
3628
url: https://github.com/01-ai/Yi
37-
model_card: https://huggingface.co/01-ai/Yi-VL-34B
38-
modality: text; text
39-
analysis: Yi-VL outperforms all existing open-source models in MMMU and CMMMU,
40-
two advanced benchmarks that include massive multi-discipline multimodal questions
29+
- access: open
30+
analysis: Yi-VL outperforms all existing open-source models in MMMU and CMMMU, two
31+
advanced benchmarks that include massive multi-discipline multimodal questions
4132
(based on data available up to January 2024).
42-
size: 34B parameters (dense)
33+
created_date: 2024-01-23
4334
dependencies: []
44-
training_emissions: unknown
45-
training_time: 10 days
46-
training_hardware: 128 NVIDIA A800 (80G) GPUs
47-
quality_control: unknown
48-
access: open
35+
description: The Yi Vision Language (Yi-VL) model is the open-source, multimodal
36+
version of the Yi Large Language Model (LLM) series, enabling content comprehension,
37+
recognition, and multi-round conversations about images.
38+
feedback: https://huggingface.co/01-ai/Yi-VL-34B/discussions
39+
intended_uses: ''
4940
license:
5041
explanation: Model license can be found at https://huggingface.co/01-ai/Yi-VL-34B/blob/main/LICENSE.
5142
Code license is under Apache 2.0
5243
value: custom
53-
intended_uses: ''
54-
prohibited_uses: ''
44+
modality: text; text
45+
model_card: https://huggingface.co/01-ai/Yi-VL-34B
5546
monitoring: unknown
56-
feedback: https://huggingface.co/01-ai/Yi-VL-34B/discussions
47+
name: Yi-VL
48+
nationality: France
49+
organization: 01 AI
50+
prohibited_uses: ''
51+
quality_control: unknown
52+
size: 34B parameters (dense)
53+
training_emissions: unknown
54+
training_hardware: 128 NVIDIA A800 (80G) GPUs
55+
training_time: 10 days
56+
type: model
57+
url: https://github.com/01-ai/Yi

assets/360.yaml

+18-19
Original file line numberDiff line numberDiff line change
@@ -1,24 +1,23 @@
1-
---
2-
- type: model
3-
name: 360 Zhinao
4-
organization: 360 Security
5-
description: 360 Zhinao is a multilingual LLM in Chinese and English with chat
6-
capabilities.
7-
created_date: 2024-05-23
8-
url: https://arxiv.org/pdf/2405.13386
9-
model_card: none
10-
modality: text; text
1+
- access: open
112
analysis: Achieved competitive performance on relevant benchmarks against other
123
7B models in Chinese, English, and coding tasks.
13-
size: 7B parameters
4+
created_date: 2024-05-23
145
dependencies: []
15-
training_emissions: unknown
16-
training_time: unknown
17-
training_hardware: unknwon
18-
quality_control: ''
19-
access: open
20-
license: unknown
6+
description: 360 Zhinao is a multilingual LLM in Chinese and English with chat capabilities.
7+
feedback: none
218
intended_uses: ''
22-
prohibited_uses: ''
9+
license: unknown
10+
modality: text; text
11+
model_card: none
2312
monitoring: ''
24-
feedback: none
13+
name: 360 Zhinao
14+
nationality: China
15+
organization: 360 Security
16+
prohibited_uses: ''
17+
quality_control: ''
18+
size: 7B parameters
19+
training_emissions: unknown
20+
training_hardware: unknwon
21+
training_time: unknown
22+
type: model
23+
url: https://arxiv.org/pdf/2405.13386

assets/adept.yaml

+74-71
Original file line numberDiff line numberDiff line change
@@ -1,100 +1,103 @@
1-
---
2-
- type: model
3-
name: ACT-1
4-
organization: Adept
5-
description: ACT-1 (ACtion Transformer) is a large-scale transformer model designed
6-
and trained specifically for taking actions on computers (use software tools
7-
APIs and websites) in response to the user's natural language commands.
1+
- access: closed
2+
analysis: ''
83
created_date:
94
explanation: The date the model was announced in the [[Adept blog post]](https://www.adept.ai/blog/act-1).
105
value: 2022-09-14
11-
url: https://www.adept.ai/blog/act-1
12-
model_card: none
6+
dependencies: []
7+
description: ACT-1 (ACtion Transformer) is a large-scale transformer model designed
8+
and trained specifically for taking actions on computers (use software tools APIs
9+
and websites) in response to the user's natural language commands.
10+
feedback: ''
11+
intended_uses: ''
12+
license: unknown
1313
modality: text; text
14-
analysis: ''
14+
model_card: none
15+
monitoring: ''
16+
name: ACT-1
17+
nationality: USA
18+
organization: Adept
19+
prohibited_uses: ''
20+
quality_control: ''
1521
size: ''
16-
dependencies: []
1722
training_emissions: unknown
18-
training_time: unknown
1923
training_hardware: unknown
20-
quality_control: ''
21-
access: closed
22-
license: unknown
24+
training_time: unknown
25+
type: model
26+
url: https://www.adept.ai/blog/act-1
27+
- access: open
28+
analysis: Evaluated in comparison to LLaMA 2 and MPT Instruct, and outperforms both
29+
on standard benchmarks.
30+
created_date: 2023-09-07
31+
dependencies: []
32+
description: Persimmon is the most capable open-source, fully permissive model with
33+
fewer than 10 billion parameters, as of its release date.
34+
feedback: ''
2335
intended_uses: ''
24-
prohibited_uses: ''
36+
license: Apache 2.0
37+
modality: text; text
38+
model_card: ''
2539
monitoring: ''
26-
feedback: ''
27-
- type: model
2840
name: Persimmon
41+
nationality: USA
2942
organization: Adept
30-
description: Persimmon is the most capable open-source, fully permissive model
31-
with fewer than 10 billion parameters, as of its release date.
32-
created_date: 2023-09-07
33-
url: https://www.adept.ai/blog/persimmon-8b
34-
model_card: ''
35-
modality: text; text
36-
analysis: Evaluated in comparison to LLaMA 2 and MPT Instruct, and outperforms
37-
both on standard benchmarks.
43+
prohibited_uses: ''
44+
quality_control: ''
3845
size: 8B parameters (dense)
39-
dependencies: []
4046
training_emissions: ''
41-
training_time: ''
4247
training_hardware: ''
43-
quality_control: ''
44-
access: open
45-
license: Apache 2.0
46-
intended_uses: ''
47-
prohibited_uses: ''
48-
monitoring: ''
49-
feedback: ''
50-
- type: model
51-
name: Fuyu
52-
organization: Adept
48+
training_time: ''
49+
type: model
50+
url: https://www.adept.ai/blog/persimmon-8b
51+
- access: open
52+
analysis: Evaluated on standard image understanding benchmarks.
53+
created_date: 2023-10-17
54+
dependencies: []
5355
description: Fuyu is a small version of the multimodal model that powers Adept's
5456
core product.
55-
created_date: 2023-10-17
56-
url: https://www.adept.ai/blog/fuyu-8b
57-
model_card: https://huggingface.co/adept/fuyu-8b
57+
feedback: https://huggingface.co/adept/fuyu-8b/discussions
58+
intended_uses: The model is intended for research purposes only.
59+
license: CC-BY-NC-4.0
5860
modality: image, text; text
59-
analysis: Evaluated on standard image understanding benchmarks.
61+
model_card: https://huggingface.co/adept/fuyu-8b
62+
monitoring: ''
63+
name: Fuyu
64+
nationality: USA
65+
organization: Adept
66+
prohibited_uses: The model was not trained to be factual or true representations
67+
of people or events, and therefore using the model to generate such content is
68+
out-of-scope for the abilities of this model.
69+
quality_control: none
6070
size: 8B parameters (dense)
61-
dependencies: []
6271
training_emissions: unknown
63-
training_time: unknown
6472
training_hardware: unknown
65-
quality_control: none
66-
access: open
67-
license: CC-BY-NC-4.0
68-
intended_uses: The model is intended for research purposes only.
69-
prohibited_uses: The model was not trained to be factual or true representations
70-
of people or events, and therefore using the model to generate such content
71-
is out-of-scope for the abilities of this model.
72-
monitoring: ''
73-
feedback: https://huggingface.co/adept/fuyu-8b/discussions
74-
- type: model
75-
name: Fuyu Heavy
76-
organization: Adept
73+
training_time: unknown
74+
type: model
75+
url: https://www.adept.ai/blog/fuyu-8b
76+
- access: closed
77+
analysis: Evaluated on the MMLU, GSM8K, MATH, and HumanEval benchmarks. According
78+
to these benchmarks, Fuyu-Heavy is, as of release, the strongest multimodal model
79+
trained outside of Google or OpenAI.
80+
created_date: 2024-01-24
81+
dependencies: []
7782
description: Fuyu Heavy is a new multimodal model designed specifically for digital
7883
agents.
79-
created_date: 2024-01-24
80-
url: https://www.adept.ai/blog/adept-fuyu-heavy
81-
model_card: none
84+
feedback: none
85+
intended_uses: unknown
86+
license: unknown
8287
modality: image, text; text
83-
analysis: Evaluated on the MMLU, GSM8K, MATH, and HumanEval benchmarks. According
84-
to these benchmarks, Fuyu-Heavy is, as of release, the strongest multimodal
85-
model trained outside of Google or OpenAI.
88+
model_card: none
89+
monitoring: ''
90+
name: Fuyu Heavy
91+
nationality: USA
92+
organization: Adept
93+
prohibited_uses: none
94+
quality_control: none
8695
size:
8796
explanation: The size of the model is 10-20 times smaller than GPT-4V and Gemini
8897
Ultra, as per announcement.
8998
value: unknown
90-
dependencies: []
9199
training_emissions: unknown
92-
training_time: unknown
93100
training_hardware: unknown
94-
quality_control: none
95-
access: closed
96-
license: unknown
97-
intended_uses: unknown
98-
prohibited_uses: none
99-
monitoring: ''
100-
feedback: none
101+
training_time: unknown
102+
type: model
103+
url: https://www.adept.ai/blog/adept-fuyu-heavy

0 commit comments

Comments
 (0)