@@ -54,7 +54,7 @@ def download_alpaca():
54
54
1.00232565 , # expected_train_metric
55
55
0.02060609 , # expected_eval_loss
56
56
1.02081990 , # expected_eval_metric
57
- id = "llama_config " , # config name
57
+ id = "llama_config_samsum " , # config name
58
58
),
59
59
pytest .param (
60
60
"meta-llama/Llama-3.2-1B" , # model_name
@@ -72,7 +72,7 @@ def download_alpaca():
72
72
1.0043447 , # expected_train_metric
73
73
0.0117334 , # expected_eval_loss
74
74
1.0118025 , # expected_eval_metric
75
- id = "llama_config " , # config name
75
+ id = "llama_config_gsm8k " , # config name
76
76
),
77
77
pytest .param (
78
78
"meta-llama/Llama-3.2-1B" , # model_name
@@ -90,7 +90,7 @@ def download_alpaca():
90
90
1.0006101 , # expected_train_metric
91
91
0.0065296 , # expected_eval_loss
92
92
1.0065510 , # expected_eval_metric
93
- id = "llama_config " , # config name
93
+ id = "llama_config_alpaca " , # config name
94
94
),
95
95
pytest .param (
96
96
"google-bert/bert-base-uncased" , # model_name
@@ -108,7 +108,7 @@ def download_alpaca():
108
108
0.55554199 , # expected_train_metric
109
109
0.00738618 , # expected_eval_loss
110
110
0.70825195 , # expected_eval_metric
111
- id = "bert_config " , # config name
111
+ id = "bert_config_imdb " , # config name
112
112
),
113
113
]
114
114
0 commit comments