Skip to content

Commit a125b92

Browse files
committed
add more recipe
1 parent afec902 commit a125b92

File tree

4 files changed

+37
-4
lines changed

4 files changed

+37
-4
lines changed
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,16 @@
1+
target_task: tasks/nlp/translation.md
2+
base_model_id: yuiseki/Mistral-7B-v0.1-ja-wikipedia-v0.1
3+
model_name: Mistral-7B-v0.1-ja-wikipedia-OpenMath-v0.1
4+
output_base_dir: /data/output
5+
dataset_id: kunishou/OpenMathInstruct-1-1.8m-ja
6+
dataset_input_field_name: question_ja
7+
dataset_output_field_name: generated_solution_ja
8+
dataset_train_split_seed: 42
9+
dataset_train_split_test_size: 0.2
10+
lora_r: 8
11+
lora_alpha: 16
12+
lora_dropout: 0.05
13+
train_claim_gpu_num: 4
14+
train_per_device_train_batch_size: 8
15+
train_gradient_accumulation_steps: 2
16+
train_num_train_epochs: 4
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,17 @@
1+
target_task: tasks/text-generation/text2sql.md
2+
base_model_id: yuiseki/Mistral-7B-v0.1-ja-wikipedia-v0.1
3+
model_name: Mistral-7B-v0.1-ja-wikipedia-amenokaku-v0.1
4+
output_base_dir: /data/output
5+
dataset_id: kunishou/amenokaku-code-instruct
6+
dataset_input_field_name: instruction
7+
dataset_context_field_name: input
8+
dataset_output_field_name: output
9+
dataset_train_split_seed: 42
10+
dataset_train_split_test_size: 0.2
11+
lora_r: 8
12+
lora_alpha: 16
13+
lora_dropout: 0.05
14+
train_claim_gpu_num: 4
15+
train_per_device_train_batch_size: 8
16+
train_gradient_accumulation_steps: 4
17+
train_num_train_epochs: 4

recipes/A5000_24GB_x8/Mistral-7B-v0.1/i18n-ja-wikipedia-aya.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,6 @@ lora_r: 8
1313
lora_alpha: 16
1414
lora_dropout: 0.05
1515
train_claim_gpu_num: 8
16-
train_per_device_train_batch_size: 2
17-
train_gradient_accumulation_steps: 8
16+
train_per_device_train_batch_size: 1
17+
train_gradient_accumulation_steps: 16
1818
train_num_train_epochs: 2

recipes/A5000_24GB_x8/Mistral-7B-v0.1/i18n-ja-wikipedia.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,6 @@ lora_r: 8
1111
lora_alpha: 16
1212
lora_dropout: 0.05
1313
train_claim_gpu_num: 8
14-
train_per_device_train_batch_size: 2
15-
train_gradient_accumulation_steps: 8
14+
train_per_device_train_batch_size: 1
15+
train_gradient_accumulation_steps: 16
1616
train_num_train_epochs: 2

0 commit comments

Comments
 (0)