You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"Loading of grammar dataset failed! Please see [here](https://github.com/meta-llama/llama-recipes/blob/main/src/llama_recipes/datasets/grammar_dataset/grammar_dataset_process.ipynb) for details on how to download the dataset."
26
+
"Loading of grammar dataset failed! Please check (https://github.com/meta-llama/llama-recipes/blob/main/src/llama_recipes/datasets/grammar_dataset/grammar_dataset_process.ipynb) for details on how to download the dataset."
f"Not proceeding with epoch {epoch+1} on device {local_rank} since loss value has been <= {train_config.convergence_loss} for last {loss_0_counter.item()} steps."
119
-
)
120
-
break
121
-
else:
113
+
if (nottrain_config.enable_ddp) or (train_config.enable_ddpandlocal_rank==0):
122
114
logger.info(
123
-
f"Not proceeding with epoch {epoch+1} since loss value has been <= {train_config.convergence_loss} for last {loss_0_counter.item()} steps."
115
+
f"Skipping epoch {epoch+1} since loss value has been <= {train_config.convergence_loss} for last {loss_0_counter.item()} steps."
124
116
)
125
117
break
126
118
@@ -161,7 +153,7 @@ def train(
161
153
ifepoch==intermediate_epochandstep==0:
162
154
total_train_steps+=intermediate_step
163
155
logger.info(
164
-
f"skipping first {intermediate_step} steps for epoch {epoch+1}, since fine tuning has already completed for them."
156
+
f"Skipping first {intermediate_step} steps for epoch {epoch+1}, since fine tuning has already completed for it."
f"Loss value has been <= {train_config.convergence_loss} for last {loss_0_counter.item()} steps. Hence, stopping the fine tuning on device {local_rank}."
282
-
)
283
-
break
284
-
else:
267
+
if (nottrain_config.enable_ddp) or (train_config.enable_ddpandlocal_rank==0):
0 commit comments