Skip to content
Snippets Groups Projects
Commit ce7bd4f8 authored by Julien Breton's avatar Julien Breton
Browse files

change hyper parameters

parent 15f8302d
Branches
No related tags found
No related merge requests found
......@@ -56,13 +56,13 @@ def fine_tune(base_model, new_model):
# Training Arguments
# Hyperparameters should beadjusted based on the hardware you using
training_arguments = TrainingArguments(
per_device_train_batch_size=1,
gradient_accumulation_steps=4,
num_train_epochs=6,
per_device_train_batch_size=2,
gradient_accumulation_steps=1,
num_train_epochs=2,
learning_rate=1e-4,
logging_steps=2,
optim="adamw_torch",
save_strategy="epoch",
save_strategy="steps",
output_dir="./results"
)
......
......@@ -70,13 +70,13 @@ def fine_tuned(base_model, new_model):
model=model,
train_dataset=train_data,
args=TrainingArguments(
per_device_train_batch_size=1,
gradient_accumulation_steps=4,
num_train_epochs=6,
per_device_train_batch_size=2,
gradient_accumulation_steps=1,
num_train_epochs=2,
learning_rate=1e-4,
logging_steps=2,
optim="adamw_torch",
save_strategy="epoch",
save_strategy="steps",
output_dir="./results"
),
data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False)
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment