From 73e661e4dbdaeac2b158c2a9a89ec5cd05626161 Mon Sep 17 00:00:00 2001
From: Julien Breton <julien.breton@moncitron.fr>
Date: Sun, 4 Feb 2024 14:49:23 +0900
Subject: [PATCH] change hyper parameters

---
 modules/llm/Mistral-7b/Mistral-7b_fine_tune.py     | 1 -
 modules/llm/Mixtral-8x7b/Mixtral-8x7b_fine_tune.py | 1 -
 2 files changed, 2 deletions(-)

diff --git a/modules/llm/Mistral-7b/Mistral-7b_fine_tune.py b/modules/llm/Mistral-7b/Mistral-7b_fine_tune.py
index a41fc61..99a152d 100644
--- a/modules/llm/Mistral-7b/Mistral-7b_fine_tune.py
+++ b/modules/llm/Mistral-7b/Mistral-7b_fine_tune.py
@@ -58,7 +58,6 @@ def fine_tune(base_model, new_model):
     training_arguments = TrainingArguments(
         per_device_train_batch_size=2,
         gradient_accumulation_steps=1,
-        num_train_epochs=2,
         learning_rate=1e-4,
         logging_steps=2,
         optim="adamw_torch",
diff --git a/modules/llm/Mixtral-8x7b/Mixtral-8x7b_fine_tune.py b/modules/llm/Mixtral-8x7b/Mixtral-8x7b_fine_tune.py
index d9e8126..9210d02 100644
--- a/modules/llm/Mixtral-8x7b/Mixtral-8x7b_fine_tune.py
+++ b/modules/llm/Mixtral-8x7b/Mixtral-8x7b_fine_tune.py
@@ -72,7 +72,6 @@ def fine_tuned(base_model, new_model):
         args=TrainingArguments(
             per_device_train_batch_size=2,
             gradient_accumulation_steps=1,
-            num_train_epochs=2,
             learning_rate=1e-4,
             logging_steps=2,
             optim="adamw_torch",
-- 
GitLab