Spaces:
Sleeping
Sleeping
Update fine_tuning.py
Browse files- fine_tuning.py +1 -1
fine_tuning.py
CHANGED
|
@@ -24,6 +24,7 @@ os.environ["WANDB_MODE"] = "disabled"
|
|
| 24 |
# -----------------------------
|
| 25 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 26 |
tokenizer = transformers.AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-Chat-v1.0")
|
|
|
|
| 27 |
|
| 28 |
# -----------------------------
|
| 29 |
# LoRA / MoE Modules
|
|
@@ -157,7 +158,6 @@ def load_and_train(model_id="TinyLlama/TinyLlama-1.1B-Chat-v1.0"):
|
|
| 157 |
print("Training started")
|
| 158 |
trainer.train()
|
| 159 |
model.eval()
|
| 160 |
-
return model, tokenizer, device
|
| 161 |
|
| 162 |
# -----------------------------
|
| 163 |
# GENERATE ANSWER
|
|
|
|
| 24 |
# -----------------------------
|
| 25 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 26 |
tokenizer = transformers.AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-Chat-v1.0")
|
| 27 |
+
model = None
|
| 28 |
|
| 29 |
# -----------------------------
|
| 30 |
# LoRA / MoE Modules
|
|
|
|
| 158 |
print("Training started")
|
| 159 |
trainer.train()
|
| 160 |
model.eval()
|
|
|
|
| 161 |
|
| 162 |
# -----------------------------
|
| 163 |
# GENERATE ANSWER
|