Upload train_qwen3_vl.py with huggingface_hub
Browse files- train_qwen3_vl.py +2 -2
train_qwen3_vl.py
CHANGED
|
@@ -33,7 +33,7 @@ print(f"Train: {len(train_dataset)}, Eval: {len(eval_dataset)}")
|
|
| 33 |
|
| 34 |
# Configure trainer with VL-specific settings
|
| 35 |
trainer = SFTTrainer(
|
| 36 |
-
model="Qwen/
|
| 37 |
train_dataset=train_dataset,
|
| 38 |
eval_dataset=eval_dataset,
|
| 39 |
peft_config=LoraConfig(
|
|
@@ -47,7 +47,7 @@ trainer = SFTTrainer(
|
|
| 47 |
args=SFTConfig(
|
| 48 |
output_dir="qwen3-vl-3b-llava-instruct",
|
| 49 |
push_to_hub=True,
|
| 50 |
-
hub_model_id="merve/
|
| 51 |
num_train_epochs=3,
|
| 52 |
per_device_train_batch_size=1,
|
| 53 |
gradient_accumulation_steps=8,
|
|
|
|
| 33 |
|
| 34 |
# Configure trainer with VL-specific settings
|
| 35 |
trainer = SFTTrainer(
|
| 36 |
+
model="Qwen/Qwen2.5-VL-3B-Instruct",
|
| 37 |
train_dataset=train_dataset,
|
| 38 |
eval_dataset=eval_dataset,
|
| 39 |
peft_config=LoraConfig(
|
|
|
|
| 47 |
args=SFTConfig(
|
| 48 |
output_dir="qwen3-vl-3b-llava-instruct",
|
| 49 |
push_to_hub=True,
|
| 50 |
+
hub_model_id="merve/qwen2.5-vl-3b-llava-instruct",
|
| 51 |
num_train_epochs=3,
|
| 52 |
per_device_train_batch_size=1,
|
| 53 |
gradient_accumulation_steps=8,
|