Nwna/olmo3-190m-zh-v2-sft

SFT(有监督微调)版本:基于 Nwna/olmo3-190m-zh-v2-continue, 使用对话格式数据进行微调,学习指令遵循能力。

数据来源

训练配置

  • LR:5e-05(低 LR 避免灾难性遗忘)
  • Warmup:5.0%
  • Max Seq Length:2048

用法

from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Nwna/olmo3-190m-zh-v2-sft")
tok = AutoTokenizer.from_pretrained("Nwna/olmo3-190m-zh-v2-sft")
Downloads last month
18
Safetensors
Model size
0.2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Nwna/olmo3-190m-zh-v2-sft

Finetuned
(1)
this model

Space using Nwna/olmo3-190m-zh-v2-sft 1