finetune-e6e1a1f3

Fine-tuned version of Qwen/Qwen3-0.6B using DPO.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("Canfield/finetune-e6e1a1f3")
tokenizer = AutoTokenizer.from_pretrained("Canfield/finetune-e6e1a1f3")

# Generate text
messages = [
    {"role": "user", "content": "Hello, how are you?"}
]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
outputs = model.generate(inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))

Training Details

  • Method: DPO
  • Base Model: Qwen/Qwen3-0.6B
  • LoRA merged and quantization removed for inference compatibility
Downloads last month
1
Safetensors
Model size
0.4B params
Tensor type
F32
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Canfield/finetune-e6e1a1f3

Finetuned
Qwen/Qwen3-0.6B
Finetuned
(549)
this model