finetune-e6e1a1f3
Fine-tuned version of Qwen/Qwen3-0.6B using DPO.
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Canfield/finetune-e6e1a1f3")
tokenizer = AutoTokenizer.from_pretrained("Canfield/finetune-e6e1a1f3")
# Generate text
messages = [
{"role": "user", "content": "Hello, how are you?"}
]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
outputs = model.generate(inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))
Training Details
- Method: DPO
- Base Model: Qwen/Qwen3-0.6B
- LoRA merged and quantization removed for inference compatibility
- Downloads last month
- 1