File size: 1,318 Bytes
99c4a2d
 
 
df2cc83
 
 
 
 
 
 
 
 
 
 
9e042ba
df2cc83
99c4a2d
 
b88304a
 
 
 
57db17f
b88304a
 
 
99c4a2d
 
 
 
b88304a
99c4a2d
b88304a
99c4a2d
b88304a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
---
language:
- en
- zh
license: apache-2.0
base_model: Qwen/Qwen3.5-9B
tags:
- unsloth
- qwen
- qwen3.5
- reasoning
- chain-of-thought
- lora
- luau
- roblox
pipeline_tag: image-text-to-text
---

# Mini-Coder

Mini-Coder is build on top of Qwen3.5-9B model with Continual Pretraining (CPT), we feed ~500k high-quality curated luau samples to improves the luau coding tasks capability.

We also inject over 14k samples from open-source of claude 4.6 distillations with a fews additional samples for Supervised-Finetuning (SFT) to improves the model reasoning, We also see the average consumed tokens has drastically reduced.

It's fine-tuned efficiently using LoRA (16-bit) and rsLoRA with Rank (r) set to 64 and Alpha (α) set to 128, ensuring strong adaptation and retention of new complex logic, it were trained specifically to handle up to 32,768 (32k) tokens of maximum output (recommended).

# Uploaded finetuned  model

- **Developed by:** khtsly
- **License:** apache-2.0
- **Finetuned from model :** khtsly/Coder-9B

This qwen3_5_text model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)