Update README.md
Browse files
README.md
CHANGED
|
@@ -7,6 +7,16 @@ base_model:
|
|
| 7 |
- Qwen/Qwen3-30B-A3B-Base
|
| 8 |
---
|
| 9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
# Qwen3-30B-A3B
|
| 11 |
<a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;">
|
| 12 |
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
|
|
|
|
| 7 |
- Qwen/Qwen3-30B-A3B-Base
|
| 8 |
---
|
| 9 |
|
| 10 |
+
# Sharded weights checkpoints
|
| 11 |
+
|
| 12 |
+
This is derived directly from [`save_sharded_state.py`](https://github.com/vllm-project/vllm/blob/main/examples/offline_inference/save_sharded_state.py) to be used with vLLM with `-tp=2`:
|
| 13 |
+
|
| 14 |
+
```bash
|
| 15 |
+
vllm serve aarnphm/qwen3-30b-a3b-sharded-tp2 \
|
| 16 |
+
-tp=2 \
|
| 17 |
+
--load-format sharded_state
|
| 18 |
+
```
|
| 19 |
+
|
| 20 |
# Qwen3-30B-A3B
|
| 21 |
<a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;">
|
| 22 |
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
|