File size: 4,968 Bytes
e301b65 a96a027 c068180 a96a027 c068180 a96a027 b195fef 1a2171c c068180 1a2171c c068180 1a2171c c068180 1a2171c c068180 e301b65 a96a027 1a2171c a96a027 1a2171c a96a027 1a2171c a96a027 1a2171c a96a027 1a2171c c068180 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 |
---
title: ๐ค ุฅุฏุฑุงูู (Edraky) - Multilingual Educational AI Model ๐ช๐ฌ
emoji: ๐
colorFrom: indigo
colorTo: green
sdk: gradio
sdk_version: 5.38.0
app_file: app.py
pinned: false
license: apache-2.0
datasets:
- fka/awesome-chatgpt-prompts
- microsoft/rStar-Coder
- gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct
language:
- ar
- en
- he
metrics:
- accuracy
- perplexity
- wer
base_model: Qwen/Qwen2-1.5B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- multilingual
- arabic
- hebrew
- qwen
- educational
- fine-tuned
- open-source
- egyptian-curriculum
---
<style>
body {
font-family: 'Cairo', sans-serif;
background: linear-gradient(to left, #f9f9f9, #e0ecf7);
color: #222;
padding: 2em;
line-height: 1.8;
}
h1, h2, h3, h4 {
color: #003366;
border-bottom: 2px solid #eee;
padding-bottom: 0.3em;
}
code {
background-color: #f4f4f4;
padding: 0.2em 0.4em;
border-radius: 4px;
font-family: Consolas, monospace;
color: #c7254e;
}
pre {
background-color: #f0f0f0;
padding: 1em;
border-radius: 8px;
overflow-x: auto;
}
ul {
padding-left: 1.5em;
}
blockquote {
background: #f9f9f9;
border-left: 5px solid #ccc;
padding: 1em;
font-style: italic;
color: #666;
}
</style>
# ๐ค ุฅุฏุฑุงูู (Edraky) - Multilingual Educational AI Model ๐ช๐ฌ
**Edraky** is a fine-tuned multilingual model built on `Qwen2-1.5B-Instruct`, designed to provide educational support for Arabic-speaking students, especially targeting Egypt's 3rd preparatory curriculum. It supports Arabic, English, and Hebrew to ensure flexible, broad usage in multilingual environments.
## ๐ง About Edraky
Edraky is part of the **"ุฅุฏุฑุงูู"** educational initiative to democratize access to AI-powered tools for students in Egypt and the broader Arab world. By fine-tuning the powerful Qwen2 base model, Edraky delivers context-aware, curriculum-aligned, and interactive responses that help learners understand core subjects such as:
- ุงููุบุฉ ุงูุนุฑุจูุฉ (Arabic Language)
- ุงูุฏุฑุงุณุงุช ุงูุงุฌุชู
ุงุนูุฉ (Social Studies)
- ุงูุชุงุฑูุฎ ูุงูุฌุบุฑุงููุง (History and Geography)
- ุงููุบุฉ ุงูุฅูุฌููุฒูุฉ (English)
## ๐ Key Features
- ๐ค **Text Generation & Q&A**: Answer student questions in an educational and child-safe manner.
- ๐ **Curriculum Support**: Focused especially on 3rd preparatory grade in Egypt.
- ๐ **Multilingual Input**: Supports Arabic, English, and Hebrew.
- ๐ **Open-Source**: Available for research, personal, or educational use.
- ๐ **Trained on curated educational prompts** for logic, language understanding, and curriculum-based queries.
## ๐งช Training & Fine-Tuning
**Base model:** `Qwen/Qwen2-1.5B-Instruct`
**Training Data Sources:**
- fka/awesome-chatgpt-prompts
- gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct
- Additional data created from Arabic curriculum-style questions and student textbooks
**Training Methodology:**
- Supervised fine-tuning
- Prompt-optimized inputs
- Tokenized using Hugging Faceโs tokenizer compatible with Qwen2 models
## ๐ Evaluation
Model was evaluated on:
- โ๏ธ Accuracy for subject-specific answers
- โ๏ธ Perplexity for fluency and coherence
- โ๏ธ WER (Word Error Rate) for language understanding
> Evaluation still in progress for full benchmarks โ to be published soon.
## ๐งโ๐ป Example Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Edraky/Edraky")
tokenizer = AutoTokenizer.from_pretrained("Edraky/Edraky")
prompt = "ุงุดุฑุญ ุงูุซูุฑุฉ ุงูุนุฑุงุจูุฉ ุจุฅูุฌุงุฒ"
inputs = tokenizer(prompt, return_tensors="pt")
output = model.generate(**inputs, max_new_tokens=150)
print(tokenizer.decode(output[0], skip_special_tokens=True))
```
## ๐งโ๐ Intended Use
- ๐ฌ Classroom support AI assistant
- โ๏ธ Writing and summarization in Arabic
- โ Question answering for exam preparation
- ๐ Fact recall for historical, literary, and social studies content
### โ Not Intended For:
- โ Political or religious fatwa content
- โ Personal decision-making
- โ Generating offensive or misleading answers
## ๐ฑ Future Plans
- โ
Add voice input/output via Whisper integration
- โ
Online quiz companion
- โ
Add visual aids (diagrams, maps)
- โ
Full web platform integration (see [edraky.rf.gd](https://edraky.rf.gd))
## ๐ข Maintainers
**Developed by:** Edraky AI Team
๐ Website: [https://edraky.rf.gd](https://edraky.rf.gd)
๐ง Contact: edraky.edu@gmail.com
## ๐ Citation
```bibtex
@misc{edraky2025,
title={Edraky: Multilingual Educational AI Model},
author={Edraky Team},
year={2025},
howpublished={\url{https://huggingface.co/Edraky/Edraky}}
}
```
> ูุฐุง ุงูู
ุดุฑูุน ู
ู ุฃุฌู ุฏุนู
ุงูุชุนููู
ูู ู
ุตุฑ ุจุงุณุชุฎุฏุงู
ุงูุฐูุงุก ุงูุงุตุทูุงุนู. ูุฑุฌู ุฃู ูููู ู
ููุฏูุง ูุฌู
ูุน ุงูุทูุงุจ ูุงูู
ุนูู
ูู ๐ |