File size: 4,968 Bytes
e301b65
a96a027
c068180
a96a027
 
 
c068180
a96a027
 
b195fef
1a2171c
c068180
 
 
1a2171c
c068180
 
 
1a2171c
c068180
 
 
1a2171c
 
 
 
c068180
 
 
 
 
 
 
 
e301b65
a96a027
1a2171c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a96a027
1a2171c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a96a027
1a2171c
 
 
 
 
 
a96a027
1a2171c
 
 
 
 
 
 
 
 
 
 
 
a96a027
1a2171c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c068180
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
---
title: ๐Ÿค– ุฅุฏุฑุงูƒูŠ (Edraky) - Multilingual Educational AI Model ๐Ÿ‡ช๐Ÿ‡ฌ
emoji: ๐Ÿš€
colorFrom: indigo
colorTo: green
sdk: gradio
sdk_version: 5.38.0
app_file: app.py
pinned: false
license: apache-2.0
datasets:
- fka/awesome-chatgpt-prompts
- microsoft/rStar-Coder
- gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct
language:
- ar
- en
- he
metrics:
- accuracy
- perplexity
- wer
base_model: Qwen/Qwen2-1.5B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- multilingual
- arabic
- hebrew
- qwen
- educational
- fine-tuned
- open-source
- egyptian-curriculum
---

<style>
body {
  font-family: 'Cairo', sans-serif;
  background: linear-gradient(to left, #f9f9f9, #e0ecf7);
  color: #222;
  padding: 2em;
  line-height: 1.8;
}

h1, h2, h3, h4 {
  color: #003366;
  border-bottom: 2px solid #eee;
  padding-bottom: 0.3em;
}

code {
  background-color: #f4f4f4;
  padding: 0.2em 0.4em;
  border-radius: 4px;
  font-family: Consolas, monospace;
  color: #c7254e;
}

pre {
  background-color: #f0f0f0;
  padding: 1em;
  border-radius: 8px;
  overflow-x: auto;
}

ul {
  padding-left: 1.5em;
}

blockquote {
  background: #f9f9f9;
  border-left: 5px solid #ccc;
  padding: 1em;
  font-style: italic;
  color: #666;
}
</style>

# ๐Ÿค– ุฅุฏุฑุงูƒูŠ (Edraky) - Multilingual Educational AI Model ๐Ÿ‡ช๐Ÿ‡ฌ

**Edraky** is a fine-tuned multilingual model built on `Qwen2-1.5B-Instruct`, designed to provide educational support for Arabic-speaking students, especially targeting Egypt's 3rd preparatory curriculum. It supports Arabic, English, and Hebrew to ensure flexible, broad usage in multilingual environments.

## ๐Ÿง  About Edraky

Edraky is part of the **"ุฅุฏุฑุงูƒูŠ"** educational initiative to democratize access to AI-powered tools for students in Egypt and the broader Arab world. By fine-tuning the powerful Qwen2 base model, Edraky delivers context-aware, curriculum-aligned, and interactive responses that help learners understand core subjects such as:

- ุงู„ู„ุบุฉ ุงู„ุนุฑุจูŠุฉ (Arabic Language)
- ุงู„ุฏุฑุงุณุงุช ุงู„ุงุฌุชู…ุงุนูŠุฉ (Social Studies)
- ุงู„ุชุงุฑูŠุฎ ูˆุงู„ุฌุบุฑุงููŠุง (History and Geography)
- ุงู„ู„ุบุฉ ุงู„ุฅู†ุฌู„ูŠุฒูŠุฉ (English)

## ๐Ÿš€ Key Features

- ๐Ÿค– **Text Generation & Q&A**: Answer student questions in an educational and child-safe manner.
- ๐Ÿ“– **Curriculum Support**: Focused especially on 3rd preparatory grade in Egypt.
- ๐ŸŒ **Multilingual Input**: Supports Arabic, English, and Hebrew.
- ๐Ÿ”€ **Open-Source**: Available for research, personal, or educational use.
- ๐Ÿ“š **Trained on curated educational prompts** for logic, language understanding, and curriculum-based queries.

## ๐Ÿงช Training & Fine-Tuning

**Base model:** `Qwen/Qwen2-1.5B-Instruct`

**Training Data Sources:**
- fka/awesome-chatgpt-prompts
- gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct
- Additional data created from Arabic curriculum-style questions and student textbooks

**Training Methodology:**
- Supervised fine-tuning
- Prompt-optimized inputs
- Tokenized using Hugging Faceโ€™s tokenizer compatible with Qwen2 models

## ๐Ÿ” Evaluation

Model was evaluated on:
- โœ”๏ธ Accuracy for subject-specific answers
- โœ”๏ธ Perplexity for fluency and coherence
- โœ”๏ธ WER (Word Error Rate) for language understanding
> Evaluation still in progress for full benchmarks โ€” to be published soon.

## ๐Ÿง‘โ€๐Ÿ’ป Example Usage

```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Edraky/Edraky")
tokenizer = AutoTokenizer.from_pretrained("Edraky/Edraky")
prompt = "ุงุดุฑุญ ุงู„ุซูˆุฑุฉ ุงู„ุนุฑุงุจูŠุฉ ุจุฅูŠุฌุงุฒ"
inputs = tokenizer(prompt, return_tensors="pt")
output = model.generate(**inputs, max_new_tokens=150)
print(tokenizer.decode(output[0], skip_special_tokens=True))
```

## ๐Ÿง‘โ€๐Ÿ““ Intended Use

- ๐Ÿ’ฌ Classroom support AI assistant
- โœ๏ธ Writing and summarization in Arabic
- โ“ Question answering for exam preparation
- ๐Ÿ” Fact recall for historical, literary, and social studies content

### โŒ Not Intended For:

- โŒ Political or religious fatwa content
- โŒ Personal decision-making
- โŒ Generating offensive or misleading answers

## ๐ŸŒฑ Future Plans

- โœ… Add voice input/output via Whisper integration
- โœ… Online quiz companion
- โœ… Add visual aids (diagrams, maps)
- โœ… Full web platform integration (see [edraky.rf.gd](https://edraky.rf.gd))

## ๐Ÿ“ข Maintainers

**Developed by:** Edraky AI Team  
๐ŸŒ Website: [https://edraky.rf.gd](https://edraky.rf.gd)  
๐Ÿ“ง Contact: edraky.edu@gmail.com

## ๐Ÿ“œ Citation

```bibtex
@misc{edraky2025,
  title={Edraky: Multilingual Educational AI Model},
  author={Edraky Team},
  year={2025},
  howpublished={\url{https://huggingface.co/Edraky/Edraky}}
}
```

> ู‡ุฐุง ุงู„ู…ุดุฑูˆุน ู…ู† ุฃุฌู„ ุฏุนู… ุงู„ุชุนู„ูŠู… ููŠ ู…ุตุฑ ุจุงุณุชุฎุฏุงู… ุงู„ุฐูƒุงุก ุงู„ุงุตุทู†ุงุนูŠ. ู†ุฑุฌูˆ ุฃู† ูŠูƒูˆู† ู…ููŠุฏู‹ุง ู„ุฌู…ูŠุน ุงู„ุทู„ุงุจ ูˆุงู„ู…ุนู„ู…ูŠู† ๐ŸŒŸ