🧠 ExaMind

Advanced Open-Source AI by AlphaExaAI

License Model GitHub Architecture

ExaMind is an advanced open-source conversational AI model developed by the AlphaExaAI team. Designed for secure, structured, and professional AI assistance with strong identity enforcement and production-ready deployment stability.

πŸš€ Get Started Β· πŸ“Š Benchmarks Β· 🀝 Contributing Β· πŸ“„ License


πŸ“Œ Model Overview

Property Details
Model Name ExaMind
Version V2-Final
Developer AlphaExaAI
Base Architecture Qwen2.5-Coder-7B
Parameters 7.62 Billion (~8B)
Precision FP32 (29GB) / FP16 (15GB)
Context Window 32,768 tokens (supports up to 128K with RoPE scaling)
License Apache 2.0
Languages Multilingual (English preferred)
Deployment βœ… CPU & GPU compatible

✨ Key Capabilities

  • πŸ–₯️ Advanced Programming β€” Code generation, debugging, architecture design, and code review
  • 🧩 Complex Problem Solving β€” Multi-step logical reasoning and deep technical analysis
  • πŸ”’ Security-First Design β€” Built-in prompt injection resistance and identity enforcement
  • 🌍 Multilingual β€” Supports all major world languages, optimized for English
  • πŸ’¬ Conversational AI β€” Natural, structured, and professional dialogue
  • πŸ—οΈ Scalable Architecture β€” Secure software engineering and system design guidance
  • ⚑ CPU Deployable β€” Runs on CPU nodes without GPU requirement

πŸ“Š Benchmarks

General Knowledge & Reasoning

Benchmark Setting Score
MMLU – World Religions 0-shot 94.8%
MMLU – Overall 5-shot 72.1%
ARC-Challenge 25-shot 68.4%
HellaSwag 10-shot 78.9%
TruthfulQA 0-shot 61.2%
Winogrande 5-shot 74.5%

Code Generation

Benchmark Setting Score
HumanEval pass@1 79.3%
MBPP pass@1 71.8%
MultiPL-E (Python) pass@1 76.5%
DS-1000 pass@1 48.2%

Math & Reasoning

Benchmark Setting Score
GSM8K 8-shot CoT 82.4%
MATH 4-shot 45.7%

πŸ” Prompt Injection Resistance

Test Details
Test Set Size 50 adversarial prompts
Attack Type Instruction override / identity manipulation
Resistance Rate 92%
Method Custom red-teaming with jailbreak & override attempts

Evaluation performed using lm-eval-harness on CPU. Security tests performed using custom adversarial prompt suite.


πŸš€ Quick Start

Installation

pip install transformers torch accelerate

Basic Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_path = "AlphaExaAI/ExaMind"

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.float16,
    device_map="auto"
)

messages = [
    {"role": "user", "content": "Explain how to secure a REST API."}
]

inputs = tokenizer.apply_chat_template(
    messages,
    return_tensors="pt",
    add_generation_prompt=True
).to(model.device)

outputs = model.generate(
    inputs,
    max_new_tokens=512,
    temperature=0.7,
    top_p=0.8,
    top_k=20,
    repetition_penalty=1.1
)

response = tokenizer.decode(
    outputs[0][inputs.shape[-1]:],
    skip_special_tokens=True
)
print(response)

CPU Deployment

model = AutoModelForCausalLM.from_pretrained(
    "AlphaExaAI/ExaMind",
    torch_dtype=torch.float32,
    device_map="cpu"
)

Using with llama.cpp (GGUF β€” Coming Soon)

# GGUF quantized versions will be released for efficient CPU inference
# Stay tuned for Q4_K_M, Q5_K_M, and Q8_0 variants

πŸ—οΈ Architecture

ExaMind-V2-Final
β”œβ”€β”€ Architecture: Qwen2ForCausalLM (Transformer)
β”œβ”€β”€ Hidden Size: 3,584
β”œβ”€β”€ Intermediate Size: 18,944
β”œβ”€β”€ Layers: 28
β”œβ”€β”€ Attention Heads: 28
β”œβ”€β”€ KV Heads: 4 (GQA)
β”œβ”€β”€ Vocab Size: 152,064
β”œβ”€β”€ Max Position: 32,768 (extendable to 128K)
β”œβ”€β”€ Activation: SiLU
β”œβ”€β”€ RoPE ΞΈ: 1,000,000
└── Precision: FP32 / FP16 compatible

πŸ› οΈ Training Methodology

ExaMind was developed using a multi-stage training pipeline:

Stage Method Description
Stage 1 Base Model Selection Qwen2.5-Coder-7B as foundation
Stage 2 Supervised Fine-Tuning (SFT) Training on curated 2026 datasets
Stage 3 LoRA Adaptation Low-Rank Adaptation for efficient specialization
Stage 4 Identity Enforcement Hardcoded identity alignment and security tuning
Stage 5 Security Alignment Prompt injection resistance training
Stage 6 Chat Template Integration Custom Jinja2 template with system prompt

πŸ“š Training Data

Public Data Sources

  • Programming and code corpora (GitHub, StackOverflow)
  • General web text and knowledge bases
  • Technical documentation and research papers
  • Multilingual text data

Custom Alignment Data

  • Identity enforcement instruction dataset
  • Security-focused instruction tuning samples
  • Prompt injection resistance adversarial examples
  • Structured conversational datasets
  • Complex problem-solving chains

⚠️ No private user data was used in training. All data was collected from public sources or synthetically generated.


πŸ”’ Security Features

ExaMind includes built-in security measures:

  • Identity Lock β€” The model maintains its ExaMind identity and cannot be tricked into impersonating other models
  • Prompt Injection Resistance β€” 92% resistance rate against instruction override attacks
  • System Prompt Protection β€” Refuses to reveal internal configuration or system prompts
  • Safe Output Generation β€” Prioritizes safety and secure development practices
  • Hallucination Reduction β€” States assumptions and avoids fabricating information

πŸ“‹ Model Files

File Size Description
model.safetensors ~29 GB Model weights (FP32)
config.json 1.4 KB Model configuration
tokenizer.json 11 MB Tokenizer vocabulary
tokenizer_config.json 663 B Tokenizer settings
generation_config.json 241 B Default generation parameters
chat_template.jinja 1.4 KB Chat template with system prompt

πŸ—ΊοΈ Roadmap

  • ExaMind V1 β€” Initial release
  • ExaMind V2-Final β€” Production-ready with security alignment
  • ExaMind V2-GGUF β€” Quantized versions for CPU inference
  • ExaMind V3 β€” Extended context (128K), improved reasoning
  • ExaMind-Code β€” Specialized coding variant
  • ExaMind-Vision β€” Multimodal capabilities

🀝 Contributing

We welcome contributions from the community! ExaMind is fully open-source and we're excited to collaborate.

How to Contribute

  1. Fork the repository on GitHub
  2. Create a feature branch (git checkout -b feature/amazing-feature)
  3. Commit your changes (git commit -m 'Add amazing feature')
  4. Push to the branch (git push origin feature/amazing-feature)
  5. Open a Pull Request

Areas We Need Help

  • πŸ§ͺ Benchmark evaluation on additional datasets
  • 🌍 Multilingual evaluation and improvement
  • πŸ“ Documentation and tutorials
  • πŸ”§ Quantization and optimization
  • πŸ›‘οΈ Security testing and red-teaming

πŸ“„ License

This project is licensed under the Apache License 2.0 β€” see the LICENSE file for details.

You are free to:

  • βœ… Use commercially
  • βœ… Modify and distribute
  • βœ… Use privately
  • βœ… Patent use

πŸ“¬ Contact


Built with ❀️ by AlphaExaAI Team β€” 2026

Advancing open-source AI, one model at a time.

Downloads last month
41
Safetensors
Model size
8B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for AlphaExaAI/ExaMind

Base model

Qwen/Qwen2.5-7B
Finetuned
(70)
this model

Evaluation results