prithivMLmods commited on
Commit
a9d75a5
·
verified ·
1 Parent(s): 6641cd1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -7,13 +7,13 @@ library_name: transformers
7
  tags:
8
  - mergekit
9
  - merge
10
-
11
  ---
 
 
12
  # merge
13
 
14
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
 
16
- ## Merge Details
17
  ### Merge Method
18
 
19
  This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [prithivMLmods/Megatron-Opus-14B-Exp](https://huggingface.co/prithivMLmods/Megatron-Opus-14B-Exp) as a base.
 
7
  tags:
8
  - mergekit
9
  - merge
 
10
  ---
11
+ # **Megatron-Opus-14B-Stock**
12
+ [ Megatron+Primal+Elite2 ] is based on the Qwen 2.5 14B modality architecture, designed to enhance the reasoning capabilities of 14B-parameter models. It has been fine-tuned on a Synthetic dataset entries based on one half of Qwen’s QWQ and DeepSeek R1, further optimizing its chain-of-thought (CoT) reasoning and logical problem-solving abilities. The model demonstrates significant improvements in context understanding, structured data processing, and long-context comprehension, making it ideal for complex reasoning tasks, instruction-following, and text generation.
13
  # merge
14
 
15
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
 
 
17
  ### Merge Method
18
 
19
  This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [prithivMLmods/Megatron-Opus-14B-Exp](https://huggingface.co/prithivMLmods/Megatron-Opus-14B-Exp) as a base.