Update README.md
Browse files
README.md
CHANGED
|
@@ -13,6 +13,8 @@ pipeline_tag: fill-mask
|
|
| 13 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
|
| 15 |
|
|
|
|
|
|
|
| 16 |
|
| 17 |
## Evaluation
|
| 18 |
|
|
@@ -32,6 +34,16 @@ Read the [paper](https://aclanthology.org/2024.unlp-1.14/) for more detailed tas
|
|
| 32 |
| [liberta-large](https://huggingface.co/Goader/liberta-large) | 91.27 (1.22) | 92.50 (0.07) | 98.62 (0.08) | 95.44 (0.04) |
|
| 33 |
| [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2) | __91.73 (1.81)__ | __93.22 (0.14)__ | __98.79 (0.06)__ | 95.67 (0.12) |
|
| 34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
|
| 36 |
|
| 37 |
## How to Get Started with the Model
|
|
|
|
| 13 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
|
| 15 |
|
| 16 |
+
All the code is available in the [Goader/ukr-lm](https://github.com/Goader/ukr-lm) repository.
|
| 17 |
+
|
| 18 |
|
| 19 |
## Evaluation
|
| 20 |
|
|
|
|
| 34 |
| [liberta-large](https://huggingface.co/Goader/liberta-large) | 91.27 (1.22) | 92.50 (0.07) | 98.62 (0.08) | 95.44 (0.04) |
|
| 35 |
| [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2) | __91.73 (1.81)__ | __93.22 (0.14)__ | __98.79 (0.06)__ | 95.67 (0.12) |
|
| 36 |
|
| 37 |
+
## Fine-Tuning Hyperparameters
|
| 38 |
+
|
| 39 |
+
| Hyperparameter | Value |
|
| 40 |
+
|:---------------|:-----:|
|
| 41 |
+
| Peak Learning Rate | 3e-5 |
|
| 42 |
+
| Warm-up Ratio | 0.05 |
|
| 43 |
+
| Learning Rate Decay | Linear |
|
| 44 |
+
| Batch Size | 16 |
|
| 45 |
+
| Epochs | 10 |
|
| 46 |
+
| Weight Decay | 0.05 |
|
| 47 |
|
| 48 |
|
| 49 |
## How to Get Started with the Model
|