Update README.md
Browse files
README.md
CHANGED
|
@@ -152,7 +152,7 @@ model-index:
|
|
| 152 |
---
|
| 153 |
# 🔬 Einstein-v4-7B
|
| 154 |
|
| 155 |
-
This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on diverse datasets.
|
| 156 |
|
| 157 |
This model is finetuned using `7xRTX3090` + `1xRTXA6000` using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
|
| 158 |
|
|
@@ -293,6 +293,30 @@ gen_input = tokenizer.apply_chat_template(message, return_tensors="pt")
|
|
| 293 |
model.generate(**gen_input)
|
| 294 |
```
|
| 295 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 296 |
# 🤝 Acknowledgments
|
| 297 |
|
| 298 |
Thanks to [sablo.ai](https://sablo.ai) for sponsoring this model.
|
|
@@ -308,16 +332,3 @@ Thanks to all open source AI community.
|
|
| 308 |
If you would like to support me:
|
| 309 |
|
| 310 |
[☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)
|
| 311 |
-
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
| 312 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v4-7B)
|
| 313 |
-
|
| 314 |
-
| Metric |Value|
|
| 315 |
-
|---------------------------------|----:|
|
| 316 |
-
|Avg. |66.62|
|
| 317 |
-
|AI2 Reasoning Challenge (25-Shot)|64.68|
|
| 318 |
-
|HellaSwag (10-Shot) |83.75|
|
| 319 |
-
|MMLU (5-Shot) |62.31|
|
| 320 |
-
|TruthfulQA (0-shot) |55.15|
|
| 321 |
-
|Winogrande (5-shot) |76.24|
|
| 322 |
-
|GSM8k (5-shot) |57.62|
|
| 323 |
-
|
|
|
|
| 152 |
---
|
| 153 |
# 🔬 Einstein-v4-7B
|
| 154 |
|
| 155 |
+
This model is a full fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on diverse datasets.
|
| 156 |
|
| 157 |
This model is finetuned using `7xRTX3090` + `1xRTXA6000` using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
|
| 158 |
|
|
|
|
| 293 |
model.generate(**gen_input)
|
| 294 |
```
|
| 295 |
|
| 296 |
+
# 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
| 297 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v4-7B)
|
| 298 |
+
|
| 299 |
+
| Metric |Value|
|
| 300 |
+
|---------------------------------|----:|
|
| 301 |
+
|Avg. |66.62|
|
| 302 |
+
|AI2 Reasoning Challenge (25-Shot)|64.68|
|
| 303 |
+
|HellaSwag (10-Shot) |83.75|
|
| 304 |
+
|MMLU (5-Shot) |62.31|
|
| 305 |
+
|TruthfulQA (0-shot) |55.15|
|
| 306 |
+
|Winogrande (5-shot) |76.24|
|
| 307 |
+
|GSM8k (5-shot) |57.62|
|
| 308 |
+
|
| 309 |
+
# 🤖 Additional information about training
|
| 310 |
+
|
| 311 |
+
This model is full fine-tuned for 1.5 epoch.
|
| 312 |
+
|
| 313 |
+
Total number of steps was 1562.
|
| 314 |
+
|
| 315 |
+
<details><summary>Loss graph</summary>
|
| 316 |
+
|
| 317 |
+

|
| 318 |
+
</details><br>
|
| 319 |
+
|
| 320 |
# 🤝 Acknowledgments
|
| 321 |
|
| 322 |
Thanks to [sablo.ai](https://sablo.ai) for sponsoring this model.
|
|
|
|
| 332 |
If you would like to support me:
|
| 333 |
|
| 334 |
[☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|