| --- |
| base_model: |
| - TheDrummer/Llama-3SOME-8B-v2 |
| - cgato/L3-TheSpice-8b-v0.8.3 |
| - Sao10K/L3-8B-Stheno-v3.2 |
| - SicariusSicariiStuff/Wingless_Imp_8B |
| - deepseek-ai/DeepSeek-R1-Distill-Llama-8B |
| - NeverSleep/Lumimaid-v0.2-8B |
| library_name: transformers |
| tags: |
| - mergekit |
| - merge |
| license: cc-by-nc-4.0 |
| --- |
| # What is this? |
|
|
| A Llama3 model with Deepseek R1 Distill merge. Maybe it's not suit for RP? |
|
|
|  |
|
|
| Overall, this merge model is the best and smartest RP, ERP model. But the IFEval score is lower than other model, so I think it's wont follow well your instructions? I didn't test yet, will have a test later |
|
|
| <details> |
| <summary>## Merge Detail</summary> |
| <p> |
| ### Models Merged |
| |
| The following models were included in the merge: |
| * [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2) |
| * [cgato/L3-TheSpice-8b-v0.8.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.8.3) |
| * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2) |
| * [SicariusSicariiStuff/Wingless_Imp_8B](https://huggingface.co/SicariusSicariiStuff/Wingless_Imp_8B) |
| * [deepseek-ai/DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B) |
|
|
| ### Configuration |
|
|
| The following YAML configuration was used to produce this model: |
|
|
| ```yaml |
| base_model: NeverSleep/Lumimaid-v0.2-8B |
| merge_method: model_stock |
| dtype: bfloat16 |
| models: |
| - model: cgato/L3-TheSpice-8b-v0.8.3 |
| - model: Sao10K/L3-8B-Stheno-v3.2 |
| - model: TheDrummer/Llama-3SOME-8B-v2 |
| - model: SicariusSicariiStuff/Wingless_Imp_8B |
| - model: deepseek-ai/DeepSeek-R1-Distill-Llama-8B |
| ``` |
|
|
| </p> |
| </details> |