license | language | |
---|---|---|
llama2 |
|
An experimental merging of Several Models using two various methods, Ties-Merge and BlockMerge_Gradient
I plan for this to be the base of my Model with my own [Stheno: ERP-Based LORA] merged in, some time in the future.
Stheno:
Gradient Merge of Stheno-P1 & Stheno-P2.
SISTER MODEL HERE: Stheno-Inverted-L2-13B
Quants courtesy of TheBloke!
GPTQ
GGUF
GGML
Test Checklist:
Censorship - Fairly Uncensored
Writing - Good Prose, Fairly Descriptive
NSFW - Yes
IQ Level - Pretty Smart
Formatting - Proper Formatting with Examples
Stheno-P1 [Ties-Merge]
-----elinas/chronos-13b-v2
-----jondurbin/airoboros-l2-13b-2.1
-----NousResearch/Nous-Hermes-Llama2-13b+nRuaif/Kimiko-v2 LORA
Stheno-P2 [Ties-Merge]
-----CalderaAI/13B-Legerdemain-L2+lemonilia/limarp-llama2-v2 LORA
-----ehartford/WizardLM-1.0-Uncensored-Llama2-13b
-----Henk717/spring-dragon
Most formats could work, but my tests have all been done in Alpaca format and it works well.
### Instruction:
Your instruction or question here.
For roleplay purposes, I suggest the following - Write <CHAR NAME>'s next reply in a chat between <YOUR NAME> and <CHAR NAME>. Write a single reply only.
### Response:
Below is the Illustration for the Final Merge:
Once Again, thanks to Chargoddard for his amazing and simple ties-merge script, and Gryphe for their great BlockMerge_Gradient script. Thanks to the original model creators too!
support me here :)
Art by wada_kazu / わだかず (pixiv page private?)
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 53.48 |
ARC (25-shot) | 61.01 |
HellaSwag (10-shot) | 83.95 |
MMLU (5-shot) | 56.33 |
TruthfulQA (0-shot) | 50.18 |
Winogrande (5-shot) | 75.14 |
GSM8K (5-shot) | 11.98 |
DROP (3-shot) | 35.76 |