Skip to content

Latest commit

 

History

History
152 lines (133 loc) · 4.54 KB

File metadata and controls

152 lines (133 loc) · 4.54 KB
license model-index
apache-2.0
name results
Dolphin2.1-OpenOrca-7B
task dataset metrics source
type name
text-generation
Text Generation
name type config split args
AI2 Reasoning Challenge (25-Shot)
ai2_arc
ARC-Challenge
test
num_few_shot
25
type value name
acc_norm
63.91
normalized accuracy
task dataset metrics source
type name
text-generation
Text Generation
name type split args
HellaSwag (10-Shot)
hellaswag
validation
num_few_shot
10
type value name
acc_norm
84.26
normalized accuracy
task dataset metrics source
type name
text-generation
Text Generation
name type config split args
MMLU (5-Shot)
cais/mmlu
all
test
num_few_shot
5
type value name
acc
62.66
accuracy
task dataset metrics source
type name
text-generation
Text Generation
name type config split args
TruthfulQA (0-shot)
truthful_qa
multiple_choice
validation
num_few_shot
0
type value
mc2
53.84
task dataset metrics source
type name
text-generation
Text Generation
name type config split args
Winogrande (5-shot)
winogrande
winogrande_xl
validation
num_few_shot
5
type value name
acc
78.22
accuracy
task dataset metrics source
type name
text-generation
Text Generation
name type config split args
GSM8k (5-shot)
gsm8k
main
test
num_few_shot
5
type value name
acc
19.94
accuracy

Buy Me A Coffee

Merge of ehartford/dolphin-2.1-mistral-7b and Open-Orca/Mistral-7B-OpenOrca using ties merge.

Weights

Density

Quantizationed versions

Quantizationed versions of this model is available thanks to TheBloke.

GPTQ
GGUF
AWQ

Detailed results can be found here

Metric Value
Avg. 60.47
AI2 Reasoning Challenge (25-Shot) 63.91
HellaSwag (10-Shot) 84.26
MMLU (5-Shot) 62.66
TruthfulQA (0-shot) 53.84
Winogrande (5-shot) 78.22
GSM8k (5-shot) 19.94