end_to_end_test |
1 |
1 |
1149.92 |
1177.75 |
0.73 |
5876.0 |
neversleep-noromaid-v0-_8068_v87 |
47B |
NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3 |
neversleep-noromaid-v0-_8068_v87 |
12582 |
24076 |
0.46 |
0.50 |
basic |
ChaiML/reward_models_100_170000000_cp_498032 |
4 |
512 |
64 |
MixtralForCausalLM |
NeverSleep/Noromaid-v0.1 |
2024-07-01 |
0.37 |
1.46 |
2 |
1 |
2 |
1 |
end_to_end_test |
2 |
1 |
1177.75 |
1177.75 |
0.71 |
5451.0 |
neversleep-noromaid-v0_8068_v102 |
47B |
NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3 |
neversleep-noromaid-v0_8068_v102 |
11494 |
24076 |
0.50 |
0.50 |
basic |
ChaiML/gpt2_xl_pairwise_89m_step_347634 |
4 |
512 |
64 |
MixtralForCausalLM |
NeverSleep/Noromaid-v0.1 |
2024-07-01 |
1.05 |
-0.27 |
1 |
1 |
1 |
2 |
end_to_end_test |
3 |
3 |
1130.55 |
1130.55 |
0.71 |
11994.0 |
zeuslabs-l3-aethora-15b-v2_v8 |
15B |
ZeusLabs/L3-Aethora-15B-V2 |
zeuslabs-l3-aethora-15b-v2_v8 |
25299 |
25299 |
0.43 |
0.43 |
basic |
ChaiML/reward_models_100_170000000_cp_498032 |
4 |
512 |
64 |
LlamaForCausalLM |
ZeusLabs/L3-Aethora-15B- |
2024-06-28 |
-0.10 |
-0.40 |
3 |
3 |
3 |
3 |
end_to_end_test |
4 |
4 |
1080.81 |
1080.81 |
0.71 |
13338.0 |
nousresearch-meta-llama-3-8b_v19 |
8B |
NousResearch/Meta-Llama-3-8B |
nousresearch-meta-llama-3-8b_v19 |
28422 |
28422 |
0.34 |
0.34 |
basic |
ChaiML/reward_models_100_170000000_cp_498032 |
4 |
512 |
64 |
LlamaForCausalLM |
NousResearch/Meta-Llama- |
2024-06-18 |
-1.32 |
-0.79 |
4 |
4 |
4 |
4 |