developer_uid: rirv938
submission_id: function_hohub_2025-01-29
model_name: retune_with_base
model_group:
status: torndown
timestamp: 2025-01-29T18:04:25+00:00
num_battles: 7140
num_wins: 3743
celo_rating: 1286.39
family_friendly_score: 0.5524
family_friendly_standard_error: 0.007032129691636808
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2025-01-29
win_ratio: 0.5242296918767507
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.1597883701324463s
Received healthy response to inference request in 5.965150833129883s
Received healthy response to inference request in 4.156012773513794s
Received healthy response to inference request in 4.234748363494873s
Received healthy response to inference request in 3.685058116912842s
5 requests
0 failed requests
5th percentile: 2.4648423194885254
10th percentile: 2.7698962688446045
20th percentile: 3.3800041675567627
30th percentile: 3.7792490482330323
40th percentile: 3.967630910873413
50th percentile: 4.156012773513794
60th percentile: 4.187507009506225
70th percentile: 4.219001245498657
80th percentile: 4.580828857421875
90th percentile: 5.272989845275879
95th percentile: 5.619070339202881
99th percentile: 5.895934734344483
mean time: 4.040151691436767
%s, retrying in %s seconds...
Received healthy response to inference request in 3.011744260787964s
Received healthy response to inference request in 3.041499376296997s
Received healthy response to inference request in 3.8624298572540283s
Received healthy response to inference request in 2.3056485652923584s
Received healthy response to inference request in 3.336170196533203s
5 requests
0 failed requests
5th percentile: 2.4468677043914795
10th percentile: 2.5880868434906006
20th percentile: 2.8705251216888428
30th percentile: 3.0176952838897706
40th percentile: 3.0295973300933836
50th percentile: 3.041499376296997
60th percentile: 3.1593677043914794
70th percentile: 3.2772360324859617
80th percentile: 3.441422128677368
90th percentile: 3.6519259929656984
95th percentile: 3.7571779251098634
99th percentile: 3.841379470825195
mean time: 3.1114984512329102
Pipeline stage StressChecker completed in 38.36s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.63s
Shutdown handler de-registered
function_hohub_2025-01-29 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3850.03s
Shutdown handler de-registered
function_hohub_2025-01-29 status is now inactive due to auto deactivation removed underperforming models
function_hohub_2025-01-29 status is now torndown due to DeploymentManager action
ChatRequest
Generation Params
Prompt Formatter
Chat History
ChatMessage 1