developer_uid: chai_backend_admin
submission_id: function_turam_2025-05-27
model_name: function_turam_2025-05-27
model_group:
status: torndown
timestamp: 2025-05-27T22:19:00+00:00
num_battles: 5364
num_wins: 2735
celo_rating: 1291.61
family_friendly_score: 0.5186
family_friendly_standard_error: 0.007066173504804421
submission_type: function
display_name: function_turam_2025-05-27
is_internal_developer: True
ranking_group: single
us_pacific_date: 2025-05-27
win_ratio: 0.5098806860551827
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': '### Instruction:\n{memory}\n', 'prompt_template': '### Input:\n{prompt}\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '### Response:\n{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 3.4299304485321045s
Received healthy response to inference request in 5.8285229206085205s
Received healthy response to inference request in 3.4358108043670654s
Received healthy response to inference request in 2.900285482406616s
Received healthy response to inference request in 4.543742656707764s
5 requests
0 failed requests
5th percentile: 3.006214475631714
10th percentile: 3.1121434688568117
20th percentile: 3.3240014553070067
30th percentile: 3.431106519699097
40th percentile: 3.433458662033081
50th percentile: 3.4358108043670654
60th percentile: 3.8789835453033445
70th percentile: 4.322156286239624
80th percentile: 4.800698709487915
90th percentile: 5.314610815048217
95th percentile: 5.571566867828369
99th percentile: 5.77713171005249
mean time: 4.027658462524414
%s, retrying in %s seconds...
Received healthy response to inference request in 2.70682954788208s
Received healthy response to inference request in 2.8016202449798584s
Received healthy response to inference request in 2.1643412113189697s
Received healthy response to inference request in 3.2551629543304443s
Received healthy response to inference request in 2.973440408706665s
5 requests
0 failed requests
5th percentile: 2.2728388786315916
10th percentile: 2.381336545944214
20th percentile: 2.598331880569458
30th percentile: 2.7257876873016356
40th percentile: 2.763703966140747
50th percentile: 2.8016202449798584
60th percentile: 2.870348310470581
70th percentile: 2.939076375961304
80th percentile: 3.029784917831421
90th percentile: 3.1424739360809326
95th percentile: 3.1988184452056885
99th percentile: 3.2438940525054933
mean time: 2.7802788734436037
Pipeline stage StressChecker completed in 36.08s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.62s
Shutdown handler de-registered
function_turam_2025-05-27 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3124.40s
Shutdown handler de-registered
function_turam_2025-05-27 status is now inactive due to auto deactivation removed underperforming models
function_turam_2025-05-27 status is now torndown due to DeploymentManager action