developer_uid: chai_backend_admin
submission_id: function_gulal_2025-07-29
model_name: function_gulal_2025-07-29
model_group:
status: torndown
timestamp: 2025-07-29T18:49:36+00:00
num_battles: 6131
num_wins: 3299
celo_rating: 1290.2
family_friendly_score: 0.5286
family_friendly_standard_error: 0.007059490633183105
submission_type: function
display_name: function_gulal_2025-07-29
is_internal_developer: True
ranking_group: single
us_pacific_date: 2025-07-29
win_ratio: 0.5380851410862828
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': '### Instruction:\n{memory}\n', 'prompt_template': '### Input:\n{prompt}\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '### Response:\n{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 4.3341851234436035s
Received healthy response to inference request in 4.613892555236816s
Received healthy response to inference request in 5.74458646774292s
Received healthy response to inference request in 4.578588247299194s
Received healthy response to inference request in 2.6467909812927246s
5 requests
0 failed requests
5th percentile: 2.9842698097229006
10th percentile: 3.321748638153076
20th percentile: 3.996706295013428
30th percentile: 4.3830657482147215
40th percentile: 4.480826997756958
50th percentile: 4.578588247299194
60th percentile: 4.592709970474243
70th percentile: 4.606831693649292
80th percentile: 4.840031337738037
90th percentile: 5.292308902740478
95th percentile: 5.518447685241699
99th percentile: 5.699358711242676
mean time: 4.383608675003051
%s, retrying in %s seconds...
Received healthy response to inference request in 2.66634464263916s
Received healthy response to inference request in 3.1693460941314697s
Received healthy response to inference request in 2.8647544384002686s
Received healthy response to inference request in 3.335064172744751s
Received healthy response to inference request in 2.1564619541168213s
5 requests
0 failed requests
5th percentile: 2.258438491821289
10th percentile: 2.3604150295257567
20th percentile: 2.5643681049346925
30th percentile: 2.7060266017913817
40th percentile: 2.785390520095825
50th percentile: 2.8647544384002686
60th percentile: 2.986591100692749
70th percentile: 3.1084277629852295
80th percentile: 3.202489709854126
90th percentile: 3.2687769412994383
95th percentile: 3.3019205570220946
99th percentile: 3.3284354496002195
mean time: 2.8383942604064942
Pipeline stage StressChecker completed in 38.33s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.69s
Shutdown handler de-registered
function_gulal_2025-07-29 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3516.31s
Shutdown handler de-registered
function_gulal_2025-07-29 status is now inactive due to auto deactivation removed underperforming models
function_gulal_2025-07-29 status is now torndown due to DeploymentManager action