developer_uid: chai_backend_admin
submission_id: function_bonur_2025-12-19
model_name: function_bonur_2025-12-19
model_group:
status: torndown
timestamp: 2025-12-22T22:41:23+00:00
num_battles: 6019
num_wins: 3133
celo_rating: 1307.45
family_friendly_score: 0.0
family_friendly_standard_error: 0.0
submission_type: function
display_name: function_bonur_2025-12-19
is_internal_developer: True
ranking_group: single
us_pacific_date: 2025-12-22
win_ratio: 0.5205183585313174
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': '### Instruction:\n{memory}\n', 'prompt_template': '### Input:\n{prompt}\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '### Response:\n{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
HTTPConnectionPool(host='guanaco-submitter.guanaco-backend.k2.chaiverse.com', port=80): Read timed out. (read timeout=20)
Received unhealthy response to inference request!
Received healthy response to inference request in 10.039057970046997s
Received healthy response to inference request in 10.35119342803955s
Received healthy response to inference request in 8.954373836517334s
Received healthy response to inference request in 8.66072392463684s
Received healthy response to inference request in 8.008966207504272s
Received healthy response to inference request in 6.788231372833252s
Received healthy response to inference request in 1.4568226337432861s
Received healthy response to inference request in 3.875722646713257s
Received healthy response to inference request in 11.952643156051636s
10 requests
1 failed requests
5th percentile: 2.545327639579773
10th percentile: 3.63383264541626
20th percentile: 6.205729627609253
30th percentile: 7.642745757102966
40th percentile: 8.400020837783813
50th percentile: 8.807548880577087
60th percentile: 9.3882474899292
70th percentile: 10.132698607444762
80th percentile: 10.671483373641967
90th percentile: 12.76866965293884
95th percentile: 16.440788888931266
99th percentile: 19.37848427772522
mean time: 9.020064330101013
%s, retrying in %s seconds...
Received healthy response to inference request in 5.989790439605713s
Received healthy response to inference request in 1.7445948123931885s
Received healthy response to inference request in 5.133419990539551s
Received healthy response to inference request in 14.088727474212646s
Received healthy response to inference request in 6.961201906204224s
Received healthy response to inference request in 3.412156105041504s
Received healthy response to inference request in 9.445550203323364s
Received healthy response to inference request in 4.2743518352508545s
Received healthy response to inference request in 2.22170090675354s
Received healthy response to inference request in 2.7759957313537598s
10 requests
0 failed requests
5th percentile: 1.9592925548553466
10th percentile: 2.1739902973175047
20th percentile: 2.665136766433716
30th percentile: 3.2213079929351807
40th percentile: 3.929473543167114
50th percentile: 4.703885912895203
60th percentile: 5.4759681701660154
70th percentile: 6.281213879585266
80th percentile: 7.458071565628052
90th percentile: 9.909867930412291
95th percentile: 11.999297702312465
99th percentile: 13.670841519832612
mean time: 5.604748940467834
Pipeline stage StressChecker completed in 150.01s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 1.06s
Shutdown handler de-registered
function_bonur_2025-12-19 status is now deployed due to DeploymentManager action
function_bonur_2025-12-19 status is now inactive due to auto deactivation removed underperforming models
function_bonur_2025-12-19 status is now torndown due to DeploymentManager action