developer_uid: chai_backend_admin
submission_id: function_harim_2024-12-08
model_name: retune_with_base
model_group:
status: inactive
timestamp: 2024-12-08T05:30:00+00:00
num_battles: 9640
num_wins: 5000
celo_rating: 1271.71
family_friendly_score: 0.597
family_friendly_standard_error: 0.006936728335461898
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2024-12-07
win_ratio: 0.5186721991701245
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.4945366382598877s
Received healthy response to inference request in 2.1415858268737793s
Received healthy response to inference request in 2.853837490081787s
Received healthy response to inference request in 2.2928032875061035s
read tcp 127.0.0.1:35916->127.0.0.1:8080: read: connection reset by peer
Received unhealthy response to inference request!
5 requests
1 failed requests
5th percentile: 1.050295352935791
10th percentile: 1.323117971420288
20th percentile: 1.8687632083892822
30th percentile: 2.1718293190002442
40th percentile: 2.2323163032531737
50th percentile: 2.2928032875061035
60th percentile: 2.3734966278076173
70th percentile: 2.454189968109131
80th percentile: 2.5663968086242677
90th percentile: 2.710117149353027
95th percentile: 2.781977319717407
99th percentile: 2.839465456008911
mean time: 2.1120471954345703
%s, retrying in %s seconds...
Received healthy response to inference request in 3.182626485824585s
Received healthy response to inference request in 2.3850934505462646s
Received healthy response to inference request in 2.35448956489563s
Received healthy response to inference request in 3.302128791809082s
Received healthy response to inference request in 1.5882236957550049s
5 requests
0 failed requests
5th percentile: 1.74147686958313
10th percentile: 1.8947300434112548
20th percentile: 2.201236391067505
30th percentile: 2.360610342025757
40th percentile: 2.3728518962860106
50th percentile: 2.3850934505462646
60th percentile: 2.704106664657593
70th percentile: 3.0231198787689206
80th percentile: 3.2065269470214846
90th percentile: 3.2543278694152833
95th percentile: 3.2782283306121824
99th percentile: 3.297348699569702
mean time: 2.5625123977661133
Pipeline stage StressChecker completed in 26.15s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 2.21s
Shutdown handler de-registered
function_harim_2024-12-08 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3027.50s
Shutdown handler de-registered
function_harim_2024-12-08 status is now inactive due to auto deactivation removed underperforming models