submission_id: function_nibul_2024-11-14
developer_uid: chai_backend_admin
celo_rating: 1249.07
display_name: retune_with_base
family_friendly_score: 0.585
family_friendly_standard_error: 0.00696814178960216
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
is_internal_developer: True
model_group:
model_name: retune_with_base
num_battles: 8164
num_wins: 4075
ranking_group: single
status: inactive
submission_type: function
timestamp: 2024-11-14T19:36:02+00:00
us_pacific_date: 2024-11-14
win_ratio: 0.4991425771680549
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 5.170519590377808s
Received healthy response to inference request in 4.535920858383179s
Received healthy response to inference request in 5.296108961105347s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Received healthy response to inference request in 5.159756660461426s
Received healthy response to inference request in 4.185844898223877s
5 requests
0 failed requests
5th percentile: 4.255860090255737
10th percentile: 4.325875282287598
20th percentile: 4.465905666351318
30th percentile: 4.660688018798828
40th percentile: 4.910222339630127
50th percentile: 5.159756660461426
60th percentile: 5.164061832427978
70th percentile: 5.168367004394531
80th percentile: 5.195637464523315
90th percentile: 5.245873212814331
95th percentile: 5.270991086959839
99th percentile: 5.2910853862762455
mean time: 4.869630193710327
%s, retrying in %s seconds...
Received healthy response to inference request in 3.7024831771850586s
Received healthy response to inference request in 2.069373846054077s
Received healthy response to inference request in 3.643240213394165s
Received healthy response to inference request in 2.19118070602417s
Received healthy response to inference request in 2.2495195865631104s
5 requests
0 failed requests
5th percentile: 2.0937352180480957
10th percentile: 2.1180965900421143
20th percentile: 2.1668193340301514
30th percentile: 2.202848482131958
40th percentile: 2.226184034347534
50th percentile: 2.2495195865631104
60th percentile: 2.807007837295532
70th percentile: 3.3644960880279537
80th percentile: 3.6550888061523437
90th percentile: 3.6787859916687013
95th percentile: 3.69063458442688
99th percentile: 3.700113458633423
mean time: 2.7711595058441163
Pipeline stage StressChecker completed in 41.77s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 2.86s
Shutdown handler de-registered
function_nibul_2024-11-14 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 2406.95s
Shutdown handler de-registered
function_nibul_2024-11-14 status is now inactive due to auto deactivation removed underperforming models