submission_id: function_teson_2024-11-19
developer_uid: chai_backend_admin
celo_rating: 1278.43
display_name: retune_with_base
family_friendly_score: 0.5774
family_friendly_standard_error: 0.006985831947592212
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
is_internal_developer: True
model_group:
model_name: retune_with_base
num_battles: 10084
num_wins: 5387
ranking_group: single
status: inactive
submission_type: function
timestamp: 2024-11-19T21:49:35+00:00
us_pacific_date: 2024-11-19
win_ratio: 0.5342126140420468
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.5770115852355957s
Received healthy response to inference request in 4.11623215675354s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Received healthy response to inference request in 2.630164861679077s
Received healthy response to inference request in 4.332850933074951s
Received healthy response to inference request in 3.5171937942504883s
5 requests
0 failed requests
5th percentile: 2.587642240524292
10th percentile: 2.5982728958129884
20th percentile: 2.619534206390381
30th percentile: 2.8075706481933596
40th percentile: 3.162382221221924
50th percentile: 3.5171937942504883
60th percentile: 3.756809139251709
70th percentile: 3.9964244842529295
80th percentile: 4.159555912017822
90th percentile: 4.246203422546387
95th percentile: 4.289527177810669
99th percentile: 4.324186182022094
mean time: 3.4346906661987306
%s, retrying in %s seconds...
Received healthy response to inference request in 3.36212158203125s
Received healthy response to inference request in 2.9172818660736084s
Received healthy response to inference request in 3.871727466583252s
Received healthy response to inference request in 3.283595323562622s
Received healthy response to inference request in 2.6253347396850586s
5 requests
0 failed requests
5th percentile: 2.6837241649627686
10th percentile: 2.7421135902404785
20th percentile: 2.8588924407958984
30th percentile: 2.990544557571411
40th percentile: 3.1370699405670166
50th percentile: 3.283595323562622
60th percentile: 3.315005826950073
70th percentile: 3.3464163303375245
80th percentile: 3.4640427589416505
90th percentile: 3.667885112762451
95th percentile: 3.7698062896728515
99th percentile: 3.851343231201172
mean time: 3.212012195587158
Pipeline stage StressChecker completed in 35.69s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 2.36s
Shutdown handler de-registered
function_teson_2024-11-19 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3456.75s
Shutdown handler de-registered
function_teson_2024-11-19 status is now inactive due to auto deactivation removed underperforming models