developer_uid: chai_backend_admin
submission_id: function_dorol_2024-11-29
model_name: retune_with_base
model_group:
status: inactive
timestamp: 2024-11-29T18:33:43+00:00
num_battles: 11046
num_wins: 5542
celo_rating: 1258.5
family_friendly_score: 0.5864
family_friendly_standard_error: 0.006964697265495466
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2024-11-29
win_ratio: 0.5017200796668477
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.8351218700408936s
Received healthy response to inference request in 4.660312652587891s
Received healthy response to inference request in 3.9555609226226807s
Received healthy response to inference request in 2.9334921836853027s
Received healthy response to inference request in 3.238001585006714s
5 requests
0 failed requests
5th percentile: 2.8547959327697754
10th percentile: 2.8744699954986572
20th percentile: 2.913818120956421
30th percentile: 2.994394063949585
40th percentile: 3.1161978244781494
50th percentile: 3.238001585006714
60th percentile: 3.5250253200531008
70th percentile: 3.812049055099487
80th percentile: 4.096511268615723
90th percentile: 4.378411960601807
95th percentile: 4.519362306594848
99th percentile: 4.6321225833892825
mean time: 3.5244978427886964
%s, retrying in %s seconds...
Received healthy response to inference request in 4.881753444671631s
Received healthy response to inference request in 3.1234796047210693s
Received healthy response to inference request in 1.7396445274353027s
Received healthy response to inference request in 2.6476948261260986s
Received healthy response to inference request in 3.08951997756958s
5 requests
0 failed requests
5th percentile: 1.9212545871734619
10th percentile: 2.102864646911621
20th percentile: 2.4660847663879393
30th percentile: 2.736059856414795
40th percentile: 2.9127899169921876
50th percentile: 3.08951997756958
60th percentile: 3.103103828430176
70th percentile: 3.1166876792907714
80th percentile: 3.475134372711182
90th percentile: 4.178443908691406
95th percentile: 4.5300986766815186
99th percentile: 4.811422491073609
mean time: 3.0964184761047364
Pipeline stage StressChecker completed in 35.62s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 2.40s
Shutdown handler de-registered
function_dorol_2024-11-29 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 4004.15s
Shutdown handler de-registered
function_dorol_2024-11-29 status is now inactive due to auto deactivation removed underperforming models