developer_uid: chai_backend_admin
submission_id: function_fatil_2024-11-27
model_name: retune_with_base
model_group:
status: inactive
timestamp: 2024-11-27T22:12:20+00:00
num_battles: 9789
num_wins: 5012
celo_rating: 1265.97
family_friendly_score: 0.5944
family_friendly_standard_error: 0.006943898616771417
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2024-11-27
win_ratio: 0.5120032689753805
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 4.171592950820923s
Received healthy response to inference request in 2.8192784786224365s
Received healthy response to inference request in 4.3365373611450195s
Received healthy response to inference request in 3.2698471546173096s
Received healthy response to inference request in 2.251113176345825s
5 requests
0 failed requests
5th percentile: 2.3647462368011474
10th percentile: 2.4783792972564695
20th percentile: 2.7056454181671143
30th percentile: 2.909392213821411
40th percentile: 3.08961968421936
50th percentile: 3.2698471546173096
60th percentile: 3.630545473098755
70th percentile: 3.9912437915802
80th percentile: 4.204581832885742
90th percentile: 4.270559597015381
95th percentile: 4.3035484790802006
99th percentile: 4.329939584732056
mean time: 3.369673824310303
%s, retrying in %s seconds...
Received healthy response to inference request in 3.0858983993530273s
Received healthy response to inference request in 3.4575304985046387s
Received healthy response to inference request in 3.6143980026245117s
Received healthy response to inference request in 5.419351100921631s
Received healthy response to inference request in 2.841597318649292s
5 requests
0 failed requests
5th percentile: 2.890457534790039
10th percentile: 2.939317750930786
20th percentile: 3.0370381832122804
30th percentile: 3.1602248191833495
40th percentile: 3.3088776588439943
50th percentile: 3.4575304985046387
60th percentile: 3.520277500152588
70th percentile: 3.583024501800537
80th percentile: 3.975388622283936
90th percentile: 4.697369861602783
95th percentile: 5.0583604812622065
99th percentile: 5.347152976989746
mean time: 3.68375506401062
Pipeline stage StressChecker completed in 37.33s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 2.06s
Shutdown handler de-registered
function_fatil_2024-11-27 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 4801.10s
Shutdown handler de-registered
function_fatil_2024-11-27 status is now inactive due to auto deactivation removed underperforming models