developer_uid: chai_backend_admin
submission_id: function_rolam_2024-11-29
model_name: retune_with_base
model_group:
status: inactive
timestamp: 2024-11-29T02:32:19+00:00
num_battles: 9140
num_wins: 4546
celo_rating: 1263.84
family_friendly_score: 0.5882000000000001
family_friendly_standard_error: 0.006960183330918806
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2024-11-28
win_ratio: 0.4973741794310722
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.286323308944702s
Received healthy response to inference request in 4.0670812129974365s
Received healthy response to inference request in 3.882652521133423s
Received healthy response to inference request in 4.2112836837768555s
Received healthy response to inference request in 3.125636100769043s
5 requests
0 failed requests
5th percentile: 2.4541858673095702
10th percentile: 2.6220484256744383
20th percentile: 2.957773542404175
30th percentile: 3.277039384841919
40th percentile: 3.5798459529876707
50th percentile: 3.882652521133423
60th percentile: 3.9564239978790283
70th percentile: 4.030195474624634
80th percentile: 4.0959217071533205
90th percentile: 4.1536026954650875
95th percentile: 4.1824431896209715
99th percentile: 4.205515584945679
mean time: 3.514595365524292
%s, retrying in %s seconds...
Received healthy response to inference request in 5.147648096084595s
Received healthy response to inference request in 3.5166845321655273s
Received healthy response to inference request in 2.5937728881835938s
Received healthy response to inference request in 3.157517671585083s
Received healthy response to inference request in 3.935335397720337s
5 requests
0 failed requests
5th percentile: 2.7065218448638917
10th percentile: 2.8192708015441896
20th percentile: 3.044768714904785
30th percentile: 3.229351043701172
40th percentile: 3.3730177879333496
50th percentile: 3.5166845321655273
60th percentile: 3.684144878387451
70th percentile: 3.851605224609375
80th percentile: 4.177797937393189
90th percentile: 4.662723016738892
95th percentile: 4.905185556411743
99th percentile: 5.099155588150024
mean time: 3.6701917171478273
%s, retrying in %s seconds...
Received healthy response to inference request in 2.3243467807769775s
Received healthy response to inference request in 2.2845709323883057s
Received healthy response to inference request in 3.741060972213745s
Received healthy response to inference request in 3.8401036262512207s
Received healthy response to inference request in 2.118525743484497s
5 requests
0 failed requests
5th percentile: 2.151734781265259
10th percentile: 2.1849438190460204
20th percentile: 2.2513618946075438
30th percentile: 2.29252610206604
40th percentile: 2.3084364414215086
50th percentile: 2.3243467807769775
60th percentile: 2.8910324573516846
70th percentile: 3.457718133926391
80th percentile: 3.7608695030212402
90th percentile: 3.8004865646362305
95th percentile: 3.8202950954437256
99th percentile: 3.8361419200897218
mean time: 2.8617216110229493
Pipeline stage StressChecker completed in 53.70s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 2.15s
Shutdown handler de-registered
function_rolam_2024-11-29 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 4096.86s
Shutdown handler de-registered
function_rolam_2024-11-29 status is now inactive due to auto deactivation removed underperforming models