developer_uid: rirv938
submission_id: function_subos_2024-12-28
model_name: retune_with_base
model_group:
status: torndown
timestamp: 2024-12-28T15:40:49+00:00
num_battles: 15209
num_wins: 7473
celo_rating: 1259.94
family_friendly_score: 0.6036
family_friendly_standard_error: 0.006917615774238983
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2024-12-28
win_ratio: 0.49135380366888026
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 3.297438144683838s
Received healthy response to inference request in 4.365864038467407s
Received healthy response to inference request in 3.1103289127349854s
Received healthy response to inference request in 3.6727306842803955s
Received healthy response to inference request in 5.484554290771484s
5 requests
0 failed requests
5th percentile: 3.1477507591247558
10th percentile: 3.185172605514526
20th percentile: 3.2600162982940675
30th percentile: 3.3724966526031492
40th percentile: 3.5226136684417724
50th percentile: 3.6727306842803955
60th percentile: 3.9499840259552004
70th percentile: 4.227237367630004
80th percentile: 4.589602088928223
90th percentile: 5.037078189849853
95th percentile: 5.260816240310668
99th percentile: 5.439806680679321
mean time: 3.986183214187622
%s, retrying in %s seconds...
Received healthy response to inference request in 2.9343719482421875s
Received healthy response to inference request in 4.612874984741211s
Received healthy response to inference request in 3.1632628440856934s
Received healthy response to inference request in 2.77195143699646s
Received healthy response to inference request in 2.7638907432556152s
5 requests
0 failed requests
5th percentile: 2.7655028820037844
10th percentile: 2.767115020751953
20th percentile: 2.770339298248291
30th percentile: 2.8044355392456053
40th percentile: 2.8694037437438964
50th percentile: 2.9343719482421875
60th percentile: 3.0259283065795897
70th percentile: 3.1174846649169923
80th percentile: 3.453185272216797
90th percentile: 4.0330301284790036
95th percentile: 4.322952556610107
99th percentile: 4.5548904991149906
mean time: 3.2492703914642336
Pipeline stage StressChecker completed in 38.90s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.66s
Shutdown handler de-registered
function_subos_2024-12-28 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3047.00s
Shutdown handler de-registered
function_subos_2024-12-28 status is now inactive due to auto deactivation removed underperforming models
function_subos_2024-12-28 status is now torndown due to DeploymentManager action
ChatRequest
Generation Params
Prompt Formatter
Chat History
ChatMessage 1