developer_uid: rirv938
submission_id: function_lobor_2025-02-14
model_name: retune_with_base
model_group:
status: torndown
timestamp: 2025-02-14T20:20:05+00:00
num_battles: 6562
num_wins: 3307
celo_rating: 1274.14
family_friendly_score: 0.5804
family_friendly_standard_error: 0.006979052084631551
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2025-02-14
win_ratio: 0.5039622066443158
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 14.474820613861084s
Received healthy response to inference request in 10.789604902267456s
Received healthy response to inference request in 10.04557728767395s
Received healthy response to inference request in 3.196687698364258s
Received healthy response to inference request in 2.515998125076294s
5 requests
0 failed requests
5th percentile: 2.652136039733887
10th percentile: 2.7882739543914794
20th percentile: 3.060549783706665
30th percentile: 4.566465616226196
40th percentile: 7.306021451950073
50th percentile: 10.04557728767395
60th percentile: 10.343188333511353
70th percentile: 10.640799379348755
80th percentile: 11.526648044586182
90th percentile: 13.000734329223633
95th percentile: 13.737777471542358
99th percentile: 14.327411985397339
mean time: 8.204537725448608
%s, retrying in %s seconds...
Received healthy response to inference request in 2.4996395111083984s
Received healthy response to inference request in 1.9609484672546387s
Received healthy response to inference request in 3.0733299255371094s
Received healthy response to inference request in 5.606405258178711s
Received healthy response to inference request in 3.649379014968872s
5 requests
0 failed requests
5th percentile: 2.068686676025391
10th percentile: 2.1764248847961425
20th percentile: 2.3919013023376463
30th percentile: 2.6143775939941407
40th percentile: 2.843853759765625
50th percentile: 3.0733299255371094
60th percentile: 3.3037495613098145
70th percentile: 3.5341691970825195
80th percentile: 4.04078426361084
90th percentile: 4.823594760894776
95th percentile: 5.215000009536743
99th percentile: 5.5281242084503175
mean time: 3.357940435409546
Pipeline stage StressChecker completed in 60.26s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.71s
Shutdown handler de-registered
function_lobor_2025-02-14 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3301.86s
Shutdown handler de-registered
function_lobor_2025-02-14 status is now inactive due to auto deactivation removed underperforming models
function_lobor_2025-02-14 status is now torndown due to DeploymentManager action
function_lobor_2025-02-14 status is now torndown due to DeploymentManager action
function_lobor_2025-02-14 status is now torndown due to DeploymentManager action
ChatRequest
Generation Params
Prompt Formatter
Chat History
ChatMessage 1