developer_uid: rirv938
submission_id: function_lefan_2025-03-14
model_name: dpo_data_collection
model_group:
status: torndown
timestamp: 2025-03-14T21:22:38+00:00
num_battles: 8646
num_wins: 4441
celo_rating: 1289.51
family_friendly_score: 0.5042
family_friendly_standard_error: 0.007070818340192315
submission_type: function
display_name: dpo_data_collection
is_internal_developer: True
ranking_group: single
us_pacific_date: 2025-03-14
win_ratio: 0.5136479296784641
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 3.65213942527771s
Received healthy response to inference request in 4.004378795623779s
Received healthy response to inference request in 3.5437490940093994s
Received healthy response to inference request in 8.885801792144775s
Received healthy response to inference request in 2.9731032848358154s
5 requests
0 failed requests
5th percentile: 3.0872324466705323
10th percentile: 3.201361608505249
20th percentile: 3.4296199321746825
30th percentile: 3.5654271602630616
40th percentile: 3.6087832927703856
50th percentile: 3.65213942527771
60th percentile: 3.7930351734161376
70th percentile: 3.9339309215545653
80th percentile: 4.980663394927979
90th percentile: 6.933232593536378
95th percentile: 7.909517192840576
99th percentile: 8.690544872283935
mean time: 4.611834478378296
%s, retrying in %s seconds...
Received healthy response to inference request in 3.384202241897583s
Received healthy response to inference request in 3.961641311645508s
Received healthy response to inference request in 3.1788740158081055s
Received healthy response to inference request in 3.4103081226348877s
Received healthy response to inference request in 3.704298496246338s
5 requests
0 failed requests
5th percentile: 3.219939661026001
10th percentile: 3.2610053062438964
20th percentile: 3.3431365966796873
30th percentile: 3.3894234180450438
40th percentile: 3.3998657703399657
50th percentile: 3.4103081226348877
60th percentile: 3.5279042720794678
70th percentile: 3.645500421524048
80th percentile: 3.755767059326172
90th percentile: 3.85870418548584
95th percentile: 3.9101727485656737
99th percentile: 3.951347599029541
mean time: 3.5278648376464843
Pipeline stage StressChecker completed in 42.94s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.67s
Shutdown handler de-registered
function_lefan_2025-03-14 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3609.14s
Shutdown handler de-registered
function_lefan_2025-03-14 status is now inactive due to auto deactivation removed underperforming models
function_lefan_2025-03-14 status is now torndown due to DeploymentManager action
ChatRequest
Generation Params
Prompt Formatter
Chat History
ChatMessage 1