developer_uid: rirv938
submission_id: function_kuhor_2024-12-28
model_name: retune_with_base
model_group:
status: torndown
timestamp: 2024-12-28T15:35:13+00:00
num_battles: 12371
num_wins: 6342
celo_rating: 1272.84
family_friendly_score: 0.5946
family_friendly_standard_error: 0.006943354232645775
submission_type: function
display_name: retune_with_base
is_internal_developer: True
ranking_group: single
us_pacific_date: 2024-12-28
win_ratio: 0.5126505537143319
generation_params: {'temperature': 0.9, 'top_p': 0.9, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.5, 'frequency_penalty': 0.5, 'stopping_words': ['\n', '</s>'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage StressChecker
HTTPConnectionPool(host='guanaco-submitter.guanaco-backend.k2.chaiverse.com', port=80): Read timed out. (read timeout=20)
Received unhealthy response to inference request!
Received healthy response to inference request in 2.8152170181274414s
Received healthy response to inference request in 3.9475696086883545s
Received healthy response to inference request in 3.1810929775238037s
Received healthy response to inference request in 5.789816856384277s
5 requests
1 failed requests
5th percentile: 2.888392210006714
10th percentile: 2.9615674018859863
20th percentile: 3.1079177856445312
30th percentile: 3.3343883037567137
40th percentile: 3.640978956222534
50th percentile: 3.9475696086883545
60th percentile: 4.684468507766724
70th percentile: 5.421367406845093
80th percentile: 8.665357542037967
90th percentile: 14.416438913345338
95th percentile: 17.29197959899902
99th percentile: 19.592412147521973
mean time: 7.180243349075317
%s, retrying in %s seconds...
Received healthy response to inference request in 2.329822063446045s
Received healthy response to inference request in 3.588934898376465s
Received healthy response to inference request in 3.605818271636963s
Received healthy response to inference request in 3.71341609954834s
Received healthy response to inference request in 3.5401220321655273s
5 requests
0 failed requests
5th percentile: 2.5718820571899412
10th percentile: 2.813942050933838
20th percentile: 3.298062038421631
30th percentile: 3.5498846054077147
40th percentile: 3.5694097518920898
50th percentile: 3.588934898376465
60th percentile: 3.595688247680664
70th percentile: 3.602441596984863
80th percentile: 3.627337837219238
90th percentile: 3.6703769683837892
95th percentile: 3.6918965339660645
99th percentile: 3.709112186431885
mean time: 3.355622673034668
Pipeline stage StressChecker completed in 55.23s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.67s
Shutdown handler de-registered
function_kuhor_2024-12-28 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3121.14s
Shutdown handler de-registered
function_kuhor_2024-12-28 status is now inactive due to auto deactivation removed underperforming models
function_kuhor_2024-12-28 status is now torndown due to DeploymentManager action
ChatRequest
Generation Params
Prompt Formatter
Chat History
ChatMessage 1