submission_id: riverise-dpo-second_v3
developer_uid: Riverise
alignment_samples: 12438
alignment_score: -0.08492816417093157
best_of: 16
celo_rating: 1238.04
display_name: riverise-dpo-second_v1
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 1.15, 'top_p': 0.95, 'min_p': 0.05, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 512, 'best_of': 16, 'max_output_tokens': 64}
is_internal_developer: False
language_model: Riverise/dpo-second
max_input_tokens: 512
max_output_tokens: 64
model_architecture: LlamaForCausalLM
model_group: Riverise/dpo-second
model_name: riverise-dpo-second_v1
model_num_parameters: 8030261248.0
model_repo: Riverise/dpo-second
model_size: 8B
num_battles: 12438
num_wins: 6293
propriety_score: 0.7103574702108157
propriety_total_count: 1091.0
ranking_group: single
status: inactive
submission_type: basic
timestamp: 2024-09-01T09:05:36+00:00
us_pacific_date: 2024-09-01
win_ratio: 0.5059495095674545
Download Preference Data
Resubmit model
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLizer
Starting job with name riverise-dpo-second-v3-mkmlizer
Waiting for job on riverise-dpo-second-v3-mkmlizer to finish
riverise-dpo-second-v3-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
riverise-dpo-second-v3-mkmlizer: ║ _____ __ __ ║
riverise-dpo-second-v3-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
riverise-dpo-second-v3-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
riverise-dpo-second-v3-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
riverise-dpo-second-v3-mkmlizer: ║ /___/ ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ║ Version: 0.10.1 ║
riverise-dpo-second-v3-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
riverise-dpo-second-v3-mkmlizer: ║ https://mk1.ai ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ║ The license key for the current software has been verified as ║
riverise-dpo-second-v3-mkmlizer: ║ belonging to: ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ║ Chai Research Corp. ║
riverise-dpo-second-v3-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
riverise-dpo-second-v3-mkmlizer: ║ Expiration: 2024-10-15 23:59:59 ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
riverise-dpo-second-v3-mkmlizer: Downloaded to shared memory in 51.454s
riverise-dpo-second-v3-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmp_ikltr79, device:0
riverise-dpo-second-v3-mkmlizer: Saving flywheel model at /dev/shm/model_cache
riverise-dpo-second-v3-mkmlizer: quantized model in 29.281s
riverise-dpo-second-v3-mkmlizer: Processed model Riverise/dpo-second in 80.735s
riverise-dpo-second-v3-mkmlizer: creating bucket guanaco-mkml-models
riverise-dpo-second-v3-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
riverise-dpo-second-v3-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/riverise-dpo-second-v3
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/riverise-dpo-second-v3/config.json
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/riverise-dpo-second-v3/special_tokens_map.json
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/riverise-dpo-second-v3/tokenizer_config.json
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/riverise-dpo-second-v3/tokenizer.json
Stopping job with name riverise-dpo-second-v3-mkmlizer
%s, retrying in %s seconds...
Stopping job with name riverise-dpo-second-v3-mkmlizer
%s, retrying in %s seconds...
Stopping job with name riverise-dpo-second-v3-mkmlizer
%s, retrying in %s seconds...
Starting job with name riverise-dpo-second-v3-mkmlizer
Waiting for job on riverise-dpo-second-v3-mkmlizer to finish
Stopping job with name riverise-dpo-second-v3-mkmlizer
%s, retrying in %s seconds...
Starting job with name riverise-dpo-second-v3-mkmlizer
Waiting for job on riverise-dpo-second-v3-mkmlizer to finish
riverise-dpo-second-v3-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
riverise-dpo-second-v3-mkmlizer: ║ _____ __ __ ║
riverise-dpo-second-v3-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
riverise-dpo-second-v3-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
riverise-dpo-second-v3-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
riverise-dpo-second-v3-mkmlizer: ║ /___/ ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ║ Version: 0.10.1 ║
riverise-dpo-second-v3-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
riverise-dpo-second-v3-mkmlizer: ║ https://mk1.ai ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ║ The license key for the current software has been verified as ║
riverise-dpo-second-v3-mkmlizer: ║ belonging to: ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ║ Chai Research Corp. ║
riverise-dpo-second-v3-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
riverise-dpo-second-v3-mkmlizer: ║ Expiration: 2024-10-15 23:59:59 ║
riverise-dpo-second-v3-mkmlizer: ║ ║
riverise-dpo-second-v3-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
riverise-dpo-second-v3-mkmlizer: Downloaded to shared memory in 38.679s
riverise-dpo-second-v3-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmpc2r81zu5, device:0
riverise-dpo-second-v3-mkmlizer: Saving flywheel model at /dev/shm/model_cache
riverise-dpo-second-v3-mkmlizer: quantized model in 28.302s
riverise-dpo-second-v3-mkmlizer: Processed model Riverise/dpo-second in 66.982s
riverise-dpo-second-v3-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
riverise-dpo-second-v3-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/riverise-dpo-second-v3
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/riverise-dpo-second-v3/config.json
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/riverise-dpo-second-v3/special_tokens_map.json
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/riverise-dpo-second-v3/tokenizer_config.json
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/riverise-dpo-second-v3/tokenizer.json
riverise-dpo-second-v3-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/riverise-dpo-second-v3/flywheel_model.0.safetensors
riverise-dpo-second-v3-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 2%|▏ | 5/291 [00:00<00:10, 26.18it/s] Loading 0: 4%|▍ | 12/291 [00:00<00:07, 35.82it/s] Loading 0: 5%|▌ | 16/291 [00:00<00:08, 33.94it/s] Loading 0: 7%|▋ | 21/291 [00:00<00:07, 37.50it/s] Loading 0: 9%|▊ | 25/291 [00:00<00:07, 35.64it/s] Loading 0: 10%|█ | 30/291 [00:00<00:06, 39.04it/s] Loading 0: 12%|█▏ | 34/291 [00:01<00:09, 26.72it/s] Loading 0: 13%|█▎ | 38/291 [00:01<00:08, 28.42it/s] Loading 0: 14%|█▍ | 42/291 [00:01<00:09, 27.37it/s] Loading 0: 16%|█▋ | 48/291 [00:01<00:07, 33.18it/s] Loading 0: 18%|█▊ | 52/291 [00:01<00:07, 32.65it/s] Loading 0: 20%|█▉ | 57/291 [00:01<00:06, 34.75it/s] Loading 0: 21%|██ | 61/291 [00:01<00:06, 33.30it/s] Loading 0: 23%|██▎ | 66/291 [00:01<00:06, 36.24it/s] Loading 0: 24%|██▍ | 70/291 [00:02<00:06, 34.95it/s] Loading 0: 25%|██▌ | 74/291 [00:02<00:06, 34.95it/s] Loading 0: 27%|██▋ | 78/291 [00:02<00:06, 35.37it/s] Loading 0: 28%|██▊ | 82/291 [00:02<00:08, 25.62it/s] Loading 0: 30%|██▉ | 86/291 [00:02<00:07, 28.23it/s] Loading 0: 31%|███ | 90/291 [00:02<00:06, 30.16it/s] Loading 0: 32%|███▏ | 94/291 [00:02<00:06, 30.44it/s] Loading 0: 34%|███▍ | 99/291 [00:03<00:05, 33.59it/s] Loading 0: 35%|███▌ | 103/291 [00:03<00:05, 32.84it/s] Loading 0: 37%|███▋ | 108/291 [00:03<00:05, 36.24it/s] Loading 0: 38%|███▊ | 112/291 [00:03<00:05, 35.20it/s] Loading 0: 40%|███▉ | 116/291 [00:03<00:04, 35.13it/s] Loading 0: 42%|████▏ | 122/291 [00:03<00:04, 39.64it/s] Loading 0: 44%|████▎ | 127/291 [00:03<00:04, 37.39it/s] Loading 0: 46%|████▌ | 133/291 [00:04<00:04, 31.66it/s] Loading 0: 47%|████▋ | 137/291 [00:04<00:04, 31.98it/s] Loading 0: 48%|████▊ | 141/291 [00:04<00:05, 29.94it/s] Loading 0: 51%|█████ | 147/291 [00:04<00:04, 34.46it/s] Loading 0: 52%|█████▏ | 151/291 [00:04<00:04, 33.89it/s] Loading 0: 54%|█████▎ | 156/291 [00:04<00:03, 36.53it/s] Loading 0: 55%|█████▍ | 160/291 [00:04<00:03, 34.26it/s] Loading 0: 57%|█████▋ | 165/291 [00:04<00:03, 36.51it/s] Loading 0: 58%|█████▊ | 169/291 [00:05<00:03, 33.04it/s] Loading 0: 60%|█████▉ | 174/291 [00:05<00:03, 35.57it/s] Loading 0: 61%|██████ | 178/291 [00:05<00:03, 33.64it/s] Loading 0: 63%|██████▎ | 184/291 [00:05<00:02, 39.94it/s] Loading 0: 65%|██████▍ | 189/291 [00:05<00:04, 24.98it/s] Loading 0: 67%|██████▋ | 194/291 [00:05<00:03, 26.29it/s] Loading 0: 69%|██████▉ | 201/291 [00:06<00:02, 32.78it/s] Loading 0: 70%|███████ | 205/291 [00:06<00:02, 31.98it/s] Loading 0: 72%|███████▏ | 210/291 [00:06<00:02, 34.81it/s] Loading 0: 74%|███████▎ | 214/291 [00:06<00:02, 33.38it/s] Loading 0: 75%|███████▌ | 219/291 [00:06<00:01, 36.02it/s] Loading 0: 77%|███████▋ | 223/291 [00:06<00:01, 34.54it/s] Loading 0: 78%|███████▊ | 227/291 [00:06<00:01, 34.90it/s] Loading 0: 79%|███████▉ | 231/291 [00:06<00:01, 35.23it/s] Loading 0: 81%|████████ | 235/291 [00:07<00:02, 26.35it/s] Loading 0: 82%|████████▏ | 239/291 [00:07<00:01, 26.10it/s] Loading 0: 85%|████████▍ | 246/291 [00:07<00:01, 33.61it/s] Loading 0: 86%|████████▌ | 250/291 [00:07<00:01, 32.75it/s] Loading 0: 88%|████████▊ | 255/291 [00:07<00:01, 35.68it/s] Loading 0: 89%|████████▉ | 259/291 [00:07<00:00, 34.71it/s] Loading 0: 91%|█████████ | 264/291 [00:07<00:00, 36.85it/s] Loading 0: 92%|█████████▏| 268/291 [00:08<00:00, 35.20it/s] Loading 0: 94%|█████████▍| 273/291 [00:08<00:00, 37.94it/s] Loading 0: 95%|█████████▌| 277/291 [00:08<00:00, 35.81it/s] Loading 0: 97%|█████████▋| 281/291 [00:08<00:00, 35.24it/s] Loading 0: 98%|█████████▊| 286/291 [00:13<00:01, 2.62it/s] Loading 0: 99%|█████████▉| 289/291 [00:14<00:00, 3.26it/s]
Job riverise-dpo-second-v3-mkmlizer completed after 85.46s with status: succeeded
Stopping job with name riverise-dpo-second-v3-mkmlizer
Pipeline stage MKMLizer completed in 193.24s
run pipeline stage %s
Running pipeline stage MKMLTemplater
Pipeline stage MKMLTemplater completed in 0.09s
run pipeline stage %s
Running pipeline stage MKMLDeployer
Creating inference service riverise-dpo-second-v3
Ignoring service riverise-dpo-second-v3 already deployed
Waiting for inference service riverise-dpo-second-v3 to be ready
Failed to get response for submission riverise-dpo-second_v2: ('http://riverise-dpo-second-v2-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'read tcp 127.0.0.1:40778->127.0.0.1:8080: read: connection reset by peer\n')
Inference service riverise-dpo-second-v3 ready after 110.4555549621582s
Pipeline stage MKMLDeployer completed in 111.29s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.6135590076446533s
Received healthy response to inference request in 1.4427003860473633s
Received healthy response to inference request in 1.5289685726165771s
Received healthy response to inference request in 1.8897480964660645s
Received healthy response to inference request in 1.7863225936889648s
5 requests
0 failed requests
5th percentile: 1.459954023361206
10th percentile: 1.4772076606750488
20th percentile: 1.5117149353027344
30th percentile: 1.5804393768310547
40th percentile: 1.6833809852600097
50th percentile: 1.7863225936889648
60th percentile: 1.8276927947998047
70th percentile: 1.8690629959106446
80th percentile: 2.0345102787017826
90th percentile: 2.324034643173218
95th percentile: 2.468796825408935
99th percentile: 2.5846065711975097
mean time: 1.8522597312927247
Pipeline stage StressChecker completed in 10.45s
run pipeline stage %s
Running pipeline stage TriggerMKMLProfilingPipeline
starting trigger_guanaco_pipeline %s
Pipeline stage TriggerMKMLProfilingPipeline completed in 8.32s
riverise-dpo-second_v3 status is now deployed due to DeploymentManager action
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLProfilerTemplater
Pipeline stage MKMLProfilerTemplater completed in 0.14s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeployer
Creating inference service riverise-dpo-second-v3-profiler
Waiting for inference service riverise-dpo-second-v3-profiler to be ready
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLProfilerTemplater
Pipeline stage %s skipped, reason=%s
Pipeline stage MKMLProfilerTemplater completed in 0.07s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeployer
Creating inference service riverise-dpo-second-v3-profiler
Ignoring service riverise-dpo-second-v3-profiler already deployed
Waiting for inference service riverise-dpo-second-v3-profiler to be ready
Tearing down inference service riverise-dpo-second-v3-profiler
%s, retrying in %s seconds...
Creating inference service riverise-dpo-second-v3-profiler
Waiting for inference service riverise-dpo-second-v3-profiler to be ready
Inference service riverise-dpo-second-v3-profiler ready after 160.38577914237976s
Pipeline stage MKMLProfilerDeployer completed in 341.68s
run pipeline stage %s
Running pipeline stage MKMLProfilerRunner
script pods %s
Pipeline stage MKMLProfilerRunner completed in 0.51s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeleter
Checking if service riverise-dpo-second-v3-profiler is running
Tearing down inference service riverise-dpo-second-v3-profiler
Service riverise-dpo-second-v3-profiler has been torndown
Pipeline stage MKMLProfilerDeleter completed in 1.98s
riverise-dpo-second_v3 status is now inactive due to auto deactivation removed underperforming models

Usage Metrics

Latency Metrics