developer_uid: Nitral-AI
submission_id: nitral-ai-captain-bmo-12b_v13
model_name: nitral-ai-captain-bmo-12b_v1
model_group: Nitral-AI/Captain_BMO-12
status: inactive
timestamp: 2024-11-15T02:57:42+00:00
num_battles: 10255
num_wins: 5734
celo_rating: 1289.81
family_friendly_score: 0.5726
family_friendly_standard_error: 0.006996130930735931
submission_type: basic
model_repo: Nitral-AI/Captain_BMO-12B
model_architecture: MistralForCausalLM
model_num_parameters: 12772070400.0
best_of: 8
max_input_tokens: 1024
max_output_tokens: 64
latencies: [{'batch_size': 1, 'throughput': 0.6115451706803039, 'latency_mean': 1.6351123976707458, 'latency_p50': 1.6513822078704834, 'latency_p90': 1.7859539270401001}, {'batch_size': 3, 'throughput': 1.1447453468464557, 'latency_mean': 2.6144292545318604, 'latency_p50': 2.615292549133301, 'latency_p90': 2.8605011224746706}, {'batch_size': 5, 'throughput': 1.384411492609272, 'latency_mean': 3.593752369880676, 'latency_p50': 3.6261651515960693, 'latency_p90': 3.998184037208557}, {'batch_size': 6, 'throughput': 1.457287835208021, 'latency_mean': 4.094273961782456, 'latency_p50': 4.100279450416565, 'latency_p90': 4.587147831916809}, {'batch_size': 8, 'throughput': 1.533744346416595, 'latency_mean': 5.186730515956879, 'latency_p50': 5.15315568447113, 'latency_p90': 5.843928408622742}, {'batch_size': 10, 'throughput': 1.579242929365689, 'latency_mean': 6.298097944259643, 'latency_p50': 6.284879207611084, 'latency_p90': 7.179172825813294}]
gpu_counts: {'NVIDIA RTX A5000': 1}
display_name: nitral-ai-captain-bmo-12b_v1
is_internal_developer: False
language_model: Nitral-AI/Captain_BMO-12B
model_size: 13B
ranking_group: single
throughput_3p7s: 1.41
us_pacific_date: 2024-11-14
win_ratio: 0.5591418820087762
generation_params: {'temperature': 1.2, 'top_p': 1.0, 'min_p': 0.1, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['</s>', '\n\n', '\nYou:', '<|im_end|>', '[/INST]'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': '[INST]system\n{memory}[/INST]\n', 'prompt_template': '[INST]user\n{prompt}[/INST]\n', 'bot_template': '[INST]assistant\n{bot_name}: {message}[/INST]\n', 'user_template': '[INST]user\n{user_name}: {message}[/INST]\n', 'response_template': '[INST]assistant\n{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLizer
Starting job with name nitral-ai-captain-bmo-12b-v13-mkmlizer
Waiting for job on nitral-ai-captain-bmo-12b-v13-mkmlizer to finish
nitral-ai-captain-bmo-12b-v13-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ _____ __ __ ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ /___/ ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ Version: 0.11.12 ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ https://mk1.ai ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ The license key for the current software has been verified as ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ belonging to: ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ Chai Research Corp. ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ Expiration: 2025-01-15 23:59:59 ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v13-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
Connection pool is full, discarding connection: %s. Connection pool size: %s
nitral-ai-captain-bmo-12b-v13-mkmlizer: Downloaded to shared memory in 32.021s
nitral-ai-captain-bmo-12b-v13-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmpovjfikox, device:0
nitral-ai-captain-bmo-12b-v13-mkmlizer: Saving flywheel model at /dev/shm/model_cache
nitral-ai-captain-bmo-12b-v13-mkmlizer: quantized model in 36.405s
nitral-ai-captain-bmo-12b-v13-mkmlizer: Processed model Nitral-AI/Captain_BMO-12B in 68.426s
nitral-ai-captain-bmo-12b-v13-mkmlizer: creating bucket guanaco-mkml-models
nitral-ai-captain-bmo-12b-v13-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
nitral-ai-captain-bmo-12b-v13-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v13
nitral-ai-captain-bmo-12b-v13-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v13/config.json
nitral-ai-captain-bmo-12b-v13-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v13/special_tokens_map.json
nitral-ai-captain-bmo-12b-v13-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v13/tokenizer_config.json
nitral-ai-captain-bmo-12b-v13-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v13/tokenizer.json
Failed to get response for submission rica40325-10-14dpo_v2: ('http://rica40325-10-14dpo-v2-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
nitral-ai-captain-bmo-12b-v13-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v13/flywheel_model.0.safetensors
nitral-ai-captain-bmo-12b-v13-mkmlizer: Loading 0: 0%| | 0/363 [00:00<?, ?it/s] Loading 0: 1%| | 2/363 [00:06<18:25, 3.06s/it] Loading 0: 2%|▏ | 6/363 [00:06<04:53, 1.22it/s] Loading 0: 4%|▎ | 13/363 [00:06<01:44, 3.34it/s] Loading 0: 5%|▍ | 17/363 [00:06<01:10, 4.88it/s] Loading 0: 6%|▋ | 23/363 [00:06<00:42, 7.97it/s] Loading 0: 8%|▊ | 29/363 [00:06<00:30, 11.03it/s] Loading 0: 9%|▉ | 33/363 [00:06<00:25, 13.19it/s] Loading 0: 11%|█ | 40/363 [00:07<00:20, 15.84it/s] Loading 0: 12%|█▏ | 44/363 [00:07<00:17, 18.37it/s] Loading 0: 13%|█▎ | 49/363 [00:07<00:13, 22.66it/s] Loading 0: 15%|█▍ | 53/363 [00:07<00:12, 24.90it/s] Loading 0: 16%|█▌ | 58/363 [00:07<00:10, 29.24it/s] Loading 0: 17%|█▋ | 62/363 [00:07<00:09, 31.04it/s] Loading 0: 19%|█▊ | 68/363 [00:07<00:08, 36.44it/s] Loading 0: 20%|██ | 74/363 [00:08<00:07, 36.80it/s] Loading 0: 22%|██▏ | 79/363 [00:08<00:07, 37.79it/s] Loading 0: 23%|██▎ | 85/363 [00:08<00:06, 42.95it/s] Loading 0: 25%|██▍ | 90/363 [00:08<00:06, 43.54it/s] Loading 0: 26%|██▌ | 95/363 [00:08<00:06, 44.28it/s] Loading 0: 28%|██▊ | 100/363 [00:08<00:05, 45.59it/s] Loading 0: 29%|██▉ | 105/363 [00:08<00:06, 37.97it/s] Loading 0: 31%|███ | 112/363 [00:08<00:05, 45.51it/s] Loading 0: 32%|███▏ | 117/363 [00:09<00:05, 45.38it/s] Loading 0: 34%|███▎ | 122/363 [00:09<00:07, 31.08it/s] Loading 0: 35%|███▍ | 127/363 [00:09<00:06, 34.85it/s] Loading 0: 36%|███▋ | 132/363 [00:09<00:07, 32.23it/s] Loading 0: 38%|███▊ | 139/363 [00:09<00:05, 39.17it/s] Loading 0: 40%|███▉ | 144/363 [00:09<00:05, 40.08it/s] Loading 0: 41%|████ | 149/363 [00:09<00:05, 39.85it/s] Loading 0: 43%|████▎ | 155/363 [00:10<00:05, 39.11it/s] Loading 0: 44%|████▍ | 160/363 [00:10<00:05, 38.93it/s] Loading 0: 46%|████▌ | 166/363 [00:10<00:04, 43.46it/s] Loading 0: 47%|████▋ | 171/363 [00:10<00:04, 42.64it/s] Loading 0: 48%|████▊ | 176/363 [00:10<00:04, 42.47it/s] Loading 0: 50%|████▉ | 181/363 [00:10<00:04, 44.10it/s] Loading 0: 51%|█████ | 186/363 [00:10<00:04, 36.04it/s] Loading 0: 53%|█████▎ | 191/363 [00:10<00:04, 39.20it/s] Loading 0: 54%|█████▍ | 196/363 [00:11<00:04, 38.66it/s] Loading 0: 56%|█████▌ | 202/363 [00:11<00:05, 29.51it/s] Loading 0: 57%|█████▋ | 206/363 [00:11<00:05, 30.33it/s] Loading 0: 58%|█████▊ | 211/363 [00:11<00:04, 33.80it/s] Loading 0: 59%|█████▉ | 215/363 [00:11<00:04, 34.39it/s] Loading 0: 61%|██████ | 220/363 [00:11<00:03, 38.12it/s] Loading 0: 62%|██████▏ | 225/363 [00:11<00:03, 39.96it/s] Loading 0: 63%|██████▎ | 230/363 [00:12<00:03, 40.19it/s] Loading 0: 65%|██████▍ | 235/363 [00:12<00:03, 42.55it/s] Loading 0: 66%|██████▌ | 240/363 [00:12<00:03, 36.03it/s] Loading 0: 68%|██████▊ | 247/363 [00:12<00:02, 42.53it/s] Loading 0: 69%|██████▉ | 252/363 [00:12<00:02, 41.25it/s] Loading 0: 71%|███████ | 257/363 [00:12<00:02, 41.30it/s] Loading 0: 72%|███████▏ | 263/363 [00:12<00:02, 40.03it/s] Loading 0: 74%|███████▍ | 268/363 [00:13<00:02, 39.39it/s] Loading 0: 75%|███████▌ | 274/363 [00:13<00:02, 43.56it/s] Loading 0: 77%|███████▋ | 279/363 [00:13<00:01, 43.65it/s] Loading 0: 78%|███████▊ | 284/363 [00:13<00:02, 29.30it/s] Loading 0: 80%|███████▉ | 289/363 [00:13<00:02, 33.02it/s] Loading 0: 81%|████████ | 293/363 [00:13<00:02, 32.31it/s] Loading 0: 82%|████████▏ | 298/363 [00:13<00:01, 36.18it/s] Loading 0: 83%|████████▎ | 303/363 [00:14<00:01, 32.30it/s] Loading 0: 85%|████████▌ | 310/363 [00:14<00:01, 40.11it/s] Loading 0: 87%|████████▋ | 315/363 [00:14<00:01, 40.82it/s] Loading 0: 88%|████████▊ | 320/363 [00:14<00:01, 39.86it/s] Loading 0: 90%|████████▉ | 325/363 [00:14<00:00, 41.47it/s] Loading 0: 91%|█████████ | 330/363 [00:14<00:00, 35.39it/s] Loading 0: 93%|█████████▎| 337/363 [00:14<00:00, 40.97it/s] Loading 0: 94%|█████████▍| 342/363 [00:15<00:00, 40.74it/s] Loading 0: 96%|█████████▌| 347/363 [00:15<00:00, 41.14it/s] Loading 0: 97%|█████████▋| 352/363 [00:15<00:00, 43.33it/s] Loading 0: 98%|█████████▊| 357/363 [00:15<00:00, 36.38it/s]
Job nitral-ai-captain-bmo-12b-v13-mkmlizer completed after 94.65s with status: succeeded
Stopping job with name nitral-ai-captain-bmo-12b-v13-mkmlizer
Pipeline stage MKMLizer completed in 95.22s
run pipeline stage %s
Running pipeline stage MKMLTemplater
Pipeline stage MKMLTemplater completed in 0.19s
run pipeline stage %s
Running pipeline stage MKMLDeployer
Creating inference service nitral-ai-captain-bmo-12b-v13
Waiting for inference service nitral-ai-captain-bmo-12b-v13 to be ready
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Inference service nitral-ai-captain-bmo-12b-v13 ready after 191.52574586868286s
Pipeline stage MKMLDeployer completed in 192.14s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.329514980316162s
Received healthy response to inference request in 2.341792345046997s
Received healthy response to inference request in 2.171627998352051s
Received healthy response to inference request in 1.7053894996643066s
Received healthy response to inference request in 1.8314151763916016s
5 requests
0 failed requests
5th percentile: 1.7305946350097656
10th percentile: 1.7557997703552246
20th percentile: 1.8062100410461426
30th percentile: 1.8994577407836915
40th percentile: 2.035542869567871
50th percentile: 2.171627998352051
60th percentile: 2.2347827911376954
70th percentile: 2.29793758392334
80th percentile: 2.331970453262329
90th percentile: 2.336881399154663
95th percentile: 2.33933687210083
99th percentile: 2.3413012504577635
mean time: 2.0759479999542236
Pipeline stage StressChecker completed in 11.90s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 2.77s
run pipeline stage %s
Running pipeline stage TriggerMKMLProfilingPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage TriggerMKMLProfilingPipeline completed in 2.79s
Shutdown handler de-registered
nitral-ai-captain-bmo-12b_v13 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeleter
Skipping teardown as no inference service was successfully deployed
Pipeline stage MKMLProfilerDeleter completed in 0.16s
run pipeline stage %s
Running pipeline stage MKMLProfilerTemplater
Pipeline stage MKMLProfilerTemplater completed in 0.13s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeployer
Creating inference service nitral-ai-captain-bmo-12b-v13-profiler
Waiting for inference service nitral-ai-captain-bmo-12b-v13-profiler to be ready
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 2723.20s
Shutdown handler de-registered
nitral-ai-captain-bmo-12b_v13 status is now inactive due to auto deactivation removed underperforming models