developer_uid: rirv938
submission_id: nitral-ai-captain-bmo-12b_v46
model_name: nitral-ai-captain-bmo-12b_v46
model_group: Nitral-AI/Captain_BMO-12
status: torndown
timestamp: 2025-04-30T21:01:25+00:00
num_battles: 9215
num_wins: 4595
celo_rating: 1287.44
family_friendly_score: 0.5536
family_friendly_standard_error: 0.00703032061857779
submission_type: basic
model_repo: Nitral-AI/Captain_BMO-12B
model_architecture: MistralForCausalLM
model_num_parameters: 12772070400.0
best_of: 8
max_input_tokens: 768
max_output_tokens: 64
reward_model: default
latencies: [{'batch_size': 1, 'throughput': 0.6423633376021006, 'latency_mean': 1.5566903603076936, 'latency_p50': 1.5613194704055786, 'latency_p90': 1.7107466459274292}, {'batch_size': 3, 'throughput': 1.2239771719530879, 'latency_mean': 2.443779841661453, 'latency_p50': 2.4494848251342773, 'latency_p90': 2.745772695541382}, {'batch_size': 5, 'throughput': 1.5021836636051011, 'latency_mean': 3.311479901075363, 'latency_p50': 3.308358907699585, 'latency_p90': 3.629803729057312}, {'batch_size': 6, 'throughput': 1.5814704551281968, 'latency_mean': 3.7677639734745028, 'latency_p50': 3.798625349998474, 'latency_p90': 4.145830678939819}, {'batch_size': 8, 'throughput': 1.6751625847819103, 'latency_mean': 4.744619374275207, 'latency_p50': 4.76392388343811, 'latency_p90': 5.305978751182556}, {'batch_size': 10, 'throughput': 1.7185065446262, 'latency_mean': 5.772483135461807, 'latency_p50': 5.810183167457581, 'latency_p90': 6.537247443199157}]
gpu_counts: {'NVIDIA RTX A5000': 1}
display_name: nitral-ai-captain-bmo-12b_v46
is_internal_developer: True
language_model: Nitral-AI/Captain_BMO-12B
model_size: 13B
ranking_group: single
throughput_3p7s: 1.58
us_pacific_date: 2025-04-30
win_ratio: 0.49864351600651113
generation_params: {'temperature': 0.9, 'top_p': 1.0, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['Bot:', 'You:', 'User:', '<|im_end|>', '\n', '</s>', '####', '<|eot_id|>'], 'max_input_tokens': 768, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': '', 'prompt_template': '', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLizer
Starting job with name nitral-ai-captain-bmo-12b-v46-mkmlizer
Waiting for job on nitral-ai-captain-bmo-12b-v46-mkmlizer to finish
nitral-ai-captain-bmo-12b-v46-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ _____ __ __ ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ /___/ ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ Version: 0.12.8 ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ https://mk1.ai ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ The license key for the current software has been verified as ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ belonging to: ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ Chai Research Corp. ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ Expiration: 2028-03-31 23:59:59 ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v46-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
nitral-ai-captain-bmo-12b-v46-mkmlizer: Downloaded to shared memory in 43.544s
nitral-ai-captain-bmo-12b-v46-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmppnve23fz, device:0
nitral-ai-captain-bmo-12b-v46-mkmlizer: Saving flywheel model at /dev/shm/model_cache
nitral-ai-captain-bmo-12b-v46-mkmlizer: quantized model in 36.040s
nitral-ai-captain-bmo-12b-v46-mkmlizer: Processed model Nitral-AI/Captain_BMO-12B in 79.585s
nitral-ai-captain-bmo-12b-v46-mkmlizer: creating bucket guanaco-mkml-models
nitral-ai-captain-bmo-12b-v46-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
nitral-ai-captain-bmo-12b-v46-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v46
nitral-ai-captain-bmo-12b-v46-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v46/config.json
nitral-ai-captain-bmo-12b-v46-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v46/special_tokens_map.json
nitral-ai-captain-bmo-12b-v46-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v46/tokenizer_config.json
nitral-ai-captain-bmo-12b-v46-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v46/tokenizer.json
nitral-ai-captain-bmo-12b-v46-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v46/flywheel_model.0.safetensors
nitral-ai-captain-bmo-12b-v46-mkmlizer: Loading 0: 0%| | 0/363 [00:00<?, ?it/s] Loading 0: 1%| | 2/363 [00:06<18:29, 3.07s/it] Loading 0: 2%|▏ | 6/363 [00:06<04:55, 1.21it/s] Loading 0: 4%|▎ | 13/363 [00:06<01:44, 3.33it/s] Loading 0: 5%|▍ | 18/363 [00:06<01:05, 5.30it/s] Loading 0: 7%|▋ | 24/363 [00:06<00:41, 8.08it/s] Loading 0: 9%|▉ | 32/363 [00:06<00:25, 13.22it/s] Loading 0: 10%|█ | 38/363 [00:06<00:19, 16.96it/s] Loading 0: 12%|█▏ | 43/363 [00:07<00:19, 16.83it/s] Loading 0: 14%|█▍ | 50/363 [00:07<00:13, 22.54it/s] Loading 0: 15%|█▌ | 56/363 [00:07<00:11, 26.13it/s] Loading 0: 17%|█▋ | 61/363 [00:07<00:10, 28.94it/s] Loading 0: 19%|█▊ | 68/363 [00:07<00:08, 35.37it/s] Loading 0: 20%|██ | 73/363 [00:07<00:07, 37.67it/s] Loading 0: 21%|██▏ | 78/363 [00:08<00:08, 33.93it/s] Loading 0: 24%|██▎ | 86/363 [00:08<00:06, 42.56it/s] Loading 0: 25%|██▌ | 92/363 [00:08<00:06, 42.70it/s] Loading 0: 27%|██▋ | 97/363 [00:08<00:06, 42.86it/s] Loading 0: 29%|██▊ | 104/363 [00:08<00:05, 47.35it/s] Loading 0: 30%|███ | 110/363 [00:08<00:05, 45.87it/s] Loading 0: 32%|███▏ | 115/363 [00:08<00:05, 44.75it/s] Loading 0: 33%|███▎ | 121/363 [00:09<00:06, 34.97it/s] Loading 0: 34%|███▍ | 125/363 [00:09<00:06, 35.13it/s] Loading 0: 36%|███▌ | 131/363 [00:09<00:05, 39.32it/s] Loading 0: 38%|███▊ | 137/363 [00:09<00:05, 39.68it/s] Loading 0: 39%|███▉ | 142/363 [00:09<00:05, 39.61it/s] Loading 0: 41%|████ | 149/363 [00:09<00:04, 44.77it/s] Loading 0: 43%|████▎ | 155/363 [00:09<00:04, 43.73it/s] Loading 0: 44%|████▍ | 160/363 [00:09<00:04, 42.82it/s] Loading 0: 46%|████▌ | 167/363 [00:10<00:04, 47.35it/s] Loading 0: 48%|████▊ | 173/363 [00:10<00:04, 44.49it/s] Loading 0: 49%|████▉ | 178/363 [00:10<00:04, 43.52it/s] Loading 0: 51%|█████ | 185/363 [00:10<00:03, 47.98it/s] Loading 0: 53%|█████▎ | 191/363 [00:10<00:03, 45.03it/s] Loading 0: 54%|█████▍ | 196/363 [00:10<00:03, 42.49it/s] Loading 0: 56%|█████▌ | 202/363 [00:10<00:04, 33.47it/s] Loading 0: 57%|█████▋ | 206/363 [00:11<00:04, 33.17it/s] Loading 0: 58%|█████▊ | 211/363 [00:11<00:04, 36.69it/s] Loading 0: 59%|█████▉ | 215/363 [00:11<00:04, 36.12it/s] Loading 0: 61%|██████ | 221/363 [00:11<00:03, 40.82it/s] Loading 0: 62%|██████▏ | 226/363 [00:11<00:03, 42.91it/s] Loading 0: 64%|██████▎ | 231/363 [00:11<00:03, 37.14it/s] Loading 0: 66%|██████▌ | 239/363 [00:11<00:02, 45.69it/s] Loading 0: 67%|██████▋ | 245/363 [00:11<00:02, 44.89it/s] Loading 0: 69%|██████▉ | 250/363 [00:12<00:02, 43.12it/s] Loading 0: 71%|███████ | 256/363 [00:12<00:02, 47.07it/s] Loading 0: 72%|███████▏ | 261/363 [00:12<00:02, 46.92it/s] Loading 0: 73%|███████▎ | 266/363 [00:12<00:02, 47.14it/s] Loading 0: 75%|███████▍ | 272/363 [00:12<00:02, 44.17it/s] Loading 0: 76%|███████▋ | 277/363 [00:12<00:02, 41.53it/s] Loading 0: 78%|███████▊ | 283/363 [00:12<00:02, 31.95it/s] Loading 0: 79%|███████▉ | 287/363 [00:13<00:02, 32.07it/s] Loading 0: 81%|████████ | 293/363 [00:13<00:01, 36.43it/s] Loading 0: 82%|████████▏ | 297/363 [00:13<00:01, 37.18it/s] Loading 0: 83%|████████▎ | 302/363 [00:13<00:01, 39.02it/s] Loading 0: 85%|████████▍ | 307/363 [00:13<00:01, 41.34it/s] Loading 0: 86%|████████▌ | 312/363 [00:13<00:01, 35.73it/s] Loading 0: 88%|████████▊ | 319/363 [00:13<00:01, 42.39it/s] Loading 0: 89%|████████▉ | 324/363 [00:13<00:00, 41.61it/s] Loading 0: 91%|█████████ | 329/363 [00:14<00:00, 41.54it/s] Loading 0: 92%|█████████▏| 334/363 [00:14<00:00, 43.20it/s] Loading 0: 93%|█████████▎| 339/363 [00:14<00:00, 35.87it/s] Loading 0: 95%|█████████▌| 346/363 [00:14<00:00, 43.22it/s] Loading 0: 97%|█████████▋| 351/363 [00:14<00:00, 43.50it/s] Loading 0: 98%|█████████▊| 356/363 [00:14<00:00, 41.98it/s] Loading 0: 99%|█████████▉| 361/363 [00:14<00:00, 43.65it/s]
Job nitral-ai-captain-bmo-12b-v46-mkmlizer completed after 104.86s with status: succeeded
Stopping job with name nitral-ai-captain-bmo-12b-v46-mkmlizer
Pipeline stage MKMLizer completed in 105.60s
run pipeline stage %s
Running pipeline stage MKMLTemplater
Pipeline stage MKMLTemplater completed in 0.16s
run pipeline stage %s
Running pipeline stage MKMLDeployer
Creating inference service nitral-ai-captain-bmo-12b-v46
Waiting for inference service nitral-ai-captain-bmo-12b-v46 to be ready
Inference service nitral-ai-captain-bmo-12b-v46 ready after 150.52695393562317s
Pipeline stage MKMLDeployer completed in 150.94s
run pipeline stage %s
Running pipeline stage StressChecker
HTTPConnectionPool(host='guanaco-submitter.guanaco-backend.k2.chaiverse.com', port=80): Read timed out. (read timeout=20)
Received unhealthy response to inference request!
Received healthy response to inference request in 2.099019765853882s
Received healthy response to inference request in 1.5857698917388916s
Received healthy response to inference request in 1.747715950012207s
Received healthy response to inference request in 1.7683544158935547s
5 requests
1 failed requests
5th percentile: 1.6181591033935547
10th percentile: 1.6505483150482179
20th percentile: 1.715326738357544
30th percentile: 1.7518436431884765
40th percentile: 1.7600990295410157
50th percentile: 1.7683544158935547
60th percentile: 1.9006205558776856
70th percentile: 2.032886695861816
80th percentile: 5.7066487312316925
90th percentile: 12.921906661987308
95th percentile: 16.52953562736511
99th percentile: 19.41563879966736
mean time: 5.467604923248291
%s, retrying in %s seconds...
Received healthy response to inference request in 1.8531484603881836s
Received healthy response to inference request in 1.5272808074951172s
Received healthy response to inference request in 1.5359601974487305s
Received healthy response to inference request in 1.5104379653930664s
Received healthy response to inference request in 1.7982022762298584s
5 requests
0 failed requests
5th percentile: 1.5138065338134765
10th percentile: 1.5171751022338866
20th percentile: 1.523912239074707
30th percentile: 1.5290166854858398
40th percentile: 1.5324884414672852
50th percentile: 1.5359601974487305
60th percentile: 1.6408570289611817
70th percentile: 1.7457538604736327
80th percentile: 1.8091915130615235
90th percentile: 1.8311699867248534
95th percentile: 1.8421592235565185
99th percentile: 1.8509506130218505
mean time: 1.6450059413909912
Pipeline stage StressChecker completed in 38.25s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.68s
run pipeline stage %s
Running pipeline stage TriggerMKMLProfilingPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage TriggerMKMLProfilingPipeline completed in 0.74s
Shutdown handler de-registered
nitral-ai-captain-bmo-12b_v46 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
%s, retrying in %s seconds...
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3593.21s
Shutdown handler de-registered
nitral-ai-captain-bmo-12b_v46 status is now inactive due to auto deactivation removed underperforming models
nitral-ai-captain-bmo-12b_v46 status is now torndown due to DeploymentManager action