developer_uid: zmeeks
submission_id: allura-org-mn-12b-rp-ink_v3
model_name: mn-v101001
model_group: allura-org/MN-12b-RP-Ink
status: torndown
timestamp: 2025-06-28T10:13:52+00:00
num_battles: 6181
num_wins: 2916
celo_rating: 1243.45
family_friendly_score: 0.5666
family_friendly_standard_error: 0.0070080587897077455
submission_type: basic
model_repo: allura-org/MN-12b-RP-Ink
model_architecture: MistralForCausalLM
model_num_parameters: 12772070400.0
best_of: 8
max_input_tokens: 1024
max_output_tokens: 64
reward_model: default
latencies: [{'batch_size': 1, 'throughput': 0.5993279402977062, 'latency_mean': 1.6684300005435944, 'latency_p50': 1.6798468828201294, 'latency_p90': 1.8415301322937012}, {'batch_size': 3, 'throughput': 1.0883730235122706, 'latency_mean': 2.7469677555561067, 'latency_p50': 2.7307939529418945, 'latency_p90': 3.0212652683258057}, {'batch_size': 5, 'throughput': 1.326776126409888, 'latency_mean': 3.752902706861496, 'latency_p50': 3.7233476638793945, 'latency_p90': 4.2098345994949335}, {'batch_size': 6, 'throughput': 1.3804998498082897, 'latency_mean': 4.316525510549545, 'latency_p50': 4.366760492324829, 'latency_p90': 4.808940052986145}, {'batch_size': 8, 'throughput': 1.4464386341649207, 'latency_mean': 5.490331867933273, 'latency_p50': 5.452619314193726, 'latency_p90': 6.1555483579635615}, {'batch_size': 10, 'throughput': 1.488326400913478, 'latency_mean': 6.660800248384476, 'latency_p50': 6.664451360702515, 'latency_p90': 7.531081676483154}]
gpu_counts: {'NVIDIA RTX A5000': 1}
display_name: mn-v101001
is_internal_developer: False
language_model: allura-org/MN-12b-RP-Ink
model_size: 13B
ranking_group: single
throughput_3p7s: 1.32
us_pacific_date: 2025-06-28
win_ratio: 0.47176832227794857
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.1, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 1024, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': '### Instruction:\n{memory}\n', 'prompt_template': '### Input:\n{prompt}\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '### Response:\n{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLizer
Starting job with name allura-org-mn-12b-rp-ink-v3-mkmlizer
Waiting for job on allura-org-mn-12b-rp-ink-v3-mkmlizer to finish
allura-org-mn-12b-rp-ink-v3-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ██████ ██████ █████ ████ ████ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ░░██████ ██████ ░░███ ███░ ░░███ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ░███░█████░███ ░███ ███ ░███ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ░███░░███ ░███ ░███████ ░███ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ░███ ░░░ ░███ ░███░░███ ░███ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ░███ ░███ ░███ ░░███ ░███ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ █████ █████ █████ ░░████ █████ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ░░░░░ ░░░░░ ░░░░░ ░░░░ ░░░░░ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ Version: 0.29.3 ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ Features: FLYWHEEL, CUDA ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ Copyright 2023-2025 MK ONE TECHNOLOGIES Inc. ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ https://mk1.ai ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ The license key for the current software has been verified as ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ belonging to: ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ Chai Research Corp. ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ Expiration: 2028-03-31 23:59:59 ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ║ ║
allura-org-mn-12b-rp-ink-v3-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
allura-org-mn-12b-rp-ink-v3-mkmlizer: Xet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`
allura-org-mn-12b-rp-ink-v3-mkmlizer: Xet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`
allura-org-mn-12b-rp-ink-v3-mkmlizer: Xet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`
allura-org-mn-12b-rp-ink-v3-mkmlizer: Xet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`
allura-org-mn-12b-rp-ink-v3-mkmlizer: Xet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`
allura-org-mn-12b-rp-ink-v3-mkmlizer: Xet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`
allura-org-mn-12b-rp-ink-v3-mkmlizer: Downloaded to shared memory in 48.561s
allura-org-mn-12b-rp-ink-v3-mkmlizer: Checking if allura-org/MN-12b-RP-Ink already exists in ChaiML
allura-org-mn-12b-rp-ink-v3-mkmlizer: Creating repo ChaiML/MN-12b-RP-Ink and uploading /tmp/tmpgqkrf658 to it
allura-org-mn-12b-rp-ink-v3-mkmlizer: 0%| | 0/6 [00:00<?, ?it/s] 17%|█▋ | 1/6 [00:07<00:38, 7.69s/it] 33%|███▎ | 2/6 [00:11<00:21, 5.47s/it] 50%|█████ | 3/6 [00:16<00:15, 5.15s/it] 67%|██████▋ | 4/6 [00:22<00:11, 5.56s/it] 83%|████████▎ | 5/6 [00:26<00:04, 4.87s/it] 100%|██████████| 6/6 [00:27<00:00, 3.63s/it] 100%|██████████| 6/6 [00:27<00:00, 4.57s/it]
allura-org-mn-12b-rp-ink-v3-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmpgqkrf658, device:0
allura-org-mn-12b-rp-ink-v3-mkmlizer: Saving flywheel model at /dev/shm/model_cache
allura-org-mn-12b-rp-ink-v3-mkmlizer: quantized model in 29.257s
allura-org-mn-12b-rp-ink-v3-mkmlizer: Processed model allura-org/MN-12b-RP-Ink in 132.338s
allura-org-mn-12b-rp-ink-v3-mkmlizer: creating bucket guanaco-mkml-models
allura-org-mn-12b-rp-ink-v3-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
allura-org-mn-12b-rp-ink-v3-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/allura-org-mn-12b-rp-ink-v3
allura-org-mn-12b-rp-ink-v3-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/allura-org-mn-12b-rp-ink-v3/config.json
allura-org-mn-12b-rp-ink-v3-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/allura-org-mn-12b-rp-ink-v3/special_tokens_map.json
allura-org-mn-12b-rp-ink-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/allura-org-mn-12b-rp-ink-v3/tokenizer_config.json
allura-org-mn-12b-rp-ink-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/allura-org-mn-12b-rp-ink-v3/tokenizer.json
allura-org-mn-12b-rp-ink-v3-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/allura-org-mn-12b-rp-ink-v3/flywheel_model.0.safetensors
allura-org-mn-12b-rp-ink-v3-mkmlizer: Loading 0: 0%| | 0/363 [00:00<?, ?it/s] Loading 0: 1%|▏ | 5/363 [00:00<00:10, 32.62it/s] Loading 0: 4%|▎ | 13/363 [00:00<00:06, 50.99it/s] Loading 0: 5%|▌ | 19/363 [00:00<00:07, 44.80it/s] Loading 0: 7%|▋ | 24/363 [00:00<00:07, 43.44it/s] Loading 0: 9%|▊ | 31/363 [00:00<00:06, 49.47it/s] Loading 0: 10%|█ | 37/363 [00:00<00:07, 45.68it/s] Loading 0: 12%|█▏ | 42/363 [00:00<00:07, 44.31it/s] Loading 0: 13%|█▎ | 49/363 [00:01<00:06, 49.29it/s] Loading 0: 15%|█▌ | 55/363 [00:01<00:06, 45.82it/s] Loading 0: 17%|█▋ | 61/363 [00:01<00:08, 34.29it/s] Loading 0: 18%|█▊ | 65/363 [00:01<00:08, 33.78it/s] Loading 0: 20%|█▉ | 72/363 [00:01<00:07, 40.20it/s] Loading 0: 21%|██ | 77/363 [00:01<00:06, 42.40it/s] Loading 0: 23%|██▎ | 82/363 [00:02<00:07, 36.33it/s] Loading 0: 25%|██▍ | 90/363 [00:02<00:06, 43.92it/s] Loading 0: 26%|██▋ | 96/363 [00:02<00:06, 42.12it/s] Loading 0: 28%|██▊ | 101/363 [00:02<00:06, 41.96it/s] Loading 0: 29%|██▉ | 107/363 [00:02<00:05, 46.13it/s] Loading 0: 31%|███ | 112/363 [00:02<00:05, 46.67it/s] Loading 0: 32%|███▏ | 117/363 [00:02<00:05, 44.07it/s] Loading 0: 34%|███▍ | 123/363 [00:02<00:05, 42.08it/s] Loading 0: 35%|███▌ | 128/363 [00:03<00:05, 41.59it/s] Loading 0: 37%|███▋ | 134/363 [00:03<00:05, 44.99it/s] Loading 0: 38%|███▊ | 139/363 [00:03<00:04, 44.82it/s] Loading 0: 40%|███▉ | 144/363 [00:03<00:07, 28.46it/s] Loading 0: 41%|████ | 149/363 [00:03<00:06, 30.87it/s] Loading 0: 43%|████▎ | 156/363 [00:03<00:05, 37.76it/s] Loading 0: 44%|████▍ | 161/363 [00:03<00:05, 39.41it/s] Loading 0: 46%|████▌ | 166/363 [00:04<00:04, 40.49it/s] Loading 0: 47%|████▋ | 172/363 [00:04<00:04, 39.50it/s] Loading 0: 49%|████▉ | 177/363 [00:04<00:04, 39.59it/s] Loading 0: 50%|█████ | 183/363 [00:04<00:04, 43.30it/s] Loading 0: 52%|█████▏ | 188/363 [00:04<00:04, 42.80it/s] Loading 0: 53%|█████▎ | 193/363 [00:04<00:03, 43.09it/s] Loading 0: 55%|█████▍ | 199/363 [00:04<00:03, 41.47it/s] Loading 0: 56%|█████▌ | 204/363 [00:04<00:03, 41.26it/s] Loading 0: 58%|█████▊ | 210/363 [00:05<00:03, 45.30it/s] Loading 0: 59%|█████▉ | 215/363 [00:05<00:03, 44.97it/s] Loading 0: 61%|██████ | 221/363 [00:05<00:02, 48.47it/s] Loading 0: 62%|██████▏ | 226/363 [00:05<00:04, 30.13it/s] Loading 0: 63%|██████▎ | 230/363 [00:05<00:04, 30.24it/s] Loading 0: 65%|██████▌ | 237/363 [00:05<00:03, 37.18it/s] Loading 0: 67%|██████▋ | 242/363 [00:05<00:03, 38.62it/s] Loading 0: 68%|██████▊ | 247/363 [00:06<00:02, 40.39it/s] Loading 0: 70%|██████▉ | 253/363 [00:06<00:02, 39.49it/s] Loading 0: 71%|███████ | 258/363 [00:06<00:02, 39.33it/s] Loading 0: 73%|███████▎ | 264/363 [00:06<00:02, 43.50it/s] Loading 0: 74%|███████▍ | 269/363 [00:06<00:02, 43.61it/s] Loading 0: 75%|███████▌ | 274/363 [00:06<00:02, 43.27it/s] Loading 0: 77%|███████▋ | 279/363 [00:06<00:01, 44.70it/s] Loading 0: 78%|███████▊ | 284/363 [00:07<00:02, 36.52it/s] Loading 0: 80%|████████ | 291/363 [00:07<00:01, 43.11it/s] Loading 0: 82%|████████▏ | 296/363 [00:07<00:01, 43.31it/s] Loading 0: 83%|████████▎ | 301/363 [00:07<00:01, 44.83it/s] Loading 0: 84%|████████▍ | 306/363 [00:07<00:02, 24.26it/s] Loading 0: 85%|████████▌ | 310/363 [00:07<00:02, 25.56it/s] Loading 0: 87%|████████▋ | 314/363 [00:08<00:01, 28.08it/s] Loading 0: 88%|████████▊ | 320/363 [00:08<00:01, 33.95it/s] Loading 0: 90%|████████▉ | 326/363 [00:08<00:01, 35.74it/s] Loading 0: 91%|█████████ | 331/363 [00:08<00:00, 37.40it/s] Loading 0: 93%|█████████▎| 338/363 [00:08<00:00, 43.40it/s] Loading 0: 95%|█████████▍| 344/363 [00:08<00:00, 41.73it/s] Loading 0: 96%|█████████▌| 349/363 [00:08<00:00, 41.82it/s] Loading 0: 98%|█████████▊| 356/363 [00:08<00:00, 46.87it/s] Loading 0: 99%|█████████▉| 361/363 [00:09<00:00, 47.49it/s]
Job allura-org-mn-12b-rp-ink-v3-mkmlizer completed after 156.62s with status: succeeded
Stopping job with name allura-org-mn-12b-rp-ink-v3-mkmlizer
Pipeline stage MKMLizer completed in 157.17s
run pipeline stage %s
Running pipeline stage MKMLTemplater
Pipeline stage MKMLTemplater completed in 0.17s
run pipeline stage %s
Running pipeline stage MKMLDeployer
Creating inference service allura-org-mn-12b-rp-ink-v3
Waiting for inference service allura-org-mn-12b-rp-ink-v3 to be ready
Inference service allura-org-mn-12b-rp-ink-v3 ready after 140.7993986606598s
Pipeline stage MKMLDeployer completed in 141.37s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.614915609359741s
Received healthy response to inference request in 1.4853136539459229s
Received healthy response to inference request in 1.8527734279632568s
Received healthy response to inference request in 2.0674421787261963s
Received healthy response to inference request in 1.8596763610839844s
5 requests
0 failed requests
5th percentile: 1.5588056087493896
10th percentile: 1.6322975635528565
20th percentile: 1.7792814731597901
30th percentile: 1.8541540145874023
40th percentile: 1.8569151878356933
50th percentile: 1.8596763610839844
60th percentile: 1.9427826881408692
70th percentile: 2.025889015197754
80th percentile: 2.1769368648529053
90th percentile: 2.3959262371063232
95th percentile: 2.5054209232330322
99th percentile: 2.5930166721343992
mean time: 1.9760242462158204
Pipeline stage StressChecker completed in 11.25s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.68s
run pipeline stage %s
Running pipeline stage TriggerMKMLProfilingPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage TriggerMKMLProfilingPipeline completed in 0.69s
Shutdown handler de-registered
allura-org-mn-12b-rp-ink_v3 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 3159.46s
Shutdown handler de-registered
allura-org-mn-12b-rp-ink_v3 status is now inactive due to auto deactivation removed underperforming models
allura-org-mn-12b-rp-ink_v3 status is now torndown due to DeploymentManager action