Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLizer
Starting job with name nitral-ai-captain-bmo-12b-v79-mkmlizer
Waiting for job on nitral-ai-captain-bmo-12b-v79-mkmlizer to finish
nitral-ai-captain-bmo-12b-v79-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ██████ ██████ █████ ████ ████ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ░░██████ ██████ ░░███ ███░ ░░███ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ░███░█████░███ ░███ ███ ░███ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ░███░░███ ░███ ░███████ ░███ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ░███ ░░░ ░███ ░███░░███ ░███ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ░███ ░███ ░███ ░░███ ░███ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ █████ █████ █████ ░░████ █████ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ░░░░░ ░░░░░ ░░░░░ ░░░░ ░░░░░ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ Version: 0.29.3 ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ Features: FLYWHEEL, CUDA ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ Copyright 2023-2025 MK ONE TECHNOLOGIES Inc. ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ https://mk1.ai ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ The license key for the current software has been verified as ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ belonging to: ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ Chai Research Corp. ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ Expiration: 2028-03-31 23:59:59 ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ║ ║
nitral-ai-captain-bmo-12b-v79-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
Failed to get response for submission function_patif_2025-06-06: ('http://chaiml-20250611-retune-u-1558-v3-predictor.tenant-chaiml-guanaco.k2.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'read tcp 127.0.0.1:35970->127.0.0.1:8080: read: connection reset by peer\n')
nitral-ai-captain-bmo-12b-v79-mkmlizer: Downloaded to shared memory in 300.303s
nitral-ai-captain-bmo-12b-v79-mkmlizer: Checking if Nitral-AI/Captain_BMO-12B already exists in ChaiML
nitral-ai-captain-bmo-12b-v79-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmp3aycmf4u, device:0
nitral-ai-captain-bmo-12b-v79-mkmlizer: Saving flywheel model at /dev/shm/model_cache
nitral-ai-captain-bmo-12b-v79-mkmlizer: quantized model in 30.465s
nitral-ai-captain-bmo-12b-v79-mkmlizer: Processed model Nitral-AI/Captain_BMO-12B in 330.850s
nitral-ai-captain-bmo-12b-v79-mkmlizer: creating bucket guanaco-mkml-models
nitral-ai-captain-bmo-12b-v79-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
nitral-ai-captain-bmo-12b-v79-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v79
nitral-ai-captain-bmo-12b-v79-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v79/tokenizer_config.json
nitral-ai-captain-bmo-12b-v79-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v79/tokenizer.json
nitral-ai-captain-bmo-12b-v79-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/nitral-ai-captain-bmo-12b-v79/flywheel_model.0.safetensors
nitral-ai-captain-bmo-12b-v79-mkmlizer:
Loading 0: 0%| | 0/363 [00:00<?, ?it/s]
Loading 0: 1%| | 2/363 [00:06<19:02, 3.16s/it]
Loading 0: 2%|▏ | 6/363 [00:06<05:02, 1.18it/s]
Loading 0: 4%|▍ | 14/363 [00:06<01:38, 3.54it/s]
Loading 0: 6%|▌ | 20/363 [00:06<00:59, 5.75it/s]
Loading 0: 7%|▋ | 24/363 [00:06<00:45, 7.52it/s]
Loading 0: 9%|▉ | 32/363 [00:06<00:26, 12.55it/s]
Loading 0: 10%|█ | 38/363 [00:07<00:20, 16.12it/s]
Loading 0: 12%|█▏ | 43/363 [00:07<00:19, 16.01it/s]
Loading 0: 14%|█▍ | 50/363 [00:07<00:14, 21.71it/s]
Loading 0: 15%|█▌ | 56/363 [00:07<00:12, 24.97it/s]
Loading 0: 17%|█▋ | 61/363 [00:07<00:10, 27.83it/s]
Loading 0: 19%|█▊ | 68/363 [00:07<00:08, 34.16it/s]
Loading 0: 20%|██ | 74/363 [00:08<00:08, 35.56it/s]
Loading 0: 22%|██▏ | 79/363 [00:08<00:07, 36.77it/s]
Loading 0: 24%|██▎ | 86/363 [00:08<00:06, 42.38it/s]
Loading 0: 25%|██▌ | 91/363 [00:08<00:06, 43.89it/s]
Loading 0: 26%|██▋ | 96/363 [00:08<00:07, 36.80it/s]
Loading 0: 28%|██▊ | 103/363 [00:08<00:06, 43.27it/s]
Loading 0: 30%|██▉ | 108/363 [00:08<00:05, 43.19it/s]
Loading 0: 31%|███ | 113/363 [00:09<00:05, 43.06it/s]
Loading 0: 33%|███▎ | 118/363 [00:09<00:05, 44.62it/s]
Loading 0: 34%|███▍ | 123/363 [00:09<00:09, 25.33it/s]
Loading 0: 36%|███▌ | 130/363 [00:09<00:07, 32.59it/s]
Loading 0: 37%|███▋ | 135/363 [00:09<00:06, 34.90it/s]
Loading 0: 39%|███▊ | 140/363 [00:09<00:06, 36.84it/s]
Loading 0: 40%|███▉ | 145/363 [00:09<00:05, 39.57it/s]
Loading 0: 41%|████▏ | 150/363 [00:10<00:06, 34.13it/s]
Loading 0: 43%|████▎ | 157/363 [00:10<00:04, 41.34it/s]
Loading 0: 45%|████▍ | 162/363 [00:10<00:04, 41.59it/s]
Loading 0: 46%|████▌ | 167/363 [00:10<00:04, 42.12it/s]
Loading 0: 47%|████▋ | 172/363 [00:10<00:04, 43.45it/s]
Loading 0: 49%|████▉ | 177/363 [00:10<00:05, 35.73it/s]
Loading 0: 51%|█████ | 184/363 [00:10<00:04, 42.45it/s]
Loading 0: 52%|█████▏ | 189/363 [00:11<00:04, 41.61it/s]
Loading 0: 53%|█████▎ | 194/363 [00:11<00:04, 41.53it/s]
Loading 0: 55%|█████▍ | 199/363 [00:11<00:03, 41.88it/s]
Loading 0: 56%|█████▌ | 204/363 [00:11<00:06, 24.97it/s]
Loading 0: 58%|█████▊ | 211/363 [00:11<00:04, 31.90it/s]
Loading 0: 60%|█████▉ | 216/363 [00:11<00:04, 33.89it/s]
Loading 0: 61%|██████ | 221/363 [00:12<00:04, 35.30it/s]
Loading 0: 62%|██████▏ | 226/363 [00:12<00:03, 37.81it/s]
Loading 0: 64%|██████▎ | 231/363 [00:12<00:03, 33.70it/s]
Loading 0: 66%|██████▌ | 238/363 [00:12<00:03, 40.07it/s]
Loading 0: 67%|██████▋ | 243/363 [00:12<00:02, 40.95it/s]
Loading 0: 68%|██████▊ | 248/363 [00:12<00:02, 41.75it/s]
Loading 0: 70%|██████▉ | 254/363 [00:12<00:02, 39.94it/s]
Loading 0: 71%|███████▏ | 259/363 [00:12<00:02, 39.30it/s]
Loading 0: 73%|███████▎ | 265/363 [00:13<00:02, 43.59it/s]
Loading 0: 74%|███████▍ | 270/363 [00:13<00:02, 43.46it/s]
Loading 0: 76%|███████▌ | 275/363 [00:13<00:02, 43.45it/s]
Loading 0: 77%|███████▋ | 280/363 [00:13<00:01, 44.76it/s]
Loading 0: 79%|███████▊ | 285/363 [00:13<00:02, 26.74it/s]
Loading 0: 80%|████████ | 292/363 [00:13<00:02, 33.93it/s]
Loading 0: 82%|████████▏ | 297/363 [00:14<00:01, 35.46it/s]
Loading 0: 83%|████████▎ | 302/363 [00:14<00:01, 35.94it/s]
Loading 0: 85%|████████▍ | 307/363 [00:14<00:01, 39.00it/s]
Loading 0: 86%|████████▌ | 312/363 [00:14<00:01, 33.95it/s]
Loading 0: 88%|████████▊ | 319/363 [00:14<00:01, 41.36it/s]
Loading 0: 89%|████████▉ | 324/363 [00:14<00:00, 42.25it/s]
Loading 0: 91%|█████████ | 329/363 [00:14<00:00, 42.59it/s]
Loading 0: 92%|█████████▏| 334/363 [00:14<00:00, 44.01it/s]
Loading 0: 93%|█████████▎| 339/363 [00:15<00:00, 37.20it/s]
Loading 0: 95%|█████████▌| 346/363 [00:15<00:00, 44.56it/s]
Loading 0: 97%|█████████▋| 351/363 [00:15<00:00, 43.83it/s]
Loading 0: 98%|█████████▊| 356/363 [00:15<00:00, 43.44it/s]
Loading 0: 100%|█████████▉| 362/363 [00:15<00:00, 41.79it/s]
Job nitral-ai-captain-bmo-12b-v79-mkmlizer completed after 351.02s with status: succeeded
Stopping job with name nitral-ai-captain-bmo-12b-v79-mkmlizer
Pipeline stage MKMLizer completed in 351.71s
run pipeline stage %s
Running pipeline stage MKMLTemplater
Pipeline stage MKMLTemplater completed in 0.20s
run pipeline stage %s
Running pipeline stage MKMLDeployer
Creating inference service nitral-ai-captain-bmo-12b-v79
Waiting for inference service nitral-ai-captain-bmo-12b-v79 to be ready
Inference service nitral-ai-captain-bmo-12b-v79 ready after 201.49549221992493s
Pipeline stage MKMLDeployer completed in 202.63s
run pipeline stage %s
Running pipeline stage StressChecker
HTTPConnectionPool(host='guanaco-submitter.guanaco-backend.k2.chaiverse.com', port=80): Read timed out. (read timeout=20)
Received unhealthy response to inference request!
Received healthy response to inference request in 2.470453977584839s
Received healthy response to inference request in 1.7840545177459717s
Received healthy response to inference request in 1.5816996097564697s
Received healthy response to inference request in 1.6662464141845703s
5 requests
1 failed requests
5th percentile: 1.5986089706420898
10th percentile: 1.61551833152771
20th percentile: 1.6493370532989502
30th percentile: 1.6898080348968505
40th percentile: 1.736931276321411
50th percentile: 1.7840545177459717
60th percentile: 2.0586143016815184
70th percentile: 2.333174085617065
80th percentile: 6.006681632995608
90th percentile: 13.07913694381714
95th percentile: 16.615364599227902
99th percentile: 19.444346723556517
mean time: 5.530809354782105
%s, retrying in %s seconds...
Received healthy response to inference request in 1.605454444885254s
Received healthy response to inference request in 1.8522303104400635s
Received healthy response to inference request in 1.865229845046997s
Received healthy response to inference request in 1.7350256443023682s
Received healthy response to inference request in 1.5411765575408936s
5 requests
0 failed requests
5th percentile: 1.5540321350097657
10th percentile: 1.5668877124786378
20th percentile: 1.5925988674163818
30th percentile: 1.6313686847686768
40th percentile: 1.6831971645355224
50th percentile: 1.7350256443023682
60th percentile: 1.7819075107574462
70th percentile: 1.8287893772125243
80th percentile: 1.8548302173614502
90th percentile: 1.8600300312042237
95th percentile: 1.8626299381256104
99th percentile: 1.8647098636627197
mean time: 1.7198233604431152
Pipeline stage StressChecker completed in 38.95s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.65s
run pipeline stage %s
Running pipeline stage TriggerMKMLProfilingPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage TriggerMKMLProfilingPipeline completed in 0.63s
Shutdown handler de-registered
nitral-ai-captain-bmo-12b_v79 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyScorer
Evaluating %s Family Friendly Score with %s threads
Pipeline stage OfflineFamilyFriendlyScorer completed in 6133.84s
Shutdown handler de-registered
nitral-ai-captain-bmo-12b_v79 status is now inactive due to auto deactivation removed underperforming models
nitral-ai-captain-bmo-12b_v79 status is now torndown due to DeploymentManager action