developer_uid: zonemercy
submission_id: zonemercy-lexical-nemov5_8701_v2
model_name: 0805v2-3
model_group: zonemercy/Lexical-Nemov5
status: torndown
timestamp: 2024-08-06T05:04:40+00:00
num_battles: 11177
num_wins: 5511
celo_rating: 1216.43
family_friendly_score: 0.0
submission_type: basic
model_repo: zonemercy/Lexical-Nemov5-1k1e5
model_architecture: MistralForCausalLM
reward_repo: ChaiML/gpt2_xl_pairwise_89m_step_347634
model_num_parameters: 12772070400.0
best_of: 4
max_input_tokens: 1024
max_output_tokens: 64
display_name: 0805v2-3
is_internal_developer: True
language_model: zonemercy/Lexical-Nemov5-1k1e5
model_size: 13B
ranking_group: single
us_pacific_date: 2024-08-05
win_ratio: 0.4930661179207301
generation_params: {'temperature': 0.9, 'top_p': 1.0, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n', '</s>', '###', 'Bot:', 'User:', 'You:', '<|im_end|>'], 'max_input_tokens': 1024, 'best_of': 4, 'max_output_tokens': 64, 'reward_max_token_input': 256}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
reward_formatter: {'bot_template': '{bot_name}: {message}\n', 'memory_template': '', 'prompt_template': '', 'response_template': '{bot_name}:', 'truncate_by_message': False, 'user_template': '{user_name}: {message}\n'}
Resubmit model
Running pipeline stage MKMLizer
Starting job with name zonemercy-lexical-nemov5-8701-v2-mkmlizer
Waiting for job on zonemercy-lexical-nemov5-8701-v2-mkmlizer to finish
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ _____ __ __ ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ /___/ ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ Version: 0.9.9 ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ https://mk1.ai ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ The license key for the current software has been verified as ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ belonging to: ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ Chai Research Corp. ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ Expiration: 2024-10-15 23:59:59 ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ║ ║
zonemercy-lexical-nemov5-8701-v2-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
Failed to get response for submission turboderp-cat-llama-3-7_8684_v21: ('http://turboderp-cat-llama-3-7-8684-v21-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'request timeout')
Failed to get response for submission turboderp-cat-llama-3-7_8684_v21: ('http://turboderp-cat-llama-3-7-8684-v21-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'request timeout')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'request timeout')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'EOF\n')
Failed to get response for submission zonemercy-lexical-nemov5_8701_v1: ('http://zonemercy-lexical-nemov5-8701-v1-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'dial tcp 127.0.0.1:8080: connect: connection refused\n')
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Downloaded to shared memory in 64.968s
zonemercy-lexical-nemov5-8701-v2-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmpdw2ke7c8, device:0
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Saving flywheel model at /dev/shm/model_cache
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
Failed to get response for submission zonemercy-lexical-nemo-_1518_v10: ('http://zonemercy-lexical-nemo-1518-v10-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
zonemercy-lexical-nemov5-8701-v2-mkmlizer: quantized model in 43.541s
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Processed model zonemercy/Lexical-Nemov5-1k1e5 in 108.510s
zonemercy-lexical-nemov5-8701-v2-mkmlizer: creating bucket guanaco-mkml-models
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
zonemercy-lexical-nemov5-8701-v2-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/zonemercy-lexical-nemov5-8701-v2
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/zonemercy-lexical-nemov5-8701-v2/config.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/zonemercy-lexical-nemov5-8701-v2/special_tokens_map.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/zonemercy-lexical-nemov5-8701-v2/tokenizer_config.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/zonemercy-lexical-nemov5-8701-v2/tokenizer.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/zonemercy-lexical-nemov5-8701-v2/flywheel_model.0.safetensors
zonemercy-lexical-nemov5-8701-v2-mkmlizer: loading reward model from ChaiML/gpt2_xl_pairwise_89m_step_347634
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Loading 0: 0%| | 0/363 [00:00<?, ?it/s] Loading 0: 1%| | 4/363 [00:00<00:09, 39.43it/s] Loading 0: 2%|▏ | 8/363 [00:00<00:14, 25.02it/s] Loading 0: 3%|▎ | 12/363 [00:00<00:13, 25.90it/s] Loading 0: 4%|▍ | 15/363 [00:00<00:15, 22.85it/s] Loading 0: 5%|▌ | 19/363 [00:00<00:12, 26.84it/s] Loading 0: 6%|▋ | 23/363 [00:01<00:18, 17.93it/s] Loading 0: 7%|▋ | 26/363 [00:01<00:19, 16.90it/s] Loading 0: 9%|▊ | 31/363 [00:01<00:14, 22.34it/s] Loading 0: 9%|▉ | 34/363 [00:01<00:14, 22.70it/s] Loading 0: 10%|█ | 37/363 [00:01<00:13, 24.14it/s] Loading 0: 11%|█ | 40/363 [00:01<00:12, 25.51it/s] Loading 0: 12%|█▏ | 43/363 [00:01<00:12, 24.86it/s] Loading 0: 13%|█▎ | 46/363 [00:01<00:12, 25.76it/s] Loading 0: 14%|█▍ | 50/363 [00:02<00:13, 22.60it/s] Loading 0: 15%|█▌ | 55/363 [00:02<00:11, 27.40it/s] Loading 0: 17%|█▋ | 60/363 [00:02<00:10, 28.71it/s] Loading 0: 18%|█▊ | 64/363 [00:02<00:16, 18.20it/s] Loading 0: 19%|█▉ | 69/363 [00:02<00:12, 22.93it/s] Loading 0: 20%|██ | 73/363 [00:03<00:12, 23.36it/s] Loading 0: 21%|██ | 77/363 [00:03<00:13, 21.97it/s] Loading 0: 23%|██▎ | 82/363 [00:03<00:10, 26.56it/s] Loading 0: 24%|██▎ | 86/363 [00:03<00:11, 23.71it/s] Loading 0: 25%|██▌ | 91/363 [00:03<00:09, 28.19it/s] Loading 0: 26%|██▌ | 95/363 [00:03<00:10, 25.22it/s] Loading 0: 28%|██▊ | 100/363 [00:04<00:08, 29.67it/s] Loading 0: 29%|██▊ | 104/363 [00:04<00:13, 18.76it/s] Loading 0: 30%|███ | 109/363 [00:04<00:10, 23.39it/s] Loading 0: 31%|███ | 113/363 [00:04<00:11, 22.20it/s] Loading 0: 33%|███▎ | 120/363 [00:04<00:08, 28.70it/s] Loading 0: 34%|███▍ | 124/363 [00:05<00:08, 27.71it/s] Loading 0: 36%|███▌ | 129/363 [00:05<00:07, 30.10it/s] Loading 0: 37%|███▋ | 133/363 [00:05<00:08, 28.72it/s] Loading 0: 38%|███▊ | 137/363 [00:05<00:07, 28.76it/s] Loading 0: 39%|███▉ | 142/363 [00:05<00:09, 24.05it/s] Loading 0: 40%|███▉ | 145/363 [00:05<00:09, 23.33it/s] Loading 0: 41%|████ | 149/363 [00:06<00:09, 22.63it/s] Loading 0: 43%|████▎ | 156/363 [00:06<00:07, 29.18it/s] Loading 0: 44%|████▍ | 160/363 [00:06<00:07, 28.35it/s] Loading 0: 45%|████▌ | 165/363 [00:06<00:06, 30.66it/s] Loading 0: 47%|████▋ | 169/363 [00:06<00:06, 29.41it/s] Loading 0: 48%|████▊ | 174/363 [00:06<00:06, 30.69it/s] Loading 0: 49%|████▉ | 178/363 [00:07<00:06, 27.56it/s] Loading 0: 50%|████▉ | 181/363 [00:07<00:06, 27.61it/s] Loading 0: 51%|█████ | 184/363 [00:07<00:09, 18.89it/s] Loading 0: 52%|█████▏ | 187/363 [00:07<00:08, 19.92it/s] Loading 0: 52%|█████▏ | 190/363 [00:07<00:07, 21.75it/s] Loading 0: 53%|█████▎ | 194/363 [00:07<00:08, 20.45it/s] Loading 0: 55%|█████▍ | 199/363 [00:08<00:06, 26.12it/s] Loading 0: 56%|█████▌ | 203/363 [00:08<00:06, 23.79it/s] Loading 0: 57%|█████▋ | 208/363 [00:08<00:05, 28.72it/s] Loading 0: 58%|█████▊ | 212/363 [00:08<00:05, 25.43it/s] Loading 0: 60%|█████▉ | 217/363 [00:08<00:04, 29.96it/s] Loading 0: 61%|██████ | 222/363 [00:08<00:04, 30.53it/s] Loading 0: 62%|██████▏ | 226/363 [00:09<00:06, 20.46it/s] Loading 0: 63%|██████▎ | 230/363 [00:09<00:06, 19.82it/s] Loading 0: 65%|██████▍ | 235/363 [00:09<00:05, 24.22it/s] Loading 0: 66%|██████▌ | 239/363 [00:09<00:05, 22.10it/s] Loading 0: 67%|██████▋ | 244/363 [00:09<00:04, 26.49it/s] Loading 0: 68%|██████▊ | 248/363 [00:10<00:04, 23.77it/s] Loading 0: 70%|██████▉ | 253/363 [00:10<00:03, 28.08it/s] Loading 0: 71%|███████ | 257/363 [00:10<00:04, 24.46it/s] Loading 0: 72%|███████▏ | 262/363 [00:10<00:03, 28.79it/s] Loading 0: 73%|███████▎ | 266/363 [00:10<00:05, 18.51it/s] Loading 0: 75%|███████▍ | 271/363 [00:11<00:04, 22.81it/s] Loading 0: 76%|███████▌ | 275/363 [00:11<00:04, 21.45it/s] Loading 0: 77%|███████▋ | 280/363 [00:11<00:03, 25.77it/s] Loading 0: 78%|███████▊ | 284/363 [00:11<00:03, 23.47it/s] Loading 0: 80%|███████▉ | 289/363 [00:11<00:02, 27.67it/s] Loading 0: 81%|████████ | 293/363 [00:11<00:02, 24.34it/s] Loading 0: 82%|████████▏ | 298/363 [00:12<00:02, 28.50it/s] Loading 0: 83%|████████▎ | 303/363 [00:12<00:02, 29.38it/s] Loading 0: 85%|████████▍ | 307/363 [00:12<00:02, 20.66it/s] Loading 0: 86%|████████▌ | 311/363 [00:12<00:02, 20.70it/s] Loading 0: 88%|████████▊ | 318/363 [00:12<00:01, 27.19it/s] Loading 0: 89%|████████▊ | 322/363 [00:13<00:01, 26.88it/s] Loading 0: 90%|█████████ | 327/363 [00:13<00:01, 29.40it/s] Loading 0: 91%|█████████ | 331/363 [00:13<00:01, 28.72it/s] Loading 0: 93%|█████████▎| 336/363 [00:13<00:00, 30.97it/s] Loading 0: 94%|█████████▎| 340/363 [00:13<00:00, 29.78it/s] Loading 0: 95%|█████████▍| 344/363 [00:20<00:09, 1.92it/s] Loading 0: 96%|█████████▌| 348/363 [00:21<00:05, 2.59it/s] Loading 0: 97%|█████████▋| 353/363 [00:21<00:02, 3.75it/s] Loading 0: 98%|█████████▊| 357/363 [00:21<00:01, 4.86it/s] /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
zonemercy-lexical-nemov5-8701-v2-mkmlizer: warnings.warn(
zonemercy-lexical-nemov5-8701-v2-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:785: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
zonemercy-lexical-nemov5-8701-v2-mkmlizer: warnings.warn(
zonemercy-lexical-nemov5-8701-v2-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:469: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
zonemercy-lexical-nemov5-8701-v2-mkmlizer: warnings.warn(
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Downloading shards: 0%| | 0/2 [00:00<?, ?it/s] Downloading shards: 50%|█████ | 1/2 [00:05<00:05, 5.31s/it] Downloading shards: 100%|██████████| 2/2 [00:07<00:00, 3.68s/it] Downloading shards: 100%|██████████| 2/2 [00:07<00:00, 3.93s/it]
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s] Loading checkpoint shards: 50%|█████ | 1/2 [00:00<00:00, 2.35it/s] Loading checkpoint shards: 100%|██████████| 2/2 [00:00<00:00, 3.85it/s] Loading checkpoint shards: 100%|██████████| 2/2 [00:00<00:00, 3.51it/s]
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Saving duration: 1.386s
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Processed model ChaiML/gpt2_xl_pairwise_89m_step_347634 in 13.102s
zonemercy-lexical-nemov5-8701-v2-mkmlizer: creating bucket guanaco-reward-models
zonemercy-lexical-nemov5-8701-v2-mkmlizer: Bucket 's3://guanaco-reward-models/' created
zonemercy-lexical-nemov5-8701-v2-mkmlizer: uploading /tmp/reward_cache to s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /tmp/reward_cache/special_tokens_map.json s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward/special_tokens_map.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /tmp/reward_cache/config.json s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward/config.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /tmp/reward_cache/tokenizer_config.json s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward/tokenizer_config.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /tmp/reward_cache/merges.txt s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward/merges.txt
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /tmp/reward_cache/vocab.json s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward/vocab.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /tmp/reward_cache/tokenizer.json s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward/tokenizer.json
zonemercy-lexical-nemov5-8701-v2-mkmlizer: cp /tmp/reward_cache/reward.tensors s3://guanaco-reward-models/zonemercy-lexical-nemov5-8701-v2_reward/reward.tensors
Job zonemercy-lexical-nemov5-8701-v2-mkmlizer completed after 156.72s with status: succeeded
Stopping job with name zonemercy-lexical-nemov5-8701-v2-mkmlizer
Pipeline stage MKMLizer completed in 157.70s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.11s
Running pipeline stage ISVCDeployer
Creating inference service zonemercy-lexical-nemov5-8701-v2
Waiting for inference service zonemercy-lexical-nemov5-8701-v2 to be ready
Failed to get response for submission turboderp-cat-llama-3-7_8684_v21: ('http://turboderp-cat-llama-3-7-8684-v21-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'request timeout')
Inference service zonemercy-lexical-nemov5-8701-v2 ready after 191.13740611076355s
Pipeline stage ISVCDeployer completed in 192.77s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.379539966583252s
Received healthy response to inference request in 1.5917694568634033s
Received healthy response to inference request in 1.6035828590393066s
Received healthy response to inference request in 1.5970110893249512s
Received healthy response to inference request in 1.5742688179016113s
5 requests
0 failed requests
5th percentile: 1.5777689456939696
10th percentile: 1.5812690734863282
20th percentile: 1.588269329071045
30th percentile: 1.592817783355713
40th percentile: 1.594914436340332
50th percentile: 1.5970110893249512
60th percentile: 1.5996397972106933
70th percentile: 1.6022685050964356
80th percentile: 1.758774280548096
90th percentile: 2.069157123565674
95th percentile: 2.224348545074463
99th percentile: 2.348501682281494
mean time: 1.7492344379425049
Pipeline stage StressChecker completed in 9.66s
zonemercy-lexical-nemov5_8701_v2 status is now deployed due to DeploymentManager action
zonemercy-lexical-nemov5_8701_v2 status is now inactive due to auto deactivation removed underperforming models
zonemercy-lexical-nemov5_8701_v2 status is now torndown due to DeploymentManager action