developer_uid: acehao-chai
submission_id: chaiml-grpo-q235b-kimid_37540_v1
model_name: chaiml-grpo-q235b-kimid_37540_v1
model_group: ChaiML/grpo-q235b-kimid-
status: torndown
timestamp: 2026-02-28T17:27:01+00:00
num_battles: 2549
num_wins: 1399
celo_rating: 1331.49
family_friendly_score: 0.0
family_friendly_standard_error: 0.0
submission_type: basic
model_repo: ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-300
model_architecture: Qwen3MoeForCausalLM
model_num_parameters: 18790207488.0
best_of: 8
max_input_tokens: 2048
max_output_tokens: 80
reward_model: default
display_name: chaiml-grpo-q235b-kimid_37540_v1
ineligible_reason: max_output_tokens!=64
is_internal_developer: False
language_model: ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-300
model_size: 19B
ranking_group: single
us_pacific_date: 2026-02-25
win_ratio: 0.548842683405257
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['<|assistant|>', '<|im_end|>', '####', '<|user|>', '</think>', '</s>'], 'max_input_tokens': 2048, 'best_of': 8, 'max_output_tokens': 80}
formatter: {'memory_template': "<|im_start|>system\n{bot_name}'s persona: {memory}<|im_end|>\n", 'prompt_template': '', 'bot_template': '<|im_start|>assistant\n{bot_name}: {message}<|im_end|>\n', 'user_template': '<|im_start|>user\n{message}<|im_end|>\n', 'response_template': '<|im_start|>assistant\n{bot_name}:', 'truncate_by_message': True}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage VLLMUploader
Starting job with name chaiml-grpo-q235b-kimid-37540-v1-uploader
Waiting for job on chaiml-grpo-q235b-kimid-37540-v1-uploader to finish
chaiml-grpo-q235b-kimid-37540-v1-uploader: Using quantization_mode: w4a16
chaiml-grpo-q235b-kimid-37540-v1-uploader: Checking if ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-300-W4A16 already exists in ChaiML
chaiml-grpo-q235b-kimid-37540-v1-uploader: Downloading snapshot of ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-300...
HTTP Request: %s %s "%s %d %s"
chaiml-grpo-q235b-kimid-37540-v1-uploader: Downloaded in 170.319s
chaiml-grpo-q235b-kimid-37540-v1-uploader: Applying quantization...
chaiml-grpo-q235b-kimid-37540-v1-uploader: The tokenizer you are loading from '/tmp/model_input' with an incorrect regex pattern: https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503/discussions/84#69121093e8b480e709447d5e. This will lead to incorrect tokenization. You should set the `fix_mistral_regex=True` flag when loading this tokenizer to fix this issue.
chaiml-grpo-q235b-kimid-37540-v1-uploader: 2026-02-17 18:12:06 WARNING modeling_utils.py L4670: `torch_dtype` is deprecated! Use `dtype` instead!
chaiml-grpo-q235b-kimid-37540-v1-uploader: 2026-02-17 18:12:23 INFO base.py L366: using torch.bfloat16 for quantization tuning
chaiml-grpo-q235b-kimid-37540-v1-uploader: 2026-02-17 18:12:28 INFO base.py L1145: start to compute imatrix
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/local/lib/python3.12/dist-packages/torch/backends/cuda/__init__.py:131: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
chaiml-grpo-q235b-kimid-37540-v1-uploader: return torch._C._get_cublas_allow_tf32()
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:31.606000 7 torch/_dynamo/convert_frame.py:1358] [6/8] torch._dynamo hit config.recompile_limit (8)
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:31.606000 7 torch/_dynamo/convert_frame.py:1358] [6/8] function: 'forward' (/usr/local/lib/python3.12/dist-packages/transformers/models/qwen3_moe/modeling_qwen3_moe.py:208)
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:31.606000 7 torch/_dynamo/convert_frame.py:1358] [6/8] last reason: 6/7: self._modules['up_proj'].imatrix_cnt == 1340 # module.imatrix_cnt += input.shape[0] # auto_round/compressors/base.py:1179 in get_imatrix_hook (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.)
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:31.606000 7 torch/_dynamo/convert_frame.py:1358] [6/8] To log all recompilation reasons, use TORCH_LOGS="recompiles".
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:31.606000 7 torch/_dynamo/convert_frame.py:1358] [6/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:34.461000 7 torch/_dynamo/convert_frame.py:1358] [3/8] torch._dynamo hit config.recompile_limit (8)
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:34.461000 7 torch/_dynamo/convert_frame.py:1358] [3/8] function: 'forward' (/usr/local/lib/python3.12/dist-packages/transformers/models/qwen3_moe/modeling_qwen3_moe.py:305)
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:34.461000 7 torch/_dynamo/convert_frame.py:1358] [3/8] last reason: 3/7: self._modules['self_attn']._modules['k_proj'].imatrix_cnt == 56 # module.imatrix_cnt += input.shape[0] # auto_round/compressors/base.py:1179 in get_imatrix_hook (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.)
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:34.461000 7 torch/_dynamo/convert_frame.py:1358] [3/8] To log all recompilation reasons, use TORCH_LOGS="recompiles".
chaiml-grpo-q235b-kimid-37540-v1-uploader: W0217 18:13:34.461000 7 torch/_dynamo/convert_frame.py:1358] [3/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html
chaiml-grpo-q235b-kimid-37540-v1-uploader: 2026-02-17 18:13:52 WARNING gguf.py L297: please use more data via setting `nsamples` to improve accuracy as calibration activations contain 0
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
HTTP Request: %s %s "%s %d %s"
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
chaiml-grpo-q235b-kimid-37540-v1-uploader: Checking if ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-300-W4A16 already exists in ChaiML
chaiml-grpo-q235b-kimid-37540-v1-uploader: Creating repo ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-300-W4A16 and uploading /dev/shm/model_output to it
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------- 2026-02-17 19:27:11 (0:00:00) ----------
chaiml-grpo-q235b-kimid-37540-v1-uploader: Files: hashed 11/38 (26.1M/131.9G) | pre-uploaded: 0/1 (0.0/131.9G) (+27 unsure) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-37540-v1-uploader: Workers: hashing: 27 | get upload mode: 0 | pre-uploading: 1 | committing: 0 | waiting: 98
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------------------------------------------------
HTTP Request: %s %s "%s %d %s"
chaiml-grpo-q235b-kimid-37540-v1-uploader:       
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------- 2026-02-17 19:28:11 (0:01:00) ----------
chaiml-grpo-q235b-kimid-37540-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 2/28 (1.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-37540-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 26 | committing: 0 | waiting: 100
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-37540-v1-uploader:       
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------- 2026-02-17 19:29:11 (0:02:00) ----------
chaiml-grpo-q235b-kimid-37540-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 10/28 (41.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-37540-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 18 | committing: 0 | waiting: 108
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-37540-v1-uploader:       
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------- 2026-02-17 19:30:11 (0:03:00) ----------
chaiml-grpo-q235b-kimid-37540-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 18/28 (81.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-37540-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 10 | committing: 0 | waiting: 116
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-37540-v1-uploader:       
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------- 2026-02-17 19:31:11 (0:04:00) ----------
chaiml-grpo-q235b-kimid-37540-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 26/28 (121.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-37540-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 2 | committing: 0 | waiting: 124
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------------------------------------------------
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
chaiml-grpo-q235b-kimid-37540-v1-uploader:       
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------- 2026-02-17 19:32:11 (0:05:00) ----------
chaiml-grpo-q235b-kimid-37540-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 28/28 (131.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-37540-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 0 | committing: 1 | waiting: 125
chaiml-grpo-q235b-kimid-37540-v1-uploader: ---------------------------------------------------
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
chaiml-grpo-q235b-kimid-37540-v1-uploader: Processed model ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-300 in 4997.158s
chaiml-grpo-q235b-kimid-37540-v1-uploader: creating bucket guanaco-vllm-models
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:56: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-37540-v1-uploader: RE_S3_DATESTRING = re.compile('\.[0-9]*(?:[Z\\-\\+]*?)')
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:57: SyntaxWarning: invalid escape sequence '\s'
chaiml-grpo-q235b-kimid-37540-v1-uploader: RE_XML_NAMESPACE = re.compile(b'^(<?[^>]+?>\s*|\s*)(<\w+) xmlns=[\'"](https?://[^\'"]+)[\'"]', re.MULTILINE)
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:240: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-37540-v1-uploader: invalid = re.search("([^a-z0-9\.-])", bucket, re.UNICODE)
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:244: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-37540-v1-uploader: invalid = re.search("([^A-Za-z0-9\._-])", bucket, re.UNICODE)
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:255: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-37540-v1-uploader: if re.search("-\.", bucket, re.UNICODE):
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:257: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-37540-v1-uploader: if re.search("\.\.", bucket, re.UNICODE):
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/S3Uri.py:155: SyntaxWarning: invalid escape sequence '\w'
chaiml-grpo-q235b-kimid-37540-v1-uploader: _re = re.compile("^(\w+://)?(.*)", re.UNICODE)
chaiml-grpo-q235b-kimid-37540-v1-uploader: /usr/lib/python3/dist-packages/S3/FileLists.py:480: SyntaxWarning: invalid escape sequence '\*'
chaiml-grpo-q235b-kimid-37540-v1-uploader: wildcard_split_result = re.split("\*|\?", uri_str, maxsplit=1)
chaiml-grpo-q235b-kimid-37540-v1-uploader: Bucket 's3://guanaco-vllm-models/' created
chaiml-grpo-q235b-kimid-37540-v1-uploader: uploading /dev/shm/model_output to s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/added_tokens.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/added_tokens.json
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/generation_config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/generation_config.json
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/chat_template.jinja s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/chat_template.jinja
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/tokenizer_config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/tokenizer_config.json
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/config.json
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/quantization_config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/quantization_config.json
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/special_tokens_map.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/special_tokens_map.json
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/merges.txt s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/merges.txt
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/vocab.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/vocab.json
HTTP Request: %s %s "%s %d %s"
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00027-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00027-of-00027.safetensors
HTTP Request: %s %s "%s %d %s"
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00020-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00020-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00014-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00014-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00007-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00007-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00012-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00012-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00021-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00021-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00006-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00006-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00003-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00003-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00011-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00011-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00026-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00026-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00023-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00023-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00017-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00017-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00022-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00022-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00025-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00025-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00001-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00001-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00008-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00008-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00005-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00005-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00018-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00018-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00015-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00015-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00019-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00019-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00024-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00024-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00004-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00004-of-00027.safetensors
chaiml-grpo-q235b-kimid-37540-v1-uploader: cp /dev/shm/model_output/model-00002-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-37540-v1/default/model-00002-of-00027.safetensors
Job chaiml-grpo-q235b-kimid-37540-v1-uploader completed after 5284.42s with status: succeeded
Stopping job with name chaiml-grpo-q235b-kimid-37540-v1-uploader
Pipeline stage VLLMUploader completed in 5285.76s
run pipeline stage %s
Running pipeline stage VLLMTemplater
Pipeline stage VLLMTemplater completed in 0.18s
run pipeline stage %s
Running pipeline stage VLLMDeployer
Creating inference service chaiml-grpo-q235b-kimid-37540-v1
Waiting for inference service chaiml-grpo-q235b-kimid-37540-v1 to be ready
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Inference service chaiml-grpo-q235b-kimid-37540-v1 ready after 613.5019738674164s
Pipeline stage VLLMDeployer completed in 614.10s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.04915189743042s
Received healthy response to inference request in 2.324738025665283s
Received healthy response to inference request in 1.9862391948699951s
Received healthy response to inference request in 1.8771460056304932s
Received healthy response to inference request in 2.3673226833343506s
Received healthy response to inference request in 2.041288137435913s
Received healthy response to inference request in 2.227006435394287s
Received healthy response to inference request in 1.9734973907470703s
Received healthy response to inference request in 2.1080470085144043s
Received healthy response to inference request in 1.8823494911193848s
Received healthy response to inference request in 1.8706510066986084s
Received healthy response to inference request in 2.03473162651062s
Received healthy response to inference request in 2.2002651691436768s
Received healthy response to inference request in 1.9382665157318115s
Received healthy response to inference request in 1.9181437492370605s
Received healthy response to inference request in 1.9707798957824707s
Received healthy response to inference request in 2.14369797706604s
Received healthy response to inference request in 2.193751573562622s
Received healthy response to inference request in 1.9358491897583008s
Received healthy response to inference request in 2.371516704559326s
Received healthy response to inference request in 1.915137529373169s
Received healthy response to inference request in 2.421557903289795s
Received healthy response to inference request in 2.1952123641967773s
Received healthy response to inference request in 1.9466519355773926s
Received healthy response to inference request in 2.679840326309204s
Received healthy response to inference request in 2.0894622802734375s
Received healthy response to inference request in 2.063551425933838s
Received healthy response to inference request in 2.298368453979492s
Received healthy response to inference request in 2.0070624351501465s
Received healthy response to inference request in 2.0371828079223633s
30 requests
0 failed requests
5th percentile: 1.8794875741004944
10th percentile: 1.9118587255477906
20th percentile: 1.9377830505371094
30th percentile: 1.9726821422576903
40th percentile: 2.023663949966431
50th percentile: 2.0452200174331665
60th percentile: 2.096896171569824
70th percentile: 2.1941898107528686
80th percentile: 2.2412788391113283
90th percentile: 2.367742085456848
95th percentile: 2.3990393638610836
99th percentile: 2.6049384236335755
mean time: 2.102282238006592
Pipeline stage StressChecker completed in 68.62s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.95s
Shutdown handler de-registered
chaiml-grpo-q235b-kimid_37540_v1 status is now deployed due to DeploymentManager action
chaiml-grpo-q235b-kimid_37540_v1 status is now inactive due to system request
chaiml-grpo-q235b-kimid_37540_v1 status is now torndown due to DeploymentManager action