developer_uid: acehao-chai
submission_id: chaiml-grpo-q235b-kimid_17438_v1
model_name: chaiml-grpo-q235b-kimid_17438_v1
model_group: ChaiML/grpo-q235b-kimid-
status: torndown
timestamp: 2026-02-24T03:46:50+00:00
num_battles: 10810
num_wins: 5746
celo_rating: 1318.6
family_friendly_score: 0.0
family_friendly_standard_error: 0.0
submission_type: basic
model_repo: ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-600
model_architecture: Qwen3MoeForCausalLM
model_num_parameters: 18790207488.0
best_of: 8
max_input_tokens: 2048
max_output_tokens: 80
reward_model: default
display_name: chaiml-grpo-q235b-kimid_17438_v1
ineligible_reason: max_output_tokens!=64
is_internal_developer: False
language_model: ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-600
model_size: 19B
ranking_group: single
us_pacific_date: 2026-02-17
win_ratio: 0.5315448658649399
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['</think>', '</s>', '<|user|>', '<|assistant|>', '####', '<|im_end|>'], 'max_input_tokens': 2048, 'best_of': 8, 'max_output_tokens': 80}
formatter: {'memory_template': "<|im_start|>system\n{bot_name}'s persona: {memory}<|im_end|>\n", 'prompt_template': '', 'bot_template': '<|im_start|>assistant\n{bot_name}: {message}<|im_end|>\n', 'user_template': '<|im_start|>user\n{message}<|im_end|>\n', 'response_template': '<|im_start|>assistant\n{bot_name}:', 'truncate_by_message': True}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
run pipeline stage %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Running pipeline stage VLLMUploader
Starting job with name chaiml-grpo-q235b-kimid-17438-v1-uploader
Waiting for job on chaiml-grpo-q235b-kimid-17438-v1-uploader to finish
chaiml-grpo-q235b-kimid-17438-v1-uploader: Using quantization_mode: w4a16
chaiml-grpo-q235b-kimid-17438-v1-uploader: Checking if ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-600-W4A16 already exists in ChaiML
chaiml-grpo-q235b-kimid-17438-v1-uploader: Downloading snapshot of ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-600...
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Unable to record family friendly update due to error: Invalid JSON input: Expecting value: line 1 column 1 (char 0)
chaiml-grpo-q235b-kimid-17438-v1-uploader: Downloaded in 159.503s
chaiml-grpo-q235b-kimid-17438-v1-uploader: Applying quantization...
chaiml-grpo-q235b-kimid-17438-v1-uploader: The tokenizer you are loading from '/tmp/model_input' with an incorrect regex pattern: https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503/discussions/84#69121093e8b480e709447d5e. This will lead to incorrect tokenization. You should set the `fix_mistral_regex=True` flag when loading this tokenizer to fix this issue.
chaiml-grpo-q235b-kimid-17438-v1-uploader: 2026-02-17 15:44:43 WARNING modeling_utils.py L4670: `torch_dtype` is deprecated! Use `dtype` instead!
chaiml-grpo-q235b-kimid-17438-v1-uploader: 2026-02-17 15:44:54 INFO base.py L366: using torch.bfloat16 for quantization tuning
chaiml-grpo-q235b-kimid-17438-v1-uploader: 2026-02-17 15:45:00 INFO base.py L1145: start to compute imatrix
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/local/lib/python3.12/dist-packages/torch/backends/cuda/__init__.py:131: UserWarning: Please use the new API settings to control TF32 behavior, such as torch.backends.cudnn.conv.fp32_precision = 'tf32' or torch.backends.cuda.matmul.fp32_precision = 'ieee'. Old settings, e.g, torch.backends.cuda.matmul.allow_tf32 = True, torch.backends.cudnn.allow_tf32 = True, allowTF32CuDNN() and allowTF32CuBLAS() will be deprecated after Pytorch 2.9. Please see https://pytorch.org/docs/main/notes/cuda.html#tensorfloat-32-tf32-on-ampere-and-later-devices (Triggered internally at /pytorch/aten/src/ATen/Context.cpp:80.)
chaiml-grpo-q235b-kimid-17438-v1-uploader: return torch._C._get_cublas_allow_tf32()
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:04.498000 7 torch/_dynamo/convert_frame.py:1358] [6/8] torch._dynamo hit config.recompile_limit (8)
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:04.498000 7 torch/_dynamo/convert_frame.py:1358] [6/8] function: 'forward' (/usr/local/lib/python3.12/dist-packages/transformers/models/qwen3_moe/modeling_qwen3_moe.py:208)
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:04.498000 7 torch/_dynamo/convert_frame.py:1358] [6/8] last reason: 6/7: self._modules['up_proj'].imatrix_cnt == 1340 # module.imatrix_cnt += input.shape[0] # auto_round/compressors/base.py:1179 in get_imatrix_hook (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.)
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:04.498000 7 torch/_dynamo/convert_frame.py:1358] [6/8] To log all recompilation reasons, use TORCH_LOGS="recompiles".
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:04.498000 7 torch/_dynamo/convert_frame.py:1358] [6/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:07.398000 7 torch/_dynamo/convert_frame.py:1358] [3/8] torch._dynamo hit config.recompile_limit (8)
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:07.398000 7 torch/_dynamo/convert_frame.py:1358] [3/8] function: 'forward' (/usr/local/lib/python3.12/dist-packages/transformers/models/qwen3_moe/modeling_qwen3_moe.py:305)
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:07.398000 7 torch/_dynamo/convert_frame.py:1358] [3/8] last reason: 3/7: self._modules['self_attn']._modules['k_proj'].imatrix_cnt == 56 # module.imatrix_cnt += input.shape[0] # auto_round/compressors/base.py:1179 in get_imatrix_hook (HINT: torch.compile considers integer attributes of the nn.Module to be static. If you are observing recompilation, you might want to make this integer dynamic using torch._dynamo.config.allow_unspec_int_on_nn_module = True, or convert this integer into a tensor.)
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:07.398000 7 torch/_dynamo/convert_frame.py:1358] [3/8] To log all recompilation reasons, use TORCH_LOGS="recompiles".
chaiml-grpo-q235b-kimid-17438-v1-uploader: W0217 15:46:07.398000 7 torch/_dynamo/convert_frame.py:1358] [3/8] To diagnose recompilation issues, see https://pytorch.org/docs/main/torch.compiler_troubleshooting.html
chaiml-grpo-q235b-kimid-17438-v1-uploader: 2026-02-17 15:46:25 WARNING gguf.py L297: please use more data via setting `nsamples` to improve accuracy as calibration activations contain 0
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_54327_v6: ('http://chaiml-mistral-24b-2048-54327-v6-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
chaiml-grpo-q235b-kimid-17438-v1-uploader: Checking if ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-600-W4A16 already exists in ChaiML
chaiml-grpo-q235b-kimid-17438-v1-uploader: Creating repo ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-600-W4A16 and uploading /dev/shm/model_output to it
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------- 2026-02-17 16:59:44 (0:00:00) ----------
chaiml-grpo-q235b-kimid-17438-v1-uploader: Files: hashed 11/38 (26.1M/131.9G) | pre-uploaded: 0/0 (0.0/131.9G) (+30 unsure) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-17438-v1-uploader: Workers: hashing: 27 | get upload mode: 4 | pre-uploading: 0 | committing: 0 | waiting: 95
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-17438-v1-uploader:       
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------- 2026-02-17 17:00:44 (0:01:00) ----------
chaiml-grpo-q235b-kimid-17438-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 2/28 (1.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-17438-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 26 | committing: 0 | waiting: 100
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-17438-v1-uploader:       
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------- 2026-02-17 17:01:44 (0:02:00) ----------
chaiml-grpo-q235b-kimid-17438-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 10/28 (41.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-17438-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 18 | committing: 0 | waiting: 108
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-17438-v1-uploader:       
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------- 2026-02-17 17:02:44 (0:03:00) ----------
chaiml-grpo-q235b-kimid-17438-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 18/28 (81.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-17438-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 10 | committing: 0 | waiting: 116
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-17438-v1-uploader:       
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------- 2026-02-17 17:03:44 (0:04:00) ----------
chaiml-grpo-q235b-kimid-17438-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 19/28 (86.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-17438-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 9 | committing: 0 | waiting: 117
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-17438-v1-uploader:       
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------- 2026-02-17 17:04:44 (0:05:00) ----------
chaiml-grpo-q235b-kimid-17438-v1-uploader: Files: hashed 38/38 (131.9G/131.9G) | pre-uploaded: 28/28 (131.9G/131.9G) | committed: 0/38 (0.0/131.9G) | ignored: 0
chaiml-grpo-q235b-kimid-17438-v1-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 0 | committing: 1 | waiting: 125
chaiml-grpo-q235b-kimid-17438-v1-uploader: ---------------------------------------------------
chaiml-grpo-q235b-kimid-17438-v1-uploader: Processed model ChaiML/grpo-q235b-kimid-v5a-merged-nemo70b-chai-rm-step-600 in 5005.767s
chaiml-grpo-q235b-kimid-17438-v1-uploader: creating bucket guanaco-vllm-models
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:56: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-17438-v1-uploader: RE_S3_DATESTRING = re.compile('\.[0-9]*(?:[Z\\-\\+]*?)')
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:57: SyntaxWarning: invalid escape sequence '\s'
chaiml-grpo-q235b-kimid-17438-v1-uploader: RE_XML_NAMESPACE = re.compile(b'^(<?[^>]+?>\s*|\s*)(<\w+) xmlns=[\'"](https?://[^\'"]+)[\'"]', re.MULTILINE)
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:240: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-17438-v1-uploader: invalid = re.search("([^a-z0-9\.-])", bucket, re.UNICODE)
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:244: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-17438-v1-uploader: invalid = re.search("([^A-Za-z0-9\._-])", bucket, re.UNICODE)
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:255: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-17438-v1-uploader: if re.search("-\.", bucket, re.UNICODE):
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:257: SyntaxWarning: invalid escape sequence '\.'
chaiml-grpo-q235b-kimid-17438-v1-uploader: if re.search("\.\.", bucket, re.UNICODE):
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/S3Uri.py:155: SyntaxWarning: invalid escape sequence '\w'
chaiml-grpo-q235b-kimid-17438-v1-uploader: _re = re.compile("^(\w+://)?(.*)", re.UNICODE)
chaiml-grpo-q235b-kimid-17438-v1-uploader: /usr/lib/python3/dist-packages/S3/FileLists.py:480: SyntaxWarning: invalid escape sequence '\*'
chaiml-grpo-q235b-kimid-17438-v1-uploader: wildcard_split_result = re.split("\*|\?", uri_str, maxsplit=1)
chaiml-grpo-q235b-kimid-17438-v1-uploader: Bucket 's3://guanaco-vllm-models/' created
chaiml-grpo-q235b-kimid-17438-v1-uploader: uploading /dev/shm/model_output to s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/added_tokens.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/added_tokens.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/chat_template.jinja s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/chat_template.jinja
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/generation_config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/generation_config.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/tokenizer_config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/tokenizer_config.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/special_tokens_map.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/special_tokens_map.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/quantization_config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/quantization_config.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/config.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/config.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/merges.txt s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/merges.txt
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/vocab.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/vocab.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model.safetensors.index.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model.safetensors.index.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/tokenizer.json s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/tokenizer.json
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00027-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00027-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00025-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00025-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00010-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00010-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00024-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00024-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00017-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00017-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00007-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00007-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00009-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00009-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00008-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00008-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00005-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00005-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00020-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00020-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00022-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00022-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00014-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00014-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00013-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00013-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00004-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00004-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00023-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00023-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00016-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00016-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00015-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00015-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00021-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00021-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00012-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00012-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00018-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00018-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00003-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00003-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00019-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00019-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00006-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00006-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00001-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00001-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00011-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00011-of-00027.safetensors
chaiml-grpo-q235b-kimid-17438-v1-uploader: cp /dev/shm/model_output/model-00026-of-00027.safetensors s3://guanaco-vllm-models/chaiml-grpo-q235b-kimid-17438-v1/default/model-00026-of-00027.safetensors
Job chaiml-grpo-q235b-kimid-17438-v1-uploader completed after 5245.57s with status: succeeded
Stopping job with name chaiml-grpo-q235b-kimid-17438-v1-uploader
Pipeline stage VLLMUploader completed in 5247.86s
run pipeline stage %s
Running pipeline stage VLLMTemplater
Pipeline stage VLLMTemplater completed in 0.17s
run pipeline stage %s
Running pipeline stage VLLMDeployer
Creating inference service chaiml-grpo-q235b-kimid-17438-v1
Waiting for inference service chaiml-grpo-q235b-kimid-17438-v1 to be ready
HTTP Request: %s %s "%s %d %s"
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
HTTP Request: %s %s "%s %d %s"
Inference service chaiml-grpo-q235b-kimid-17438-v1 ready after 756.1131370067596s
Pipeline stage VLLMDeployer completed in 756.90s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.0115468502044678s
Received healthy response to inference request in 1.9304862022399902s
Received healthy response to inference request in 2.4916369915008545s
Received healthy response to inference request in 2.3781819343566895s
Received healthy response to inference request in 2.0975234508514404s
Received healthy response to inference request in 2.1001498699188232s
Received healthy response to inference request in 2.3452208042144775s
Received healthy response to inference request in 1.8862943649291992s
Received healthy response to inference request in 1.8691003322601318s
Received healthy response to inference request in 2.9875409603118896s
Received healthy response to inference request in 2.071094036102295s
Received healthy response to inference request in 2.233370542526245s
Received healthy response to inference request in 1.6592142581939697s
Received healthy response to inference request in 1.8291895389556885s
Received healthy response to inference request in 1.846653938293457s
Received healthy response to inference request in 1.9482526779174805s
Received healthy response to inference request in 1.996375322341919s
Received healthy response to inference request in 2.413037061691284s
Received healthy response to inference request in 2.527147054672241s
Received healthy response to inference request in 1.7941722869873047s
Received healthy response to inference request in 2.041445255279541s
Received healthy response to inference request in 2.182037353515625s
Received healthy response to inference request in 2.4027204513549805s
Received healthy response to inference request in 2.0862619876861572s
Received healthy response to inference request in 1.929225206375122s
Received healthy response to inference request in 2.2912063598632812s
Received healthy response to inference request in 2.1358015537261963s
Received healthy response to inference request in 1.7830753326416016s
Received healthy response to inference request in 1.9395122528076172s
Received healthy response to inference request in 1.9829013347625732s
30 requests
0 failed requests
5th percentile: 1.788068962097168
10th percentile: 1.8256878137588501
20th percentile: 1.8828555583953857
30th percentile: 1.936804437637329
40th percentile: 1.9909857273101808
50th percentile: 2.056269645690918
60th percentile: 2.0985740184783936
70th percentile: 2.197437310218811
80th percentile: 2.35181303024292
90th percentile: 2.4208970546722415
95th percentile: 2.511167526245117
99th percentile: 2.854026727676392
mean time: 2.106345852216085
Pipeline stage StressChecker completed in 66.56s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.81s
Shutdown handler de-registered
chaiml-grpo-q235b-kimid_17438_v1 status is now deployed due to DeploymentManager action
chaiml-grpo-q235b-kimid_17438_v1 status is now inactive due to auto deactivation removed underperforming models
admin requested tearing down of chaiml-grpo-q235b-kimid_17438_v1
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage VLLMDeleter
Checking if service chaiml-grpo-q235b-kimid-17438-v1 is running
Tearing down inference service chaiml-grpo-q235b-kimid-17438-v1
Service chaiml-grpo-q235b-kimid-17438-v1 has been torndown
Pipeline stage VLLMDeleter completed in 1.08s
run pipeline stage %s
Running pipeline stage VLLMModelDeleter
Cleaning model data from S3
Cleaning model data from model cache
%s, retrying in %s seconds...
Cleaning model data from S3
Cleaning model data from model cache
%s, retrying in %s seconds...
Cleaning model data from S3
Cleaning model data from model cache
clean up pipeline due to error=TeardownError("Got unexpected keyword argument 'request_checksum_calculation'")
Shutdown handler de-registered
chaiml-grpo-q235b-kimid_17438_v1 status is now torndown due to DeploymentManager action