developer_uid: rirv938
submission_id: chaiml-mistral-24b-2048_90555_v4
model_name: chaiml-mistral-24b-2048_90555_v4
model_group: ChaiML/mistral_24b_2048_
status: inactive
timestamp: 2026-02-17T22:38:26+00:00
num_battles: 10888
num_wins: 5814
celo_rating: 1322.9
family_friendly_score: 0.0
family_friendly_standard_error: 0.0
submission_type: basic
model_repo: ChaiML/mistral_24b_2048_120_max_top_10k_new_rm_927_instruct_v2_merged
model_architecture: MistralForCausalLM
model_num_parameters: 24096691200.0
best_of: 8
max_input_tokens: 2048
max_output_tokens: 80
reward_model: default
display_name: chaiml-mistral-24b-2048_90555_v4
ineligible_reason: max_output_tokens!=64
is_internal_developer: True
language_model: ChaiML/mistral_24b_2048_120_max_top_10k_new_rm_927_instruct_v2_merged
model_size: 24B
ranking_group: single
us_pacific_date: 2026-02-17
win_ratio: 0.533982365907421
generation_params: {'temperature': 1.0, 'top_p': 0.95, 'min_p': 0.05, 'top_k': 60, 'presence_penalty': 0.1, 'frequency_penalty': 0.0, 'stopping_words': ['<|im_start|>', '###', '<|im_end|>', '</s>', 'You:'], 'max_input_tokens': 2048, 'best_of': 8, 'max_output_tokens': 80}
formatter: {'memory_template': '[SYSTEM_PROMPT]Respond as a high quality storyteller.[/SYSTEM_PROMPT][INST]', 'prompt_template': '', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '[/INST]{bot_name}:', 'truncate_by_message': False}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage VLLMUploader
Starting job with name chaiml-mistral-24b-2048-90555-v4-uploader
Waiting for job on chaiml-mistral-24b-2048-90555-v4-uploader to finish
Failed to get response for submission chaiml-mistral-24b-2048-_2678_v3: ('http://chaiml-mistral-24b-2048-2678-v3-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
chaiml-mistral-24b-2048-90555-v4-uploader: Using quantization_mode: fp8
chaiml-mistral-24b-2048-90555-v4-uploader: Checking if ChaiML/mistral_24b_2048_120_max_top_10k_new_rm_927_instruct_v2_merged-FP8 already exists in ChaiML
chaiml-mistral-24b-2048-90555-v4-uploader: Downloading snapshot of ChaiML/mistral_24b_2048_120_max_top_10k_new_rm_927_instruct_v2_merged...
chaiml-mistral-24b-2048-90555-v4-uploader: Downloaded in 43.101s
chaiml-mistral-24b-2048-90555-v4-uploader: Loading /tmp/model_input...
chaiml-mistral-24b-2048-90555-v4-uploader: The tokenizer you are loading from '/tmp/model_input' with an incorrect regex pattern: https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503/discussions/84#69121093e8b480e709447d5e. This will lead to incorrect tokenization. You should set the `fix_mistral_regex=True` flag when loading this tokenizer to fix this issue.
chaiml-mistral-24b-2048-90555-v4-uploader: `torch_dtype` is deprecated! Use `dtype` instead!
chaiml-mistral-24b-2048-90555-v4-uploader: Some parameters are on the meta device because they were offloaded to the cpu.
chaiml-mistral-24b-2048-90555-v4-uploader: Applying quantization...
chaiml-mistral-24b-2048-90555-v4-uploader: The tokenizer you are loading from '/tmp/model_input' with an incorrect regex pattern: https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Instruct-2503/discussions/84#69121093e8b480e709447d5e. This will lead to incorrect tokenization. You should set the `fix_mistral_regex=True` flag when loading this tokenizer to fix this issue.
chaiml-mistral-24b-2048-90555-v4-uploader: 2026-02-17T11:15:19.778002-0800 | reset | INFO - Compression lifecycle reset
chaiml-mistral-24b-2048-90555-v4-uploader: 2026-02-17T11:15:19.778929-0800 | from_modifiers | INFO - Creating recipe from modifiers
chaiml-mistral-24b-2048-90555-v4-uploader: 2026-02-17T11:15:19.868956-0800 | initialize | INFO - Compression lifecycle initialized for 1 modifiers
chaiml-mistral-24b-2048-90555-v4-uploader: 2026-02-17T11:15:19.869233-0800 | IndependentPipeline | INFO - Inferred `DataFreePipeline` for `QuantizationModifier`
chaiml-mistral-24b-2048-90555-v4-uploader: Some parameters are on the meta device because they were offloaded to the cpu.
chaiml-mistral-24b-2048-90555-v4-uploader: 2026-02-17T11:16:11.762002-0800 | finalize | INFO - Compression lifecycle finalized for 1 modifiers
chaiml-mistral-24b-2048-90555-v4-uploader: 2026-02-17T11:16:14.926272-0800 | post_process | WARNING - Optimized model is not saved. To save, please provide`output_dir` as input arg.Ex. `oneshot(..., output_dir=...)`
chaiml-mistral-24b-2048-90555-v4-uploader: Saving to /dev/shm/model_output...
chaiml-mistral-24b-2048-90555-v4-uploader: 2026-02-17T11:16:14.953511-0800 | get_model_compressor | INFO - skip_sparsity_compression_stats set to True. Skipping sparsity compression statistic calculations. No sparsity compressor will be applied.
chaiml-mistral-24b-2048-90555-v4-uploader: ---------- 2026-02-17 11:17:18 (0:00:00) ----------
chaiml-mistral-24b-2048-90555-v4-uploader: Files: hashed 7/14 (277.6K/27.6G) | pre-uploaded: 0/0 (0.0/27.6G) (+14 unsure) | committed: 0/14 (0.0/27.6G) | ignored: 0
chaiml-mistral-24b-2048-90555-v4-uploader: Workers: hashing: 7 | get upload mode: 5 | pre-uploading: 0 | committing: 0 | waiting: 114
chaiml-mistral-24b-2048-90555-v4-uploader: ---------------------------------------------------
chaiml-mistral-24b-2048-90555-v4-uploader:       
chaiml-mistral-24b-2048-90555-v4-uploader: ---------- 2026-02-17 11:18:18 (0:01:00) ----------
chaiml-mistral-24b-2048-90555-v4-uploader: Files: hashed 14/14 (27.6G/27.6G) | pre-uploaded: 7/7 (27.6G/27.6G) | committed: 0/14 (0.0/27.6G) | ignored: 0
chaiml-mistral-24b-2048-90555-v4-uploader: Workers: hashing: 0 | get upload mode: 0 | pre-uploading: 0 | committing: 1 | waiting: 125
chaiml-mistral-24b-2048-90555-v4-uploader: ---------------------------------------------------
chaiml-mistral-24b-2048-90555-v4-uploader: Processed model ChaiML/mistral_24b_2048_120_max_top_10k_new_rm_927_instruct_v2_merged in 247.830s
chaiml-mistral-24b-2048-90555-v4-uploader: creating bucket guanaco-vllm-models
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:56: SyntaxWarning: invalid escape sequence '\.'
chaiml-mistral-24b-2048-90555-v4-uploader: RE_S3_DATESTRING = re.compile('\.[0-9]*(?:[Z\\-\\+]*?)')
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:57: SyntaxWarning: invalid escape sequence '\s'
chaiml-mistral-24b-2048-90555-v4-uploader: RE_XML_NAMESPACE = re.compile(b'^(<?[^>]+?>\s*|\s*)(<\w+) xmlns=[\'"](https?://[^\'"]+)[\'"]', re.MULTILINE)
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:240: SyntaxWarning: invalid escape sequence '\.'
chaiml-mistral-24b-2048-90555-v4-uploader: invalid = re.search("([^a-z0-9\.-])", bucket, re.UNICODE)
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:244: SyntaxWarning: invalid escape sequence '\.'
chaiml-mistral-24b-2048-90555-v4-uploader: invalid = re.search("([^A-Za-z0-9\._-])", bucket, re.UNICODE)
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:255: SyntaxWarning: invalid escape sequence '\.'
chaiml-mistral-24b-2048-90555-v4-uploader: if re.search("-\.", bucket, re.UNICODE):
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:257: SyntaxWarning: invalid escape sequence '\.'
chaiml-mistral-24b-2048-90555-v4-uploader: if re.search("\.\.", bucket, re.UNICODE):
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/S3Uri.py:155: SyntaxWarning: invalid escape sequence '\w'
chaiml-mistral-24b-2048-90555-v4-uploader: _re = re.compile("^(\w+://)?(.*)", re.UNICODE)
chaiml-mistral-24b-2048-90555-v4-uploader: /usr/lib/python3/dist-packages/S3/FileLists.py:480: SyntaxWarning: invalid escape sequence '\*'
chaiml-mistral-24b-2048-90555-v4-uploader: wildcard_split_result = re.split("\*|\?", uri_str, maxsplit=1)
chaiml-mistral-24b-2048-90555-v4-uploader: Bucket 's3://guanaco-vllm-models/' created
chaiml-mistral-24b-2048-90555-v4-uploader: uploading /dev/shm/model_output to s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/chat_template.jinja s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/chat_template.jinja
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/recipe.yaml s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/recipe.yaml
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/config.json s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/config.json
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/generation_config.json s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/generation_config.json
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/model.safetensors.index.json s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/model.safetensors.index.json
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/special_tokens_map.json s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/special_tokens_map.json
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/tokenizer_config.json s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/tokenizer_config.json
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/tokenizer.json s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/tokenizer.json
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/model-00006-of-00006.safetensors s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/model-00006-of-00006.safetensors
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/model-00005-of-00006.safetensors s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/model-00005-of-00006.safetensors
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/model-00004-of-00006.safetensors s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/model-00004-of-00006.safetensors
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/model-00002-of-00006.safetensors s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/model-00002-of-00006.safetensors
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/model-00003-of-00006.safetensors s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/model-00003-of-00006.safetensors
chaiml-mistral-24b-2048-90555-v4-uploader: cp /dev/shm/model_output/model-00001-of-00006.safetensors s3://guanaco-vllm-models/chaiml-mistral-24b-2048-90555-v4/default/model-00001-of-00006.safetensors
Job chaiml-mistral-24b-2048-90555-v4-uploader completed after 413.1s with status: succeeded
Stopping job with name chaiml-mistral-24b-2048-90555-v4-uploader
Pipeline stage VLLMUploader completed in 413.97s
run pipeline stage %s
Running pipeline stage VLLMTemplater
Pipeline stage VLLMTemplater completed in 0.15s
run pipeline stage %s
Running pipeline stage VLLMDeployer
Creating inference service chaiml-mistral-24b-2048-90555-v4
Waiting for inference service chaiml-mistral-24b-2048-90555-v4 to be ready
HTTP Request: %s %s "%s %d %s"
Inference service chaiml-mistral-24b-2048-90555-v4 ready after 201.52392888069153s
Pipeline stage VLLMDeployer completed in 202.07s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 1.3410325050354004s
Received healthy response to inference request in 1.8276572227478027s
Received healthy response to inference request in 1.3470067977905273s
Received healthy response to inference request in 1.4935426712036133s
Received healthy response to inference request in 1.3112828731536865s
Received healthy response to inference request in 1.3627588748931885s
Received healthy response to inference request in 1.5080304145812988s
Received healthy response to inference request in 1.2184076309204102s
Received healthy response to inference request in 1.6364264488220215s
Received healthy response to inference request in 1.158949613571167s
Received healthy response to inference request in 1.2555484771728516s
Received healthy response to inference request in 2.0270469188690186s
Received healthy response to inference request in 1.6786534786224365s
Received healthy response to inference request in 1.3025777339935303s
Received healthy response to inference request in 1.3075032234191895s
Received healthy response to inference request in 1.6469306945800781s
Received healthy response to inference request in 1.4269213676452637s
Received healthy response to inference request in 1.6911938190460205s
Received healthy response to inference request in 1.1530849933624268s
Received healthy response to inference request in 1.7377820014953613s
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Received healthy response to inference request in 1.191014289855957s
Received healthy response to inference request in 1.235865592956543s
Received healthy response to inference request in 1.623711109161377s
Received healthy response to inference request in 1.35355544090271s
Received healthy response to inference request in 1.2329041957855225s
Received healthy response to inference request in 1.8874497413635254s
Received healthy response to inference request in 1.166905403137207s
Received healthy response to inference request in 1.1626920700073242s
Failed to get response for submission chaiml-mistral-24b-2048_15988_v1: ('http://chaiml-mistral-24b-2048-15988-v1-predictor.tenant-chaiml-guanaco.kchai-coreweave-us-east-04a.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', '')
Received healthy response to inference request in 1.461045265197754s
Received healthy response to inference request in 1.5270519256591797s
30 requests
0 failed requests
5th percentile: 1.1606337189674378
10th percentile: 1.1664840698242187
20th percentile: 1.2300048828125
30th percentile: 1.2884689569473267
40th percentile: 1.329132652282715
50th percentile: 1.3581571578979492
60th percentile: 1.4740442276000976
70th percentile: 1.5560496807098385
80th percentile: 1.6532752513885498
90th percentile: 1.7467695236206056
95th percentile: 1.86054310798645
99th percentile: 1.9865637373924256
mean time: 1.4424844264984131
Pipeline stage StressChecker completed in 47.29s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.71s
Shutdown handler de-registered
chaiml-mistral-24b-2048_90555_v4 status is now deployed due to DeploymentManager action
chaiml-mistral-24b-2048_90555_v4 status is now inactive due to auto deactivation removed underperforming models