developer_uid: chai_evaluation_service
submission_id: qwen-qwen2-5-7b-instruct_v2
model_name: qwen-qwen2-5-7b-instruct_v2
model_group: Qwen/Qwen2.5-7B-Instruct
status: inactive
timestamp: 2026-02-08T01:37:45+00:00
num_battles: 11740
num_wins: 4336
celo_rating: 1213.7
family_friendly_score: 0.0
family_friendly_standard_error: 0.0
submission_type: basic
model_repo: Qwen/Qwen2.5-7B-Instruct
model_architecture: Qwen2ForCausalLM
model_num_parameters: 7615487488.0
best_of: 8
max_input_tokens: 2048
max_output_tokens: 64
reward_model: default
display_name: qwen-qwen2-5-7b-instruct_v2
is_internal_developer: True
language_model: Qwen/Qwen2.5-7B-Instruct
model_size: 8B
ranking_group: single
us_pacific_date: 2026-02-07
win_ratio: 0.36933560477001703
generation_params: {'temperature': 0.85, 'top_p': 0.9, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.2, 'frequency_penalty': 0.3, 'stopping_words': ['\n'], 'max_input_tokens': 2048, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': '<|im_start|>system\n{memory}<|im_end|>\n', 'prompt_template': '<|im_start|>user\n{prompt}<|im_end|>\n', 'bot_template': '<|im_start|>assistant\n{bot_name}: {message}<|im_end|>\n', 'user_template': '<|im_start|>user\n{user_name}: {message}<|im_end|>\n', 'response_template': '<|im_start|>assistant\n{bot_name}:', 'truncate_by_message': True}
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage VLLMUploader
Starting job with name qwen-qwen2-5-7b-instruct-v2-uploader
Waiting for job on qwen-qwen2-5-7b-instruct-v2-uploader to finish
HTTP Request: %s %s "%s %d %s"
qwen-qwen2-5-7b-instruct-v2-uploader: Using quantization_mode: none
qwen-qwen2-5-7b-instruct-v2-uploader: Downloading snapshot of Qwen/Qwen2.5-7B-Instruct...
HTTP Request: %s %s "%s %d %s"
qwen-qwen2-5-7b-instruct-v2-uploader: Fetching 14 files: 0%| | 0/14 [00:00<?, ?it/s] Fetching 14 files: 7%|▋ | 1/14 [00:00<00:03, 3.37it/s] Fetching 14 files: 43%|████▎ | 6/14 [00:00<00:00, 16.96it/s] Fetching 14 files: 64%|██████▍ | 9/14 [00:05<00:04, 1.21it/s] Fetching 14 files: 79%|███████▊ | 11/14 [00:08<00:02, 1.05it/s] Fetching 14 files: 100%|██████████| 14/14 [00:08<00:00, 1.65it/s]
qwen-qwen2-5-7b-instruct-v2-uploader: Downloaded in 8.619s
qwen-qwen2-5-7b-instruct-v2-uploader: Processed model Qwen/Qwen2.5-7B-Instruct in 14.581s
qwen-qwen2-5-7b-instruct-v2-uploader: creating bucket guanaco-vllm-models
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:56: SyntaxWarning: invalid escape sequence '\.'
qwen-qwen2-5-7b-instruct-v2-uploader: RE_S3_DATESTRING = re.compile('\.[0-9]*(?:[Z\\-\\+]*?)')
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/BaseUtils.py:57: SyntaxWarning: invalid escape sequence '\s'
qwen-qwen2-5-7b-instruct-v2-uploader: RE_XML_NAMESPACE = re.compile(b'^(<?[^>]+?>\s*|\s*)(<\w+) xmlns=[\'"](https?://[^\'"]+)[\'"]', re.MULTILINE)
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:240: SyntaxWarning: invalid escape sequence '\.'
qwen-qwen2-5-7b-instruct-v2-uploader: invalid = re.search("([^a-z0-9\.-])", bucket, re.UNICODE)
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:244: SyntaxWarning: invalid escape sequence '\.'
qwen-qwen2-5-7b-instruct-v2-uploader: invalid = re.search("([^A-Za-z0-9\._-])", bucket, re.UNICODE)
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:255: SyntaxWarning: invalid escape sequence '\.'
qwen-qwen2-5-7b-instruct-v2-uploader: if re.search("-\.", bucket, re.UNICODE):
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/Utils.py:257: SyntaxWarning: invalid escape sequence '\.'
qwen-qwen2-5-7b-instruct-v2-uploader: if re.search("\.\.", bucket, re.UNICODE):
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/S3Uri.py:155: SyntaxWarning: invalid escape sequence '\w'
qwen-qwen2-5-7b-instruct-v2-uploader: _re = re.compile("^(\w+://)?(.*)", re.UNICODE)
qwen-qwen2-5-7b-instruct-v2-uploader: /usr/lib/python3/dist-packages/S3/FileLists.py:480: SyntaxWarning: invalid escape sequence '\*'
qwen-qwen2-5-7b-instruct-v2-uploader: wildcard_split_result = re.split("\*|\?", uri_str, maxsplit=1)
qwen-qwen2-5-7b-instruct-v2-uploader: Bucket 's3://guanaco-vllm-models/' created
qwen-qwen2-5-7b-instruct-v2-uploader: uploading /dev/shm/model_output to s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/config.json s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/config.json
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/README.md s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/README.md
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/LICENSE s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/LICENSE
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/generation_config.json s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/generation_config.json
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/tokenizer_config.json s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/tokenizer_config.json
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/model.safetensors.index.json s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/model.safetensors.index.json
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/merges.txt s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/merges.txt
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/vocab.json s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/vocab.json
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/tokenizer.json s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/tokenizer.json
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/.gitattributes s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/.gitattributes
HTTP Request: %s %s "%s %d %s"
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/model-00003-of-00004.safetensors s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/model-00003-of-00004.safetensors
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/model-00004-of-00004.safetensors s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/model-00004-of-00004.safetensors
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/model-00002-of-00004.safetensors s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/model-00002-of-00004.safetensors
qwen-qwen2-5-7b-instruct-v2-uploader: cp /dev/shm/model_output/model-00001-of-00004.safetensors s3://guanaco-vllm-models/qwen-qwen2-5-7b-instruct-v2/model-00001-of-00004.safetensors
Job qwen-qwen2-5-7b-instruct-v2-uploader completed after 103.58s with status: succeeded
Stopping job with name qwen-qwen2-5-7b-instruct-v2-uploader
Pipeline stage VLLMUploader completed in 104.04s
run pipeline stage %s
Running pipeline stage VLLMTemplater
Pipeline stage VLLMTemplater completed in 0.15s
run pipeline stage %s
Running pipeline stage VLLMDeployer
Creating inference service qwen-qwen2-5-7b-instruct-v2
Waiting for inference service qwen-qwen2-5-7b-instruct-v2 to be ready
HTTP Request: %s %s "%s %d %s"
Inference service qwen-qwen2-5-7b-instruct-v2 ready after 170.73933935165405s
Pipeline stage VLLMDeployer completed in 171.25s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 0.8327479362487793s
Received healthy response to inference request in 0.9149084091186523s
Received healthy response to inference request in 1.2225854396820068s
Received healthy response to inference request in 1.331167221069336s
Received healthy response to inference request in 1.5951802730560303s
Received healthy response to inference request in 1.0504131317138672s
Received healthy response to inference request in 0.9547452926635742s
Received healthy response to inference request in 0.9724259376525879s
Received healthy response to inference request in 1.2369351387023926s
Received healthy response to inference request in 1.164923906326294s
Received healthy response to inference request in 1.0336439609527588s
Received healthy response to inference request in 1.3189451694488525s
Received healthy response to inference request in 1.5490844249725342s
Received healthy response to inference request in 1.0548460483551025s
Received healthy response to inference request in 0.9180729389190674s
Received healthy response to inference request in 0.8968226909637451s
Received healthy response to inference request in 1.049332857131958s
Received healthy response to inference request in 0.9328994750976562s
Received healthy response to inference request in 0.9260756969451904s
Received healthy response to inference request in 1.1324222087860107s
Received healthy response to inference request in 1.1422476768493652s
Received healthy response to inference request in 1.5189945697784424s
Received healthy response to inference request in 1.0360958576202393s
Received healthy response to inference request in 0.9001328945159912s
Received healthy response to inference request in 0.8825018405914307s
Received healthy response to inference request in 0.8292641639709473s
Received healthy response to inference request in 1.1289851665496826s
Received healthy response to inference request in 0.8142032623291016s
Received healthy response to inference request in 1.1316266059875488s
Received healthy response to inference request in 0.9993109703063965s
30 requests
0 failed requests
5th percentile: 0.8308318614959717
10th percentile: 0.8775264501571656
20th percentile: 0.9119533061981201
30th percentile: 0.9308523416519165
40th percentile: 0.9885569572448731
50th percentile: 1.0427143573760986
60th percentile: 1.0845016956329345
70th percentile: 1.135369849205017
80th percentile: 1.225455379486084
90th percentile: 1.3499499559402468
95th percentile: 1.5355439901351928
99th percentile: 1.5818124771118165
mean time: 1.082384705543518
Pipeline stage StressChecker completed in 35.70s
run pipeline stage %s
Running pipeline stage OfflineFamilyFriendlyTriggerPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
triggered trigger_guanaco_pipeline args=%s
Pipeline stage OfflineFamilyFriendlyTriggerPipeline completed in 0.60s
Shutdown handler de-registered
qwen-qwen2-5-7b-instruct_v2 status is now deployed due to DeploymentManager action
qwen-qwen2-5-7b-instruct_v2 status is now inactive due to auto deactivation removed underperforming models