submission_id: zeuslabs-l3-aethora-15b-v2_v16
developer_uid: HiImElyStay
alignment_samples: 0
best_of: 4
celo_rating: 1160.73
display_name: zeuslabs-l3-aethora-15b-v2_v15
formatter: {'memory_template': '<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{memory}<|eot_id|>', 'prompt_template': '<|start_header_id|>system<|end_header_id|>\n\n{prompt}<|eot_id|>', 'bot_template': '<|start_header_id|>assistant<|end_header_id|>\n\n{bot_name}: {message}<|eot_id|>', 'user_template': '<|start_header_id|>user<|end_header_id|>\n\n{user_name}: {message}<|eot_id|>', 'response_template': '<|start_header_id|>assistant<|end_header_id|>\n\n{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 0.95, 'top_p': 1.0, 'min_p': 0.05, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 512, 'best_of': 4, 'max_output_tokens': 64, 'reward_max_token_input': 256}
is_internal_developer: False
language_model: ZeusLabs/L3-Aethora-15B-V2
max_input_tokens: 512
max_output_tokens: 64
model_architecture: LlamaForCausalLM
model_group: ZeusLabs/L3-Aethora-15B-
model_name: zeuslabs-l3-aethora-15b-v2_v15
model_num_parameters: 15009845248.0
model_repo: ZeusLabs/L3-Aethora-15B-V2
model_size: 15B
num_battles: 10920
num_wins: 4817
propriety_score: 0.7415730337078652
propriety_total_count: 979.0
ranking_group: single
reward_formatter: {'bot_template': '{bot_name}: {message}\n', 'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'response_template': '{bot_name}:', 'truncate_by_message': False, 'user_template': '{user_name}: {message}\n'}
reward_repo: ChaiML/reward_gpt2_medium_preference_24m_e2
status: torndown
submission_type: basic
timestamp: 2024-07-26T14:13:23+00:00
us_pacific_date: 2024-07-26
win_ratio: 0.4411172161172161
Download Preference Data
Resubmit model
Running pipeline stage MKMLizer
Starting job with name zeuslabs-l3-aethora-15b-v2-v16-mkmlizer
Waiting for job on zeuslabs-l3-aethora-15b-v2-v16-mkmlizer to finish
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ _____ __ __ ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ /___/ ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ Version: 0.9.7 ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ https://mk1.ai ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ The license key for the current software has been verified as ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ belonging to: ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ Chai Research Corp. ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ Expiration: 2024-10-15 23:59:59 ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ║ ║
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Downloaded to shared memory in 45.010s
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmp_odw3vc0, device:0
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Saving flywheel model at /dev/shm/model_cache
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: quantized model in 39.938s
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Processed model ZeusLabs/L3-Aethora-15B-V2 in 84.948s
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: creating bucket guanaco-mkml-models
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/zeuslabs-l3-aethora-15b-v2-v16
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/zeuslabs-l3-aethora-15b-v2-v16/special_tokens_map.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/zeuslabs-l3-aethora-15b-v2-v16/config.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/zeuslabs-l3-aethora-15b-v2-v16/tokenizer_config.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/zeuslabs-l3-aethora-15b-v2-v16/tokenizer.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /dev/shm/model_cache/flywheel_model.1.safetensors s3://guanaco-mkml-models/zeuslabs-l3-aethora-15b-v2-v16/flywheel_model.1.safetensors
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/zeuslabs-l3-aethora-15b-v2-v16/flywheel_model.0.safetensors
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: loading reward model from ChaiML/reward_gpt2_medium_preference_24m_e2
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Loading 0: 0%| | 0/579 [00:00<?, ?it/s] Loading 0: 1%| | 5/579 [00:00<00:15, 36.58it/s] Loading 0: 2%|▏ | 13/579 [00:00<00:09, 58.41it/s] Loading 0: 3%|▎ | 20/579 [00:00<00:10, 52.78it/s] Loading 0: 4%|▍ | 26/579 [00:00<00:10, 52.06it/s] Loading 0: 6%|▌ | 32/579 [00:00<00:12, 44.49it/s] Loading 0: 7%|▋ | 40/579 [00:00<00:10, 53.81it/s] Loading 0: 8%|▊ | 46/579 [00:00<00:10, 49.83it/s] Loading 0: 9%|▉ | 52/579 [00:01<00:10, 51.15it/s] Loading 0: 10%|█ | 58/579 [00:01<00:09, 53.27it/s] Loading 0: 11%|█ | 64/579 [00:01<00:10, 48.40it/s] Loading 0: 12%|█▏ | 70/579 [00:01<00:10, 49.74it/s] Loading 0: 13%|█▎ | 76/579 [00:01<00:09, 51.92it/s] Loading 0: 14%|█▍ | 82/579 [00:01<00:10, 46.70it/s] Loading 0: 15%|█▌ | 87/579 [00:01<00:15, 31.67it/s] Loading 0: 16%|█▌ | 94/579 [00:02<00:12, 38.66it/s] Loading 0: 17%|█▋ | 100/579 [00:02<00:11, 39.93it/s] Loading 0: 18%|█▊ | 105/579 [00:02<00:11, 41.52it/s] Loading 0: 19%|█▉ | 112/579 [00:02<00:09, 47.65it/s] Loading 0: 20%|██ | 118/579 [00:02<00:10, 45.46it/s] Loading 0: 21%|██ | 123/579 [00:02<00:10, 44.88it/s] Loading 0: 22%|██▏ | 130/579 [00:02<00:08, 50.39it/s] Loading 0: 23%|██▎ | 136/579 [00:02<00:09, 47.11it/s] Loading 0: 24%|██▍ | 141/579 [00:03<00:09, 45.84it/s] Loading 0: 25%|██▌ | 147/579 [00:03<00:08, 49.36it/s] Loading 0: 26%|██▋ | 153/579 [00:03<00:08, 51.07it/s] Loading 0: 27%|██▋ | 159/579 [00:03<00:09, 45.65it/s] Loading 0: 29%|██▊ | 166/579 [00:03<00:08, 49.85it/s] Loading 0: 30%|██▉ | 172/579 [00:03<00:08, 47.06it/s] Loading 0: 31%|███ | 179/579 [00:03<00:07, 50.59it/s] Loading 0: 32%|███▏ | 185/579 [00:03<00:07, 51.90it/s] Loading 0: 33%|███▎ | 191/579 [00:04<00:11, 33.23it/s] Loading 0: 34%|███▍ | 196/579 [00:04<00:11, 34.81it/s] Loading 0: 35%|███▍ | 202/579 [00:04<00:09, 39.67it/s] Loading 0: 36%|███▌ | 208/579 [00:04<00:09, 40.33it/s] Loading 0: 37%|███▋ | 213/579 [00:04<00:08, 40.84it/s] Loading 0: 38%|███▊ | 220/579 [00:04<00:07, 46.04it/s] Loading 0: 39%|███▉ | 226/579 [00:04<00:07, 44.56it/s] Loading 0: 40%|███▉ | 231/579 [00:05<00:07, 44.78it/s] Loading 0: 41%|████ | 237/579 [00:05<00:07, 48.02it/s] Loading 0: 42%|████▏ | 242/579 [00:05<00:07, 47.62it/s] Loading 0: 43%|████▎ | 247/579 [00:05<00:06, 48.21it/s] Loading 0: 44%|████▎ | 253/579 [00:05<00:07, 43.30it/s] Loading 0: 45%|████▍ | 258/579 [00:05<00:07, 43.25it/s] Loading 0: 46%|████▌ | 265/579 [00:05<00:06, 48.94it/s] Loading 0: 47%|████▋ | 271/579 [00:05<00:06, 46.73it/s] Loading 0: 48%|████▊ | 276/579 [00:06<00:06, 46.66it/s] Loading 0: 49%|████▊ | 282/579 [00:06<00:06, 43.09it/s] Loading 0: 50%|████▉ | 287/579 [00:06<00:09, 31.93it/s] Loading 0: 51%|█████ | 293/579 [00:06<00:08, 35.47it/s] Loading 0: 52%|█████▏ | 301/579 [00:06<00:06, 43.70it/s] Loading 0: 53%|█████▎ | 307/579 [00:06<00:06, 42.80it/s] Loading 0: 54%|█████▍ | 312/579 [00:06<00:06, 42.18it/s] Loading 0: 55%|█████▌ | 319/579 [00:07<00:05, 46.94it/s] Loading 0: 56%|█████▌ | 325/579 [00:07<00:05, 44.07it/s] Loading 0: 57%|█████▋ | 330/579 [00:07<00:05, 43.68it/s] Loading 0: 58%|█████▊ | 336/579 [00:07<00:05, 47.10it/s] Loading 0: 59%|█████▉ | 341/579 [00:07<00:05, 47.26it/s] Loading 0: 60%|█████▉ | 346/579 [00:07<00:04, 47.33it/s] Loading 0: 61%|██████ | 351/579 [00:07<00:04, 47.75it/s] Loading 0: 61%|██████▏ | 356/579 [00:07<00:05, 39.18it/s] Loading 0: 63%|██████▎ | 363/579 [00:08<00:04, 45.38it/s] Loading 0: 64%|██████▎ | 368/579 [00:08<00:04, 45.56it/s] Loading 0: 64%|██████▍ | 373/579 [00:08<00:04, 46.07it/s] Loading 0: 65%|██████▌ | 378/579 [00:08<00:04, 46.97it/s] Loading 0: 66%|██████▌ | 383/579 [00:08<00:05, 38.06it/s] Loading 0: 68%|██████▊ | 392/579 [00:08<00:04, 44.08it/s] Loading 0: 69%|██████▊ | 397/579 [00:09<00:05, 31.75it/s] Loading 0: 69%|██████▉ | 401/579 [00:09<00:05, 32.55it/s] Loading 0: 70%|███████ | 408/579 [00:09<00:04, 39.86it/s] Loading 0: 71%|███████▏ | 413/579 [00:09<00:04, 41.00it/s] Loading 0: 72%|███████▏ | 418/579 [00:09<00:03, 42.21it/s] Loading 0: 73%|███████▎ | 423/579 [00:09<00:03, 44.05it/s] Loading 0: 74%|███████▍ | 428/579 [00:09<00:03, 37.80it/s] Loading 0: 75%|███████▌ | 436/579 [00:09<00:03, 46.35it/s] Loading 0: 76%|███████▋ | 442/579 [00:10<00:03, 43.76it/s] Loading 0: 77%|███████▋ | 447/579 [00:10<00:02, 44.21it/s] Loading 0: 78%|███████▊ | 454/579 [00:10<00:02, 48.96it/s] Loading 0: 79%|███████▉ | 460/579 [00:10<00:02, 47.03it/s] Loading 0: 80%|████████ | 465/579 [00:10<00:02, 46.15it/s] Loading 0: 82%|████████▏ | 472/579 [00:10<00:02, 50.36it/s] Loading 0: 83%|████████▎ | 478/579 [00:10<00:02, 47.32it/s] Loading 0: 83%|████████▎ | 483/579 [00:10<00:02, 45.64it/s] Loading 0: 85%|████████▍ | 490/579 [00:11<00:01, 51.28it/s] Loading 0: 86%|████████▌ | 496/579 [00:11<00:02, 33.35it/s] Loading 0: 87%|████████▋ | 501/579 [00:11<00:02, 35.17it/s] Loading 0: 88%|████████▊ | 507/579 [00:11<00:01, 39.76it/s] Loading 0: 88%|████████▊ | 512/579 [00:11<00:01, 41.64it/s] Loading 0: 89%|████████▉ | 517/579 [00:11<00:01, 42.70it/s] Loading 0: 89%|████████▉ | 517/579 [00:25<00:01, 42.70it/s] Loading 0: 89%|████████▉ | 518/579 [00:25<01:04, 1.06s/it] Loading 0: 90%|█████████ | 523/579 [00:25<00:39, 1.43it/s] Loading 0: 91%|█████████ | 528/579 [00:26<00:24, 2.05it/s] Loading 0: 92%|█████████▏| 535/579 [00:26<00:13, 3.33it/s] Loading 0: 93%|█████████▎| 541/579 [00:26<00:07, 4.77it/s] Loading 0: 94%|█████████▍| 546/579 [00:26<00:05, 6.37it/s] Loading 0: 96%|█████████▌| 553/579 [00:26<00:02, 9.41it/s] Loading 0: 97%|█████████▋| 559/579 [00:26<00:01, 12.34it/s] Loading 0: 97%|█████████▋| 564/579 [00:26<00:00, 15.28it/s] Loading 0: 99%|█████████▊| 571/579 [00:27<00:00, 20.89it/s] Loading 0: 100%|█████████▉| 577/579 [00:27<00:00, 24.38it/s] /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: warnings.warn(
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:785: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: warnings.warn(
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:469: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: warnings.warn(
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Saving duration: 0.331s
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Processed model ChaiML/reward_gpt2_medium_preference_24m_e2 in 6.726s
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: creating bucket guanaco-reward-models
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: Bucket 's3://guanaco-reward-models/' created
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: uploading /tmp/reward_cache to s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /tmp/reward_cache/config.json s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward/config.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /tmp/reward_cache/tokenizer_config.json s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward/tokenizer_config.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /tmp/reward_cache/special_tokens_map.json s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward/special_tokens_map.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /tmp/reward_cache/merges.txt s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward/merges.txt
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /tmp/reward_cache/vocab.json s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward/vocab.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /tmp/reward_cache/tokenizer.json s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward/tokenizer.json
zeuslabs-l3-aethora-15b-v2-v16-mkmlizer: cp /tmp/reward_cache/reward.tensors s3://guanaco-reward-models/zeuslabs-l3-aethora-15b-v2-v16_reward/reward.tensors
Job zeuslabs-l3-aethora-15b-v2-v16-mkmlizer completed after 126.94s with status: succeeded
Stopping job with name zeuslabs-l3-aethora-15b-v2-v16-mkmlizer
Pipeline stage MKMLizer completed in 128.16s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.10s
Running pipeline stage ISVCDeployer
Creating inference service zeuslabs-l3-aethora-15b-v2-v16
Waiting for inference service zeuslabs-l3-aethora-15b-v2-v16 to be ready
Inference service zeuslabs-l3-aethora-15b-v2-v16 ready after 70.65508341789246s
Pipeline stage ISVCDeployer completed in 72.37s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.8429150581359863s
Received healthy response to inference request in 1.7466599941253662s
Received healthy response to inference request in 1.8504722118377686s
Received healthy response to inference request in 1.2331023216247559s
Received healthy response to inference request in 1.3745365142822266s
5 requests
0 failed requests
5th percentile: 1.26138916015625
10th percentile: 1.2896759986877442
20th percentile: 1.3462496757507325
30th percentile: 1.4489612102508544
40th percentile: 1.5978106021881104
50th percentile: 1.7466599941253662
60th percentile: 1.7881848812103271
70th percentile: 1.829709768295288
80th percentile: 2.048960781097412
90th percentile: 2.4459379196166995
95th percentile: 2.6444264888763427
99th percentile: 2.8032173442840578
mean time: 1.8095372200012207
Pipeline stage StressChecker completed in 9.72s
zeuslabs-l3-aethora-15b-v2_v16 status is now deployed due to DeploymentManager action
zeuslabs-l3-aethora-15b-v2_v16 status is now inactive due to auto deactivation removed underperforming models
admin requested tearing down of zeuslabs-l3-aethora-15b-v2_v16
Running pipeline stage ISVCDeleter
Checking if service zeuslabs-l3-aethora-15b-v2-v16 is running
Tearing down inference service zeuslabs-l3-aethora-15b-v2-v16
Service zeuslabs-l3-aethora-15b-v2-v16 has been torndown
Pipeline stage ISVCDeleter completed in 4.78s
Running pipeline stage MKMLModelDeleter
Cleaning model data from S3
Cleaning model data from model cache
Deleting key zeuslabs-l3-aethora-15b-v2-v16/config.json from bucket guanaco-mkml-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16/flywheel_model.0.safetensors from bucket guanaco-mkml-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16/flywheel_model.1.safetensors from bucket guanaco-mkml-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16/special_tokens_map.json from bucket guanaco-mkml-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16/tokenizer.json from bucket guanaco-mkml-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16/tokenizer_config.json from bucket guanaco-mkml-models
Cleaning model data from model cache
Deleting key zeuslabs-l3-aethora-15b-v2-v16_reward/config.json from bucket guanaco-reward-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16_reward/merges.txt from bucket guanaco-reward-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16_reward/reward.tensors from bucket guanaco-reward-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16_reward/special_tokens_map.json from bucket guanaco-reward-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16_reward/tokenizer.json from bucket guanaco-reward-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16_reward/tokenizer_config.json from bucket guanaco-reward-models
Deleting key zeuslabs-l3-aethora-15b-v2-v16_reward/vocab.json from bucket guanaco-reward-models
Pipeline stage MKMLModelDeleter completed in 6.93s
zeuslabs-l3-aethora-15b-v2_v16 status is now torndown due to DeploymentManager action

Usage Metrics

Latency Metrics