submission_id: kaoeiri-hedone-l3-1-8b-v1-1_v1
developer_uid: helkero
alignment_samples: 0
best_of: 16
celo_rating: 1202.41
display_name: kaoeiri-hedone-l3-1-8b-v1-1_v1
formatter: {'memory_template': "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{bot_name}'s Persona: {memory}\n\n", 'prompt_template': '{prompt}<|eot_id|>', 'bot_template': '<|start_header_id|>assistant<|end_header_id|>\n\n{bot_name}: {message}<|eot_id|>', 'user_template': '<|start_header_id|>user<|end_header_id|>\n\n{user_name}: {message}<|eot_id|>', 'response_template': '<|start_header_id|>assistant<|end_header_id|>\n\n{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 1.0, 'top_p': 0.95, 'min_p': 0.05, 'top_k': 80, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n', '<|eot_id|>'], 'max_input_tokens': 512, 'best_of': 16, 'max_output_tokens': 64, 'reward_max_token_input': 256}
is_internal_developer: False
language_model: Kaoeiri/Hedone-L3.1-8B-v1.1
max_input_tokens: 512
max_output_tokens: 64
model_architecture: LlamaForCausalLM
model_group: Kaoeiri/Hedone-L3.1-8B-v
model_name: kaoeiri-hedone-l3-1-8b-v1-1_v1
model_num_parameters: 8030261248.0
model_repo: Kaoeiri/Hedone-L3.1-8B-v1.1
model_size: 8B
num_battles: 13588
num_wins: 6791
propriety_score: 0.7098919368246052
propriety_total_count: 1203.0
ranking_group: single
reward_formatter: {'bot_template': '{bot_name}: {message}\n', 'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'response_template': '{bot_name}:', 'truncate_by_message': False, 'user_template': '{user_name}: {message}\n'}
reward_repo: Jellywibble/gpt2_xl_pairwise_89m_step_347634
status: torndown
submission_type: basic
timestamp: 2024-08-06T15:18:32+00:00
us_pacific_date: 2024-08-06
win_ratio: 0.49977921695613775
Download Preference Data
Resubmit model
Running pipeline stage MKMLizer
Starting job with name kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer
Waiting for job on kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer to finish
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ _____ __ __ ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ /___/ ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ Version: 0.9.9 ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ https://mk1.ai ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ The license key for the current software has been verified as ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ belonging to: ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ Chai Research Corp. ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ Expiration: 2024-10-15 23:59:59 ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ║ ║
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Downloaded to shared memory in 30.592s
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmp_bw7ivfm, device:0
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Saving flywheel model at /dev/shm/model_cache
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: quantized model in 27.474s
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Processed model Kaoeiri/Hedone-L3.1-8B-v1.1 in 58.066s
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: creating bucket guanaco-mkml-models
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/kaoeiri-hedone-l3-1-8b-v1-1-v1
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/kaoeiri-hedone-l3-1-8b-v1-1-v1/config.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/kaoeiri-hedone-l3-1-8b-v1-1-v1/special_tokens_map.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/kaoeiri-hedone-l3-1-8b-v1-1-v1/tokenizer_config.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/kaoeiri-hedone-l3-1-8b-v1-1-v1/tokenizer.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/kaoeiri-hedone-l3-1-8b-v1-1-v1/flywheel_model.0.safetensors
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: loading reward model from Jellywibble/gpt2_xl_pairwise_89m_step_347634
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 1%| | 2/291 [00:04<11:40, 2.42s/it] Loading 0: 2%|▏ | 6/291 [00:04<03:06, 1.53it/s] Loading 0: 4%|▍ | 11/291 [00:05<01:22, 3.40it/s] Loading 0: 5%|▌ | 15/291 [00:05<00:52, 5.25it/s] Loading 0: 8%|▊ | 22/291 [00:05<00:27, 9.69it/s] Loading 0: 9%|▉ | 27/291 [00:05<00:20, 13.20it/s] Loading 0: 11%|█ | 32/291 [00:05<00:15, 17.13it/s] Loading 0: 13%|█▎ | 38/291 [00:05<00:11, 21.35it/s] Loading 0: 15%|█▍ | 43/291 [00:05<00:09, 25.05it/s] Loading 0: 17%|█▋ | 49/291 [00:05<00:07, 30.83it/s] Loading 0: 19%|█▊ | 54/291 [00:06<00:07, 33.46it/s] Loading 0: 20%|██ | 59/291 [00:06<00:06, 35.98it/s] Loading 0: 22%|██▏ | 64/291 [00:06<00:05, 38.28it/s] Loading 0: 24%|██▎ | 69/291 [00:06<00:06, 32.00it/s] Loading 0: 26%|██▌ | 76/291 [00:06<00:05, 39.69it/s] Loading 0: 28%|██▊ | 81/291 [00:06<00:05, 41.18it/s] Loading 0: 30%|██▉ | 86/291 [00:06<00:04, 42.01it/s] Loading 0: 31%|███▏ | 91/291 [00:06<00:04, 40.70it/s] Loading 0: 33%|███▎ | 96/291 [00:07<00:05, 35.79it/s] Loading 0: 35%|███▌ | 103/291 [00:07<00:04, 42.99it/s] Loading 0: 37%|███▋ | 108/291 [00:07<00:04, 43.06it/s] Loading 0: 39%|███▉ | 113/291 [00:07<00:04, 42.87it/s] Loading 0: 41%|████ | 118/291 [00:07<00:03, 44.18it/s] Loading 0: 42%|████▏ | 123/291 [00:08<00:07, 23.05it/s] Loading 0: 45%|████▍ | 130/291 [00:08<00:05, 30.29it/s] Loading 0: 46%|████▋ | 135/291 [00:08<00:04, 33.35it/s] Loading 0: 48%|████▊ | 140/291 [00:08<00:04, 35.75it/s] Loading 0: 50%|████▉ | 145/291 [00:08<00:03, 38.86it/s] Loading 0: 52%|█████▏ | 150/291 [00:08<00:04, 34.32it/s] Loading 0: 54%|█████▍ | 157/291 [00:08<00:03, 41.65it/s] Loading 0: 56%|█████▌ | 162/291 [00:08<00:03, 42.21it/s] Loading 0: 57%|█████▋ | 167/291 [00:08<00:02, 42.31it/s] Loading 0: 59%|█████▉ | 173/291 [00:09<00:02, 39.97it/s] Loading 0: 61%|██████ | 178/291 [00:09<00:02, 40.46it/s] Loading 0: 63%|██████▎ | 184/291 [00:09<00:02, 44.20it/s] Loading 0: 65%|██████▍ | 189/291 [00:09<00:02, 43.36it/s] Loading 0: 67%|██████▋ | 194/291 [00:09<00:02, 44.07it/s] Loading 0: 69%|██████▊ | 200/291 [00:09<00:02, 41.08it/s] Loading 0: 70%|███████ | 205/291 [00:09<00:02, 41.28it/s] Loading 0: 73%|███████▎ | 211/291 [00:10<00:01, 45.06it/s] Loading 0: 74%|███████▍ | 216/291 [00:10<00:01, 44.77it/s] Loading 0: 76%|███████▌ | 221/291 [00:10<00:01, 45.75it/s] Loading 0: 78%|███████▊ | 227/291 [00:10<00:01, 41.74it/s] Loading 0: 80%|███████▉ | 232/291 [00:10<00:01, 41.43it/s] Loading 0: 82%|████████▏ | 239/291 [00:10<00:01, 47.07it/s] Loading 0: 84%|████████▍ | 244/291 [00:10<00:00, 47.59it/s] Loading 0: 86%|████████▌ | 249/291 [00:10<00:01, 38.30it/s] Loading 0: 88%|████████▊ | 257/291 [00:11<00:00, 46.23it/s] Loading 0: 90%|█████████ | 262/291 [00:11<00:00, 46.79it/s] Loading 0: 92%|█████████▏| 267/291 [00:11<00:00, 38.24it/s] Loading 0: 94%|█████████▍| 274/291 [00:11<00:00, 45.21it/s] Loading 0: 96%|█████████▌| 280/291 [00:11<00:00, 47.28it/s] Loading 0: 98%|█████████▊| 286/291 [00:12<00:00, 26.62it/s] /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:957: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: warnings.warn(
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:785: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: warnings.warn(
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:469: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: warnings.warn(
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Failed to get response for submission turboderp-cat-llama-3-7_8684_v21: ('http://turboderp-cat-llama-3-7-8684-v21-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'request timeout')
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Saving duration: 1.475s
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Processed model Jellywibble/gpt2_xl_pairwise_89m_step_347634 in 11.512s
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: creating bucket guanaco-reward-models
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: Bucket 's3://guanaco-reward-models/' created
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: uploading /tmp/reward_cache to s3://guanaco-reward-models/kaoeiri-hedone-l3-1-8b-v1-1-v1_reward
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /tmp/reward_cache/config.json s3://guanaco-reward-models/kaoeiri-hedone-l3-1-8b-v1-1-v1_reward/config.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /tmp/reward_cache/special_tokens_map.json s3://guanaco-reward-models/kaoeiri-hedone-l3-1-8b-v1-1-v1_reward/special_tokens_map.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /tmp/reward_cache/tokenizer_config.json s3://guanaco-reward-models/kaoeiri-hedone-l3-1-8b-v1-1-v1_reward/tokenizer_config.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /tmp/reward_cache/merges.txt s3://guanaco-reward-models/kaoeiri-hedone-l3-1-8b-v1-1-v1_reward/merges.txt
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /tmp/reward_cache/vocab.json s3://guanaco-reward-models/kaoeiri-hedone-l3-1-8b-v1-1-v1_reward/vocab.json
kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer: cp /tmp/reward_cache/tokenizer.json s3://guanaco-reward-models/kaoeiri-hedone-l3-1-8b-v1-1-v1_reward/tokenizer.json
Job kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer completed after 104.64s with status: succeeded
Stopping job with name kaoeiri-hedone-l3-1-8b-v1-1-v1-mkmlizer
Pipeline stage MKMLizer completed in 105.60s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.09s
Running pipeline stage ISVCDeployer
Creating inference service kaoeiri-hedone-l3-1-8b-v1-1-v1
Waiting for inference service kaoeiri-hedone-l3-1-8b-v1-1-v1 to be ready
Failed to get response for submission turboderp-cat-llama-3-7_8684_v21: ('http://turboderp-cat-llama-3-7-8684-v21-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
Failed to get response for submission turboderp-cat-llama-3-7_8684_v21: ('http://turboderp-cat-llama-3-7-8684-v21-predictor.tenant-chaiml-guanaco.k.chaiverse.com/v1/models/GPT-J-6B-lit-v2:predict', 'activator request timeout')
Inference service kaoeiri-hedone-l3-1-8b-v1-1-v1 ready after 181.19991540908813s
Pipeline stage ISVCDeployer completed in 183.02s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.3334381580352783s
Received healthy response to inference request in 1.4792332649230957s
Received healthy response to inference request in 1.4261753559112549s
Received healthy response to inference request in 1.4298985004425049s
Received healthy response to inference request in 1.4960896968841553s
5 requests
0 failed requests
5th percentile: 1.426919984817505
10th percentile: 1.427664613723755
20th percentile: 1.4291538715362548
30th percentile: 1.439765453338623
40th percentile: 1.4594993591308594
50th percentile: 1.4792332649230957
60th percentile: 1.4859758377075196
70th percentile: 1.4927184104919433
80th percentile: 1.66355938911438
90th percentile: 1.9984987735748292
95th percentile: 2.1659684658050535
99th percentile: 2.2999442195892335
mean time: 1.6329669952392578
Pipeline stage StressChecker completed in 9.14s
kaoeiri-hedone-l3-1-8b-v1-1_v1 status is now deployed due to DeploymentManager action
kaoeiri-hedone-l3-1-8b-v1-1_v1 status is now inactive due to auto deactivation removed underperforming models
kaoeiri-hedone-l3-1-8b-v1-1_v1 status is now torndown due to DeploymentManager action

Usage Metrics

Latency Metrics