submission_id: decem-staphylus-llama-v1-9_v11
developer_uid: decem_y
status: torndown
model_repo: decem/Staphylus-Llama-v1.9
reward_repo: decem/wine
generation_params: {'temperature': 1.1, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 80, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['</s>', '<|im_end|>', '\n', '<|endoftext|>'], 'max_input_tokens': 1024, 'best_of': 4, 'max_output_tokens': 64}
formatter: {'memory_template': '### Instruction:\nAs the assistant, your task is to play the role of the assigned character, stay in character and generate detailed descriptions of actions and entertaining response to enhance the vividness of the conversation.\nYour character: {bot_name}.\nContext: {memory}\n', 'prompt_template': '{prompt}\n\n', 'bot_template': '### Response:\n{bot_name}: {message}\n\n', 'user_template': '### Input:\n{user_name}: {message}\n\n', 'response_template': '### Response:\n{bot_name}:', 'truncate_by_message': False}
timestamp: 2024-02-08T12:33:52+00:00
model_name: Dionysus
model_eval_status: success
model_group: decem/Staphylus-Llama-v1
num_battles: 1176402
num_wins: 566896
celo_rating: 1139.38
propriety_score: 0.0
propriety_total_count: 0.0
submission_type: basic
model_architecture: None
model_num_parameters: 13015864320.0
best_of: 4
max_input_tokens: 1024
max_output_tokens: 64
display_name: Dionysus
ineligible_reason: propriety_total_count < 800
language_model: decem/Staphylus-Llama-v1.9
model_size: 13B
reward_model: decem/wine
us_pacific_date: 2024-02-08
win_ratio: 0.48188969416917005
preference_data_url: None
Resubmit model
Running pipeline stage MKMLizer
Starting job with name decem-staphylus-llama-v1-9-v11-mkmlizer
Waiting for job on decem-staphylus-llama-v1-9-v11-mkmlizer to finish
decem-staphylus-llama-v1-9-v11-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ _____ __ __ ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ /___/ ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ Version: 0.6.11 ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ The license key for the current software has been verified as ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ belonging to: ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ Chai Research Corp. ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ Expiration: 2024-04-15 23:59:59 ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ║ ║
decem-staphylus-llama-v1-9-v11-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
decem-staphylus-llama-v1-9-v11-mkmlizer: .gitattributes: 0%| | 0.00/1.52k [00:00<?, ?B/s] .gitattributes: 100%|██████████| 1.52k/1.52k [00:00<00:00, 12.8MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: added_tokens.json: 0%| | 0.00/21.0 [00:00<?, ?B/s] added_tokens.json: 100%|██████████| 21.0/21.0 [00:00<00:00, 158kB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: config.json: 0%| | 0.00/659 [00:00<?, ?B/s] config.json: 100%|██████████| 659/659 [00:00<00:00, 5.19MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: model-00001-of-00003.safetensors: 0%| | 0.00/9.95G [00:00<?, ?B/s] model-00001-of-00003.safetensors: 0%| | 10.5M/9.95G [00:00<05:08, 32.2MB/s] model-00001-of-00003.safetensors: 0%| | 31.5M/9.95G [00:00<02:13, 74.2MB/s] model-00001-of-00003.safetensors: 0%| | 41.9M/9.95G [00:00<03:11, 51.8MB/s] model-00001-of-00003.safetensors: 1%| | 52.4M/9.95G [00:00<02:52, 57.5MB/s] model-00001-of-00003.safetensors: 1%| | 62.9M/9.95G [00:01<03:57, 41.6MB/s] model-00001-of-00003.safetensors: 1%| | 73.4M/9.95G [00:01<03:37, 45.5MB/s] model-00001-of-00003.safetensors: 1%| | 105M/9.95G [00:01<01:56, 84.6MB/s] model-00001-of-00003.safetensors: 1%|▏ | 126M/9.95G [00:01<01:35, 103MB/s] model-00001-of-00003.safetensors: 1%|▏ | 147M/9.95G [00:01<01:22, 119MB/s] model-00001-of-00003.safetensors: 2%|▏ | 168M/9.95G [00:02<01:14, 131MB/s] model-00001-of-00003.safetensors: 2%|▏ | 189M/9.95G [00:02<01:28, 111MB/s] model-00001-of-00003.safetensors: 2%|▏ | 210M/9.95G [00:02<01:26, 113MB/s] model-00001-of-00003.safetensors: 2%|▏ | 231M/9.95G [00:02<01:19, 123MB/s] model-00001-of-00003.safetensors: 3%|▎ | 304M/9.95G [00:02<00:40, 240MB/s] model-00001-of-00003.safetensors: 4%|▍ | 377M/9.95G [00:02<00:36, 259MB/s] model-00001-of-00003.safetensors: 5%|▌ | 503M/9.95G [00:03<00:21, 445MB/s] model-00001-of-00003.safetensors: 12%|█▏ | 1.23G/9.95G [00:03<00:04, 1.88GB/s] model-00001-of-00003.safetensors: 15%|█▍ | 1.49G/9.95G [00:03<00:06, 1.37GB/s] model-00001-of-00003.safetensors: 17%|█▋ | 1.70G/9.95G [00:03<00:05, 1.49GB/s] model-00001-of-00003.safetensors: 19%|█▉ | 1.91G/9.95G [00:03<00:05, 1.48GB/s] model-00001-of-00003.safetensors: 21%|██ | 2.10G/9.95G [00:04<00:07, 1.05GB/s] model-00001-of-00003.safetensors: 23%|██▎ | 2.25G/9.95G [00:04<00:06, 1.11GB/s] model-00001-of-00003.safetensors: 25%|██▍ | 2.47G/9.95G [00:04<00:05, 1.31GB/s] model-00001-of-00003.safetensors: 27%|██▋ | 2.64G/9.95G [00:04<00:06, 1.06GB/s] model-00001-of-00003.safetensors: 28%|██▊ | 2.78G/9.95G [00:04<00:06, 1.11GB/s] model-00001-of-00003.safetensors: 30%|██▉ | 2.96G/9.95G [00:04<00:05, 1.24GB/s] model-00001-of-00003.safetensors: 31%|███ | 3.10G/9.95G [00:04<00:05, 1.21GB/s] model-00001-of-00003.safetensors: 33%|███▎ | 3.24G/9.95G [00:05<00:06, 1.10GB/s] model-00001-of-00003.safetensors: 34%|███▍ | 3.37G/9.95G [00:05<00:07, 861MB/s] model-00001-of-00003.safetensors: 36%|███▌ | 3.57G/9.95G [00:05<00:05, 1.08GB/s] model-00001-of-00003.safetensors: 39%|███▉ | 3.88G/9.95G [00:05<00:03, 1.52GB/s] model-00001-of-00003.safetensors: 41%|████ | 4.08G/9.95G [00:05<00:03, 1.62GB/s] model-00001-of-00003.safetensors: 43%|████▎ | 4.27G/9.95G [00:05<00:03, 1.54GB/s] model-00001-of-00003.safetensors: 45%|████▍ | 4.45G/9.95G [00:05<00:03, 1.49GB/s] model-00001-of-00003.safetensors: 47%|████▋ | 4.65G/9.95G [00:05<00:03, 1.62GB/s] model-00001-of-00003.safetensors: 50%|████▉ | 4.95G/9.95G [00:06<00:02, 1.97GB/s] model-00001-of-00003.safetensors: 52%|█████▏ | 5.16G/9.95G [00:06<00:02, 1.65GB/s] model-00001-of-00003.safetensors: 54%|█████▍ | 5.35G/9.95G [00:06<00:02, 1.64GB/s] model-00001-of-00003.safetensors: 56%|█████▌ | 5.59G/9.95G [00:06<00:02, 1.77GB/s] model-00001-of-00003.safetensors: 58%|█████▊ | 5.80G/9.95G [00:06<00:02, 1.83GB/s] model-00001-of-00003.safetensors: 60%|██████ | 6.00G/9.95G [00:06<00:02, 1.86GB/s] model-00001-of-00003.safetensors: 62%|██████▏ | 6.20G/9.95G [00:06<00:02, 1.67GB/s] model-00001-of-00003.safetensors: 64%|██████▍ | 6.38G/9.95G [00:06<00:02, 1.52GB/s] model-00001-of-00003.safetensors: 66%|██████▌ | 6.56G/9.95G [00:07<00:02, 1.56GB/s] model-00001-of-00003.safetensors: 68%|██████▊ | 6.76G/9.95G [00:07<00:01, 1.67GB/s] model-00001-of-00003.safetensors: 71%|███████ | 7.09G/9.95G [00:07<00:01, 2.04GB/s] model-00001-of-00003.safetensors: 73%|███████▎ | 7.30G/9.95G [00:07<00:01, 1.75GB/s] model-00001-of-00003.safetensors: 75%|███████▌ | 7.49G/9.95G [00:07<00:01, 1.64GB/s] model-00001-of-00003.safetensors: 77%|███████▋ | 7.67G/9.95G [00:07<00:01, 1.61GB/s] model-00001-of-00003.safetensors: 79%|███████▊ | 7.83G/9.95G [00:07<00:01, 1.52GB/s] model-00001-of-00003.safetensors: 82%|████████▏ | 8.14G/9.95G [00:07<00:00, 1.85GB/s] model-00001-of-00003.safetensors: 84%|████████▍ | 8.34G/9.95G [00:08<00:00, 1.79GB/s] model-00001-of-00003.safetensors: 86%|████████▌ | 8.54G/9.95G [00:08<00:00, 1.82GB/s] model-00001-of-00003.safetensors: 88%|████████▊ | 8.75G/9.95G [00:08<00:00, 1.89GB/s] model-00001-of-00003.safetensors: 90%|████████▉ | 8.94G/9.95G [00:08<00:00, 1.53GB/s] model-00001-of-00003.safetensors: 93%|█████████▎| 9.23G/9.95G [00:08<00:00, 1.84GB/s] model-00001-of-00003.safetensors: 97%|█████████▋| 9.64G/9.95G [00:08<00:00, 2.40GB/s] model-00001-of-00003.safetensors: 100%|█████████▉| 9.95G/9.95G [00:08<00:00, 1.14GB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: model-00002-of-00003.safetensors: 0%| | 0.00/9.96G [00:00<?, ?B/s] model-00002-of-00003.safetensors: 0%| | 10.5M/9.96G [00:00<06:07, 27.0MB/s] model-00002-of-00003.safetensors: 0%| | 21.0M/9.96G [00:00<03:44, 44.2MB/s] model-00002-of-00003.safetensors: 1%| | 83.9M/9.96G [00:00<00:54, 180MB/s] model-00002-of-00003.safetensors: 1%| | 115M/9.96G [00:00<00:52, 189MB/s] model-00002-of-00003.safetensors: 2%|▏ | 189M/9.96G [00:00<00:31, 307MB/s] model-00002-of-00003.safetensors: 10%|█ | 1.01G/9.96G [00:01<00:03, 2.26GB/s] model-00002-of-00003.safetensors: 13%|█▎ | 1.31G/9.96G [00:01<00:04, 1.90GB/s] model-00002-of-00003.safetensors: 16%|█▌ | 1.56G/9.96G [00:01<00:07, 1.06GB/s] model-00002-of-00003.safetensors: 18%|█▊ | 1.75G/9.96G [00:01<00:07, 1.12GB/s] model-00002-of-00003.safetensors: 25%|██▍ | 2.49G/9.96G [00:01<00:03, 2.13GB/s] model-00002-of-00003.safetensors: 28%|██▊ | 2.83G/9.96G [00:02<00:03, 2.33GB/s] model-00002-of-00003.safetensors: 32%|███▏ | 3.17G/9.96G [00:02<00:04, 1.51GB/s] model-00002-of-00003.safetensors: 34%|███▍ | 3.43G/9.96G [00:02<00:04, 1.32GB/s] model-00002-of-00003.safetensors: 37%|███▋ | 3.64G/9.96G [00:02<00:04, 1.42GB/s] model-00002-of-00003.safetensors: 40%|███▉ | 3.94G/9.96G [00:03<00:03, 1.69GB/s] model-00002-of-00003.safetensors: 43%|████▎ | 4.28G/9.96G [00:03<00:02, 1.95GB/s] model-00002-of-00003.safetensors: 45%|████▌ | 4.53G/9.96G [00:03<00:03, 1.64GB/s] model-00002-of-00003.safetensors: 48%|████▊ | 4.75G/9.96G [00:03<00:02, 1.74GB/s] model-00002-of-00003.safetensors: 50%|████▉ | 4.97G/9.96G [00:03<00:03, 1.51GB/s] model-00002-of-00003.safetensors: 52%|█████▏ | 5.16G/9.96G [00:03<00:03, 1.58GB/s] model-00002-of-00003.safetensors: 55%|█████▌ | 5.49G/9.96G [00:03<00:02, 1.94GB/s] model-00002-of-00003.safetensors: 58%|█████▊ | 5.73G/9.96G [00:03<00:02, 1.91GB/s] model-00002-of-00003.safetensors: 60%|█████▉ | 5.93G/9.96G [00:04<00:02, 1.82GB/s] model-00002-of-00003.safetensors: 62%|██████▏ | 6.13G/9.96G [00:04<00:02, 1.45GB/s] model-00002-of-00003.safetensors: 63%|██████▎ | 6.30G/9.96G [00:04<00:02, 1.33GB/s] model-00002-of-00003.safetensors: 66%|██████▌ | 6.52G/9.96G [00:04<00:02, 1.49GB/s] model-00002-of-00003.safetensors: 68%|██████▊ | 6.81G/9.96G [00:04<00:01, 1.76GB/s] model-00002-of-00003.safetensors: 72%|███████▏ | 7.12G/9.96G [00:04<00:01, 2.05GB/s] model-00002-of-00003.safetensors: 74%|███████▍ | 7.35G/9.96G [00:04<00:01, 2.00GB/s] model-00002-of-00003.safetensors: 76%|███████▌ | 7.59G/9.96G [00:05<00:01, 2.05GB/s] model-00002-of-00003.safetensors: 78%|███████▊ | 7.81G/9.96G [00:05<00:01, 1.98GB/s] model-00002-of-00003.safetensors: 81%|████████ | 8.06G/9.96G [00:05<00:00, 2.12GB/s] model-00002-of-00003.safetensors: 83%|████████▎ | 8.28G/9.96G [00:05<00:00, 2.14GB/s] model-00002-of-00003.safetensors: 86%|████████▌ | 8.52G/9.96G [00:05<00:00, 2.21GB/s] model-00002-of-00003.safetensors: 88%|████████▊ | 8.81G/9.96G [00:05<00:00, 2.26GB/s] model-00002-of-00003.safetensors: 91%|█████████ | 9.04G/9.96G [00:05<00:00, 1.94GB/s] model-00002-of-00003.safetensors: 93%|█████████▎| 9.25G/9.96G [00:05<00:00, 1.86GB/s] model-00002-of-00003.safetensors: 95%|█████████▍| 9.44G/9.96G [00:06<00:00, 1.41GB/s] model-00002-of-00003.safetensors: 100%|█████████▉| 9.96G/9.96G [00:06<00:00, 1.60GB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: model-00003-of-00003.safetensors: 0%| | 0.00/6.13G [00:00<?, ?B/s] model-00003-of-00003.safetensors: 0%| | 10.5M/6.13G [00:00<03:52, 26.3MB/s] model-00003-of-00003.safetensors: 1%| | 73.4M/6.13G [00:00<00:58, 104MB/s] model-00003-of-00003.safetensors: 2%|▏ | 115M/6.13G [00:00<00:41, 146MB/s] model-00003-of-00003.safetensors: 2%|▏ | 136M/6.13G [00:01<01:00, 99.0MB/s] model-00003-of-00003.safetensors: 3%|▎ | 157M/6.13G [00:02<02:32, 39.0MB/s] model-00003-of-00003.safetensors: 3%|▎ | 210M/6.13G [00:02<01:24, 70.2MB/s] model-00003-of-00003.safetensors: 20%|█▉ | 1.21G/6.13G [00:03<00:05, 906MB/s] model-00003-of-00003.safetensors: 25%|██▌ | 1.55G/6.13G [00:03<00:05, 876MB/s] model-00003-of-00003.safetensors: 30%|██▉ | 1.81G/6.13G [00:03<00:05, 834MB/s] model-00003-of-00003.safetensors: 33%|███▎ | 2.02G/6.13G [00:04<00:04, 840MB/s] model-00003-of-00003.safetensors: 36%|███▌ | 2.19G/6.13G [00:04<00:05, 784MB/s] model-00003-of-00003.safetensors: 38%|███▊ | 2.33G/6.13G [00:04<00:04, 786MB/s] model-00003-of-00003.safetensors: 40%|████ | 2.45G/6.13G [00:04<00:04, 740MB/s] model-00003-of-00003.safetensors: 42%|████▏ | 2.56G/6.13G [00:04<00:05, 694MB/s] model-00003-of-00003.safetensors: 43%|████▎ | 2.66G/6.13G [00:04<00:04, 749MB/s] model-00003-of-00003.safetensors: 45%|████▌ | 2.77G/6.13G [00:05<00:04, 769MB/s] model-00003-of-00003.safetensors: 47%|████▋ | 2.86G/6.13G [00:05<00:05, 643MB/s] model-00003-of-00003.safetensors: 49%|████▉ | 3.00G/6.13G [00:05<00:04, 753MB/s] model-00003-of-00003.safetensors: 58%|█████▊ | 3.57G/6.13G [00:05<00:01, 1.70GB/s] model-00003-of-00003.safetensors: 62%|██████▏ | 3.77G/6.13G [00:05<00:01, 1.72GB/s] model-00003-of-00003.safetensors: 65%|██████▍ | 3.97G/6.13G [00:05<00:01, 1.73GB/s] model-00003-of-00003.safetensors: 68%|██████▊ | 4.17G/6.13G [00:05<00:01, 1.71GB/s] model-00003-of-00003.safetensors: 71%|███████ | 4.36G/6.13G [00:06<00:01, 1.61GB/s] model-00003-of-00003.safetensors: 74%|███████▍ | 4.54G/6.13G [00:06<00:01, 1.48GB/s] model-00003-of-00003.safetensors: 77%|███████▋ | 4.71G/6.13G [00:06<00:00, 1.51GB/s] model-00003-of-00003.safetensors: 81%|████████ | 4.96G/6.13G [00:06<00:00, 1.75GB/s] model-00003-of-00003.safetensors: 84%|████████▍ | 5.15G/6.13G [00:06<00:00, 1.68GB/s] model-00003-of-00003.safetensors: 88%|████████▊ | 5.42G/6.13G [00:06<00:00, 1.94GB/s] model-00003-of-00003.safetensors: 93%|█████████▎| 5.69G/6.13G [00:06<00:00, 2.09GB/s] model-00003-of-00003.safetensors: 100%|█████████▉| 6.13G/6.13G [00:06<00:00, 895MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: model.safetensors.index.json: 0%| | 0.00/28.4k [00:00<?, ?B/s] model.safetensors.index.json: 100%|██████████| 28.4k/28.4k [00:00<00:00, 156MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: special_tokens_map.json: 0%| | 0.00/549 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 549/549 [00:00<00:00, 6.11MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: tokenizer.json: 0%| | 0.00/1.84M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 1.84M/1.84M [00:00<00:00, 33.9MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: tokenizer.model: 0%| | 0.00/500k [00:00<?, ?B/s] tokenizer.model: 100%|██████████| 500k/500k [00:00<00:00, 18.4MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: tokenizer_config.json: 0%| | 0.00/1.07k [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 1.07k/1.07k [00:00<00:00, 8.66MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: Downloaded to shared memory in 23.723s
decem-staphylus-llama-v1-9-v11-mkmlizer: quantizing model to /dev/shm/model_cache
decem-staphylus-llama-v1-9-v11-mkmlizer: Saving mkml model at /dev/shm/model_cache
decem-staphylus-llama-v1-9-v11-mkmlizer: Reading /tmp/tmp8r6sv2c1/model.safetensors.index.json
decem-staphylus-llama-v1-9-v11-mkmlizer: Profiling: 0%| | 0/363 [00:00<?, ?it/s] Profiling: 0%| | 1/363 [00:02<12:27, 2.07s/it] Profiling: 4%|▎ | 13/363 [00:02<00:42, 8.21it/s] Profiling: 8%|▊ | 28/363 [00:02<00:16, 20.16it/s] Profiling: 11%|█▏ | 41/363 [00:02<00:10, 31.69it/s] Profiling: 16%|█▌ | 57/363 [00:02<00:06, 48.58it/s] Profiling: 19%|█▉ | 70/363 [00:02<00:04, 61.28it/s] Profiling: 24%|██▍ | 87/363 [00:02<00:03, 79.45it/s] Profiling: 28%|██▊ | 103/363 [00:02<00:02, 95.90it/s] Profiling: 33%|███▎ | 118/363 [00:02<00:02, 103.53it/s] Profiling: 36%|███▋ | 132/363 [00:03<00:02, 111.66it/s] Profiling: 40%|████ | 146/363 [00:03<00:03, 55.00it/s] Profiling: 44%|████▎ | 158/363 [00:03<00:03, 63.85it/s] Profiling: 48%|████▊ | 174/363 [00:03<00:02, 79.90it/s] Profiling: 52%|█████▏ | 188/363 [00:03<00:01, 90.62it/s] Profiling: 56%|█████▌ | 202/363 [00:03<00:01, 100.76it/s] Profiling: 60%|█████▉ | 217/363 [00:04<00:01, 112.24it/s] Profiling: 64%|██████▎ | 231/363 [00:04<00:01, 117.29it/s] Profiling: 68%|██████▊ | 246/363 [00:04<00:00, 125.48it/s] Profiling: 72%|███████▏ | 260/363 [00:04<00:00, 127.66it/s] Profiling: 75%|███████▌ | 274/363 [00:04<00:00, 128.65it/s] Profiling: 79%|███████▉ | 288/363 [00:06<00:03, 21.94it/s] Profiling: 83%|████████▎ | 303/363 [00:06<00:02, 29.80it/s] Profiling: 88%|████████▊ | 320/363 [00:06<00:01, 40.88it/s] Profiling: 92%|█████████▏| 334/363 [00:06<00:00, 51.12it/s] Profiling: 96%|█████████▌| 348/363 [00:06<00:00, 62.39it/s] Profiling: 100%|██████████| 363/363 [00:07<00:00, 50.64it/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: quantized model in 23.716s
decem-staphylus-llama-v1-9-v11-mkmlizer: Processed model decem/Staphylus-Llama-v1.9 in 49.038s
decem-staphylus-llama-v1-9-v11-mkmlizer: creating bucket guanaco-mkml-models
decem-staphylus-llama-v1-9-v11-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
decem-staphylus-llama-v1-9-v11-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11/config.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /dev/shm/model_cache/added_tokens.json s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11/added_tokens.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11/tokenizer_config.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11/special_tokens_map.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /dev/shm/model_cache/tokenizer.model s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11/tokenizer.model
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11/tokenizer.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /dev/shm/model_cache/mkml_model.tensors s3://guanaco-mkml-models/decem-staphylus-llama-v1-9-v11/mkml_model.tensors
decem-staphylus-llama-v1-9-v11-mkmlizer: loading reward model from decem/wine
decem-staphylus-llama-v1-9-v11-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:1067: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
decem-staphylus-llama-v1-9-v11-mkmlizer: warnings.warn(
decem-staphylus-llama-v1-9-v11-mkmlizer: config.json: 0%| | 0.00/1.03k [00:00<?, ?B/s] config.json: 100%|██████████| 1.03k/1.03k [00:00<00:00, 11.5MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:690: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
decem-staphylus-llama-v1-9-v11-mkmlizer: warnings.warn(
decem-staphylus-llama-v1-9-v11-mkmlizer: tokenizer_config.json: 0%| | 0.00/477 [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 477/477 [00:00<00:00, 3.75MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: vocab.json: 0%| | 0.00/798k [00:00<?, ?B/s] vocab.json: 100%|██████████| 798k/798k [00:00<00:00, 9.75MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: merges.txt: 0%| | 0.00/456k [00:00<?, ?B/s] merges.txt: 100%|██████████| 456k/456k [00:00<00:00, 22.1MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: tokenizer.json: 0%| | 0.00/2.11M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 2.11M/2.11M [00:00<00:00, 30.3MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: special_tokens_map.json: 0%| | 0.00/473 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 473/473 [00:00<00:00, 5.44MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:472: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
decem-staphylus-llama-v1-9-v11-mkmlizer: warnings.warn(
decem-staphylus-llama-v1-9-v11-mkmlizer: pytorch_model.bin: 0%| | 0.00/498M [00:00<?, ?B/s] pytorch_model.bin: 2%|▏ | 10.5M/498M [00:00<00:08, 55.7MB/s] pytorch_model.bin: 5%|▌ | 26.0M/498M [00:00<00:19, 24.8MB/s] pytorch_model.bin: 7%|▋ | 36.4M/498M [00:01<00:14, 31.0MB/s] pytorch_model.bin: 20%|█▉ | 99.4M/498M [00:01<00:03, 116MB/s] pytorch_model.bin: 26%|██▋ | 131M/498M [00:01<00:02, 149MB/s] pytorch_model.bin: 100%|█████████▉| 498M/498M [00:01<00:00, 337MB/s]
decem-staphylus-llama-v1-9-v11-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
decem-staphylus-llama-v1-9-v11-mkmlizer: Saving duration: 0.087s
decem-staphylus-llama-v1-9-v11-mkmlizer: Processed model decem/wine in 3.745s
decem-staphylus-llama-v1-9-v11-mkmlizer: creating bucket guanaco-reward-models
decem-staphylus-llama-v1-9-v11-mkmlizer: Bucket 's3://guanaco-reward-models/' created
decem-staphylus-llama-v1-9-v11-mkmlizer: uploading /tmp/reward_cache to s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /tmp/reward_cache/tokenizer_config.json s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward/tokenizer_config.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /tmp/reward_cache/config.json s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward/config.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /tmp/reward_cache/special_tokens_map.json s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward/special_tokens_map.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /tmp/reward_cache/merges.txt s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward/merges.txt
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /tmp/reward_cache/vocab.json s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward/vocab.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /tmp/reward_cache/tokenizer.json s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward/tokenizer.json
decem-staphylus-llama-v1-9-v11-mkmlizer: cp /tmp/reward_cache/reward.tensors s3://guanaco-reward-models/decem-staphylus-llama-v1-9-v11_reward/reward.tensors
Job decem-staphylus-llama-v1-9-v11-mkmlizer completed after 86.95s with status: succeeded
Stopping job with name decem-staphylus-llama-v1-9-v11-mkmlizer
Pipeline stage MKMLizer completed in 92.47s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.19s
Running pipeline stage ISVCDeployer
Creating inference service decem-staphylus-llama-v1-9-v11
Waiting for inference service decem-staphylus-llama-v1-9-v11 to be ready
Inference service decem-staphylus-llama-v1-9-v11 ready after 60.54819941520691s
Pipeline stage ISVCDeployer completed in 70.15s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.432603359222412s
Received healthy response to inference request in 1.9970407485961914s
Received healthy response to inference request in 1.9180936813354492s
Received healthy response to inference request in 1.551922082901001s
Received healthy response to inference request in 1.8928077220916748s
5 requests
0 failed requests
5th percentile: 1.6200992107391357
10th percentile: 1.6882763385772706
20th percentile: 1.8246305942535401
30th percentile: 1.8978649139404298
40th percentile: 1.9079792976379395
50th percentile: 1.9180936813354492
60th percentile: 1.9496725082397461
70th percentile: 1.981251335144043
80th percentile: 2.0841532707214356
90th percentile: 2.2583783149719237
95th percentile: 2.345490837097168
99th percentile: 2.4151808547973634
mean time: 1.9584935188293457
Pipeline stage StressChecker completed in 11.06s
Running pipeline stage DaemonicModelEvalScorer
Pipeline stage DaemonicModelEvalScorer completed in 0.07s
Running pipeline stage DaemonicSafetyScorer
Running M-Eval for topic stay_in_character
Pipeline stage DaemonicSafetyScorer completed in 0.07s
M-Eval Dataset for topic stay_in_character is loaded
decem-staphylus-llama-v1-9_v11 status is now inactive due to auto deactivation removed underperforming models
decem-staphylus-llama-v1-9_v11 status is now deployed due to admin request
decem-staphylus-llama-v1-9_v11 status is now inactive due to auto deactivation removed underperforming models
admin requested tearing down of decem-staphylus-llama-v1-9_v11
Running pipeline stage ISVCDeleter
Checking if service decem-staphylus-llama-v1-9-v11 is running
Tearing down inference service decem-staphylus-llama-v1-9-v11
Toredown service decem-staphylus-llama-v1-9-v11
Pipeline stage ISVCDeleter completed in 4.07s
Running pipeline stage MKMLModelDeleter
Cleaning model data from S3
Cleaning model data from model cache
Deleting key decem-staphylus-llama-v1-9-v11/added_tokens.json from bucket guanaco-mkml-models
Deleting key decem-staphylus-llama-v1-9-v11/config.json from bucket guanaco-mkml-models
Deleting key decem-staphylus-llama-v1-9-v11/mkml_model.tensors from bucket guanaco-mkml-models
Deleting key decem-staphylus-llama-v1-9-v11/special_tokens_map.json from bucket guanaco-mkml-models
Deleting key decem-staphylus-llama-v1-9-v11/tokenizer.json from bucket guanaco-mkml-models
Deleting key decem-staphylus-llama-v1-9-v11/tokenizer.model from bucket guanaco-mkml-models
Deleting key decem-staphylus-llama-v1-9-v11/tokenizer_config.json from bucket guanaco-mkml-models
Cleaning model data from model cache
Deleting key decem-staphylus-llama-v1-9-v11_reward/config.json from bucket guanaco-reward-models
Deleting key decem-staphylus-llama-v1-9-v11_reward/merges.txt from bucket guanaco-reward-models
Deleting key decem-staphylus-llama-v1-9-v11_reward/reward.tensors from bucket guanaco-reward-models
Deleting key decem-staphylus-llama-v1-9-v11_reward/special_tokens_map.json from bucket guanaco-reward-models
Deleting key decem-staphylus-llama-v1-9-v11_reward/tokenizer.json from bucket guanaco-reward-models
Deleting key decem-staphylus-llama-v1-9-v11_reward/tokenizer_config.json from bucket guanaco-reward-models
Deleting key decem-staphylus-llama-v1-9-v11_reward/vocab.json from bucket guanaco-reward-models
Pipeline stage MKMLModelDeleter completed in 3.19s
decem-staphylus-llama-v1-9_v11 status is now torndown due to DeploymentManager action

Usage Metrics

Latency Metrics