submission_id: thanhdaonguyen-once-upon-a-t_v35
developer_uid: chai_backend_admin
status: deployed
model_repo: thanhdaonguyen/once-upon-a-time
reward_repo: rirv938/gpt2_ties_merge_preference_plus_classic_e2_density_99
generation_params: {'temperature': 0.72, 'top_p': 0.73, 'top_k': 1000, 'presence_penalty': 0.7, 'frequency_penalty': 0.3, 'stopping_words': ['</s>', '<|user|>', '###', '\n'], 'max_input_tokens': 512, 'best_of': 8, 'max_output_tokens': 64}
formatter: {'memory_template': "### Instruction:\n\n{bot_name}'s Persona: {memory}.\n\nPlay the role of {bot_name}. Engage in a chat with {user_name} while stay in character. Do not write dialogues and narration for {user_name}. {bot_name} should response with engaging messages of medium length that encourage responses.", 'prompt_template': '{prompt}\n\n', 'bot_template': '### Response:\n\n{bot_name}: {message}\n\n', 'user_template': '### Input:\n\n{user_name}: {message}\n\n', 'response_template': '### Response:\n\n{bot_name}:'}
reward_formatter: {'memory_template': 'Memory: {memory}\n', 'prompt_template': '{prompt}\n', 'bot_template': 'Bot: {message}\n', 'user_template': 'User: {message}\n', 'response_template': 'Bot:'}
timestamp: 2024-02-25T02:01:48+00:00
model_name: thanhdaonguyen-once-upon-a-t_v35
model_eval_status: success
safety_score: 0.94
entertaining: 7.06
stay_in_character: 8.69
user_preference: 7.3
double_thumbs_up: 3379
thumbs_up: 5173
thumbs_down: 2393
num_battles: 183865
num_wins: 91867
win_ratio: 0.49964376036766106
celo_rating: 1153.44
Resubmit model
Running pipeline stage MKMLizer
Starting job with name thanhdaonguyen-once-upon-a-t-v35-mkmlizer
Waiting for job on thanhdaonguyen-once-upon-a-t-v35-mkmlizer to finish
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ _____ __ __ ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ /___/ ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ Version: 0.6.11 ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ The license key for the current software has been verified as ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ belonging to: ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ Chai Research Corp. ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ Expiration: 2024-04-15 23:59:59 ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ║ ║
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: .gitattributes: 0%| | 0.00/1.52k [00:00<?, ?B/s] .gitattributes: 100%|██████████| 1.52k/1.52k [00:00<00:00, 19.5MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: added_tokens.json: 0%| | 0.00/21.0 [00:00<?, ?B/s] added_tokens.json: 100%|██████████| 21.0/21.0 [00:00<00:00, 176kB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: config.json: 0%| | 0.00/643 [00:00<?, ?B/s] config.json: 100%|██████████| 643/643 [00:00<00:00, 5.29MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: generation_config.json: 0%| | 0.00/154 [00:00<?, ?B/s] generation_config.json: 100%|██████████| 154/154 [00:00<00:00, 1.07MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: pytorch_model-00001-of-00003.bin: 0%| | 0.00/9.95G [00:00<?, ?B/s] pytorch_model-00001-of-00003.bin: 0%| | 10.5M/9.95G [00:00<03:28, 47.7MB/s] pytorch_model-00001-of-00003.bin: 0%| | 21.0M/9.95G [00:00<03:13, 51.3MB/s] pytorch_model-00001-of-00003.bin: 1%| | 62.9M/9.95G [00:00<01:04, 152MB/s] pytorch_model-00001-of-00003.bin: 1%| | 94.4M/9.95G [00:00<00:52, 187MB/s] pytorch_model-00001-of-00003.bin: 1%|▏ | 126M/9.95G [00:00<00:48, 202MB/s] pytorch_model-00001-of-00003.bin: 3%|▎ | 315M/9.95G [00:00<00:14, 645MB/s] pytorch_model-00001-of-00003.bin: 9%|▉ | 912M/9.95G [00:00<00:04, 2.05GB/s] pytorch_model-00001-of-00003.bin: 12%|█▏ | 1.15G/9.95G [00:01<00:04, 1.96GB/s] pytorch_model-00001-of-00003.bin: 15%|█▍ | 1.45G/9.95G [00:01<00:04, 2.12GB/s] pytorch_model-00001-of-00003.bin: 17%|█▋ | 1.68G/9.95G [00:01<00:05, 1.52GB/s] pytorch_model-00001-of-00003.bin: 20%|█▉ | 1.96G/9.95G [00:01<00:04, 1.80GB/s] pytorch_model-00001-of-00003.bin: 23%|██▎ | 2.25G/9.95G [00:01<00:03, 2.04GB/s] pytorch_model-00001-of-00003.bin: 25%|██▌ | 2.50G/9.95G [00:01<00:03, 1.96GB/s] pytorch_model-00001-of-00003.bin: 27%|██▋ | 2.72G/9.95G [00:02<00:04, 1.69GB/s] pytorch_model-00001-of-00003.bin: 32%|███▏ | 3.15G/9.95G [00:02<00:03, 2.20GB/s] pytorch_model-00001-of-00003.bin: 34%|███▍ | 3.41G/9.95G [00:02<00:02, 2.24GB/s] pytorch_model-00001-of-00003.bin: 37%|███▋ | 3.66G/9.95G [00:02<00:03, 1.60GB/s] pytorch_model-00001-of-00003.bin: 40%|███▉ | 3.97G/9.95G [00:02<00:03, 1.91GB/s] pytorch_model-00001-of-00003.bin: 45%|████▍ | 4.47G/9.95G [00:02<00:02, 2.47GB/s] pytorch_model-00001-of-00003.bin: 48%|████▊ | 4.76G/9.95G [00:02<00:02, 2.24GB/s] pytorch_model-00001-of-00003.bin: 50%|█████ | 5.02G/9.95G [00:03<00:02, 1.86GB/s] pytorch_model-00001-of-00003.bin: 53%|█████▎ | 5.24G/9.95G [00:03<00:02, 1.76GB/s] pytorch_model-00001-of-00003.bin: 57%|█████▋ | 5.66G/9.95G [00:03<00:01, 2.27GB/s] pytorch_model-00001-of-00003.bin: 60%|█████▉ | 5.93G/9.95G [00:03<00:01, 2.29GB/s] pytorch_model-00001-of-00003.bin: 62%|██████▏ | 6.20G/9.95G [00:03<00:01, 2.04GB/s] pytorch_model-00001-of-00003.bin: 65%|██████▍ | 6.43G/9.95G [00:03<00:01, 1.85GB/s] pytorch_model-00001-of-00003.bin: 67%|██████▋ | 6.71G/9.95G [00:03<00:01, 2.06GB/s] pytorch_model-00001-of-00003.bin: 70%|███████ | 6.99G/9.95G [00:04<00:01, 2.21GB/s] pytorch_model-00001-of-00003.bin: 73%|███████▎ | 7.24G/9.95G [00:04<00:01, 2.10GB/s] pytorch_model-00001-of-00003.bin: 75%|███████▌ | 7.47G/9.95G [00:04<00:01, 1.83GB/s] pytorch_model-00001-of-00003.bin: 78%|███████▊ | 7.76G/9.95G [00:04<00:01, 2.08GB/s] pytorch_model-00001-of-00003.bin: 81%|████████ | 8.05G/9.95G [00:04<00:00, 2.28GB/s] pytorch_model-00001-of-00003.bin: 83%|████████▎ | 8.30G/9.95G [00:04<00:00, 2.14GB/s] pytorch_model-00001-of-00003.bin: 86%|████████▌ | 8.54G/9.95G [00:04<00:00, 1.97GB/s] pytorch_model-00001-of-00003.bin: 88%|████████▊ | 8.75G/9.95G [00:04<00:00, 1.91GB/s] pytorch_model-00001-of-00003.bin: 91%|█████████ | 9.04G/9.95G [00:05<00:00, 2.17GB/s] pytorch_model-00001-of-00003.bin: 93%|█████████▎| 9.27G/9.95G [00:05<00:00, 2.15GB/s] pytorch_model-00001-of-00003.bin: 96%|█████████▌| 9.54G/9.95G [00:05<00:00, 2.30GB/s] pytorch_model-00001-of-00003.bin: 99%|█████████▉| 9.90G/9.95G [00:05<00:00, 2.52GB/s] pytorch_model-00001-of-00003.bin: 100%|█████████▉| 9.95G/9.95G [00:06<00:00, 1.65GB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: pytorch_model-00002-of-00003.bin: 0%| | 0.00/9.90G [00:00<?, ?B/s] pytorch_model-00002-of-00003.bin: 0%| | 10.5M/9.90G [00:00<05:54, 27.9MB/s] pytorch_model-00002-of-00003.bin: 1%| | 62.9M/9.90G [00:00<01:02, 158MB/s] pytorch_model-00002-of-00003.bin: 1%| | 94.4M/9.90G [00:00<00:58, 168MB/s] pytorch_model-00002-of-00003.bin: 3%|▎ | 315M/9.90G [00:00<00:14, 665MB/s] pytorch_model-00002-of-00003.bin: 10%|▉ | 965M/9.90G [00:00<00:04, 2.14GB/s] pytorch_model-00002-of-00003.bin: 13%|█▎ | 1.26G/9.90G [00:01<00:04, 2.11GB/s] pytorch_model-00002-of-00003.bin: 15%|█▌ | 1.52G/9.90G [00:01<00:05, 1.66GB/s] pytorch_model-00002-of-00003.bin: 17%|█▋ | 1.73G/9.90G [00:01<00:05, 1.63GB/s] pytorch_model-00002-of-00003.bin: 20%|██ | 2.01G/9.90G [00:01<00:04, 1.87GB/s] pytorch_model-00002-of-00003.bin: 24%|██▍ | 2.38G/9.90G [00:01<00:03, 2.27GB/s] pytorch_model-00002-of-00003.bin: 27%|██▋ | 2.64G/9.90G [00:01<00:03, 2.27GB/s] pytorch_model-00002-of-00003.bin: 29%|██▉ | 2.89G/9.90G [00:01<00:03, 2.07GB/s] pytorch_model-00002-of-00003.bin: 32%|███▏ | 3.12G/9.90G [00:02<00:04, 1.56GB/s] pytorch_model-00002-of-00003.bin: 34%|███▍ | 3.37G/9.90G [00:02<00:03, 1.73GB/s] pytorch_model-00002-of-00003.bin: 38%|███▊ | 3.81G/9.90G [00:02<00:02, 2.33GB/s] pytorch_model-00002-of-00003.bin: 41%|████ | 4.08G/9.90G [00:02<00:02, 2.18GB/s] pytorch_model-00002-of-00003.bin: 44%|████▍ | 4.40G/9.90G [00:02<00:02, 2.42GB/s] pytorch_model-00002-of-00003.bin: 47%|████▋ | 4.68G/9.90G [00:02<00:02, 2.33GB/s] pytorch_model-00002-of-00003.bin: 50%|████▉ | 4.93G/9.90G [00:02<00:02, 1.98GB/s] pytorch_model-00002-of-00003.bin: 52%|█████▏ | 5.15G/9.90G [00:03<00:02, 1.82GB/s] pytorch_model-00002-of-00003.bin: 55%|█████▌ | 5.48G/9.90G [00:03<00:02, 2.16GB/s] pytorch_model-00002-of-00003.bin: 59%|█████▉ | 5.84G/9.90G [00:03<00:01, 2.49GB/s] pytorch_model-00002-of-00003.bin: 62%|██████▏ | 6.13G/9.90G [00:03<00:01, 2.56GB/s] pytorch_model-00002-of-00003.bin: 65%|██████▍ | 6.41G/9.90G [00:03<00:01, 2.03GB/s] pytorch_model-00002-of-00003.bin: 67%|██████▋ | 6.64G/9.90G [00:03<00:01, 1.81GB/s] pytorch_model-00002-of-00003.bin: 71%|███████ | 7.03G/9.90G [00:03<00:01, 2.26GB/s] pytorch_model-00002-of-00003.bin: 74%|███████▍ | 7.32G/9.90G [00:03<00:01, 2.38GB/s] pytorch_model-00002-of-00003.bin: 77%|███████▋ | 7.58G/9.90G [00:04<00:01, 2.13GB/s] pytorch_model-00002-of-00003.bin: 79%|███████▉ | 7.82G/9.90G [00:04<00:01, 2.01GB/s] pytorch_model-00002-of-00003.bin: 81%|████████▏ | 8.06G/9.90G [00:04<00:00, 2.09GB/s] pytorch_model-00002-of-00003.bin: 84%|████████▍ | 8.33G/9.90G [00:04<00:00, 2.23GB/s] pytorch_model-00002-of-00003.bin: 86%|████████▋ | 8.57G/9.90G [00:04<00:00, 2.17GB/s] pytorch_model-00002-of-00003.bin: 90%|████████▉ | 8.90G/9.90G [00:04<00:00, 2.49GB/s] pytorch_model-00002-of-00003.bin: 93%|█████████▎| 9.16G/9.90G [00:04<00:00, 2.21GB/s] pytorch_model-00002-of-00003.bin: 97%|█████████▋| 9.56G/9.90G [00:04<00:00, 2.65GB/s] pytorch_model-00002-of-00003.bin: 100%|█████████▉| 9.86G/9.90G [00:04<00:00, 2.75GB/s] pytorch_model-00002-of-00003.bin: 100%|█████████▉| 9.90G/9.90G [00:05<00:00, 1.84GB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: pytorch_model-00003-of-00003.bin: 0%| | 0.00/6.18G [00:00<?, ?B/s] pytorch_model-00003-of-00003.bin: 0%| | 10.5M/6.18G [00:00<02:42, 38.1MB/s] pytorch_model-00003-of-00003.bin: 1%| | 41.9M/6.18G [00:00<00:49, 125MB/s] pytorch_model-00003-of-00003.bin: 2%|▏ | 115M/6.18G [00:00<00:30, 196MB/s] pytorch_model-00003-of-00003.bin: 11%|█ | 671M/6.18G [00:00<00:04, 1.36GB/s] pytorch_model-00003-of-00003.bin: 15%|█▌ | 944M/6.18G [00:00<00:03, 1.59GB/s] pytorch_model-00003-of-00003.bin: 23%|██▎ | 1.39G/6.18G [00:01<00:02, 2.29GB/s] pytorch_model-00003-of-00003.bin: 27%|██▋ | 1.70G/6.18G [00:01<00:02, 2.16GB/s] pytorch_model-00003-of-00003.bin: 32%|███▏ | 1.97G/6.18G [00:01<00:02, 1.63GB/s] pytorch_model-00003-of-00003.bin: 35%|███▌ | 2.19G/6.18G [00:01<00:02, 1.47GB/s] pytorch_model-00003-of-00003.bin: 39%|███▊ | 2.38G/6.18G [00:01<00:02, 1.50GB/s] pytorch_model-00003-of-00003.bin: 46%|████▋ | 2.86G/6.18G [00:01<00:01, 2.18GB/s] pytorch_model-00003-of-00003.bin: 52%|█████▏ | 3.19G/6.18G [00:01<00:01, 2.42GB/s] pytorch_model-00003-of-00003.bin: 58%|█████▊ | 3.55G/6.18G [00:02<00:00, 2.71GB/s] pytorch_model-00003-of-00003.bin: 63%|██████▎ | 3.87G/6.18G [00:02<00:00, 2.72GB/s] pytorch_model-00003-of-00003.bin: 68%|██████▊ | 4.17G/6.18G [00:02<00:00, 2.13GB/s] pytorch_model-00003-of-00003.bin: 72%|███████▏ | 4.42G/6.18G [00:02<00:01, 1.70GB/s] pytorch_model-00003-of-00003.bin: 75%|███████▌ | 4.63G/6.18G [00:02<00:00, 1.63GB/s] pytorch_model-00003-of-00003.bin: 79%|███████▉ | 4.88G/6.18G [00:02<00:00, 1.77GB/s] pytorch_model-00003-of-00003.bin: 86%|████████▌ | 5.31G/6.18G [00:02<00:00, 2.33GB/s] pytorch_model-00003-of-00003.bin: 94%|█████████▎| 5.78G/6.18G [00:03<00:00, 2.88GB/s] pytorch_model-00003-of-00003.bin: 99%|█████████▉| 6.11G/6.18G [00:03<00:00, 2.45GB/s] pytorch_model-00003-of-00003.bin: 100%|█████████▉| 6.18G/6.18G [00:03<00:00, 1.56GB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: pytorch_model.bin.index.json: 0%| | 0.00/29.9k [00:00<?, ?B/s] pytorch_model.bin.index.json: 100%|██████████| 29.9k/29.9k [00:00<00:00, 128MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: special_tokens_map.json: 0%| | 0.00/438 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 438/438 [00:00<00:00, 4.32MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: tokenizer.json: 0%| | 0.00/1.84M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 1.84M/1.84M [00:00<00:00, 35.2MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: tokenizer.model: 0%| | 0.00/500k [00:00<?, ?B/s] tokenizer.model: 100%|██████████| 500k/500k [00:00<00:00, 47.9MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: tokenizer_config.json: 0%| | 0.00/828 [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 828/828 [00:00<00:00, 10.5MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Downloaded to shared memory in 18.156s
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: quantizing model to /dev/shm/model_cache
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Saving mkml model at /dev/shm/model_cache
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Reading /tmp/tmpquv62j_k/pytorch_model.bin.index.json
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Profiling: 0%| | 0/363 [00:00<?, ?it/s] Profiling: 0%| | 1/363 [00:03<20:20, 3.37s/it] Profiling: 38%|███▊ | 139/363 [00:04<00:06, 36.81it/s] Profiling: 77%|███████▋ | 278/363 [00:05<00:01, 66.86it/s] Profiling: 100%|██████████| 363/363 [00:07<00:00, 64.30it/s] Profiling: 100%|██████████| 363/363 [00:07<00:00, 51.17it/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: quantized model in 24.136s
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Processed model thanhdaonguyen/once-upon-a-time in 43.761s
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: creating bucket guanaco-mkml-models
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35/config.json
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35/special_tokens_map.json
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: cp /dev/shm/model_cache/tokenizer.model s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35/tokenizer.model
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35/tokenizer_config.json
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: cp /dev/shm/model_cache/added_tokens.json s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35/added_tokens.json
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35/tokenizer.json
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: cp /dev/shm/model_cache/mkml_model.tensors s3://guanaco-mkml-models/thanhdaonguyen-once-upon-a-t-v35/mkml_model.tensors
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: loading reward model from rirv938/gpt2_ties_merge_preference_plus_classic_e2_density_99
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:1067: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: warnings.warn(
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: config.json: 0%| | 0.00/983 [00:00<?, ?B/s] config.json: 100%|██████████| 983/983 [00:00<00:00, 12.6MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:690: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: warnings.warn(
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: tokenizer_config.json: 0%| | 0.00/445 [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 445/445 [00:00<00:00, 2.94MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: vocab.json: 0%| | 0.00/798k [00:00<?, ?B/s] vocab.json: 100%|██████████| 798k/798k [00:00<00:00, 10.4MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: merges.txt: 0%| | 0.00/456k [00:00<?, ?B/s] merges.txt: 100%|██████████| 456k/456k [00:00<00:00, 19.8MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: tokenizer.json: 0%| | 0.00/2.11M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 2.11M/2.11M [00:00<00:00, 56.3MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: special_tokens_map.json: 0%| | 0.00/441 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 441/441 [00:00<00:00, 4.97MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:472: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: warnings.warn(
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: model.safetensors.index.json: 0%| | 0.00/10.5k [00:00<?, ?B/s] model.safetensors.index.json: 100%|██████████| 10.5k/10.5k [00:00<00:00, 89.8MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Downloading shards: 0%| | 0/1 [00:00<?, ?it/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: model-00001-of-00001.safetensors: 0%| | 0.00/249M [00:00<?, ?B/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: model-00001-of-00001.safetensors: 4%|▍ | 10.5M/249M [00:00<00:03, 69.2MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: model-00001-of-00001.safetensors: 16%|█▌ | 39.2M/249M [00:00<00:01, 174MB/s] 
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: model-00001-of-00001.safetensors: 96%|█████████▌| 238M/249M [00:00<00:00, 791MB/s]  model-00001-of-00001.safetensors: 100%|█████████▉| 249M/249M [00:00<00:00, 572MB/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Downloading shards: 100%|██████████| 1/1 [00:00<00:00, 1.26it/s] Downloading shards: 100%|██████████| 1/1 [00:00<00:00, 1.26it/s]
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Saving duration: 0.076s
thanhdaonguyen-once-upon-a-t-v35-mkmlizer: Processed model rirv938/gpt2_ties_merge_preference_plus_classic_e2_density_99 in 3.244s
Job thanhdaonguyen-once-upon-a-t-v35-mkmlizer completed after 75.83s with status: succeeded
Stopping job with name thanhdaonguyen-once-upon-a-t-v35-mkmlizer
Pipeline stage MKMLizer completed in 80.32s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.11s
Running pipeline stage ISVCDeployer
Creating inference service thanhdaonguyen-once-upon-a-t-v35
Waiting for inference service thanhdaonguyen-once-upon-a-t-v35 to be ready
Inference service thanhdaonguyen-once-upon-a-t-v35 ready after 181.1257839202881s
Pipeline stage ISVCDeployer completed in 188.53s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.727611541748047s
Received healthy response to inference request in 1.7630388736724854s
Received healthy response to inference request in 1.777207851409912s
Received healthy response to inference request in 1.7796854972839355s
Received healthy response to inference request in 1.7648334503173828s
5 requests
0 failed requests
5th percentile: 1.763397789001465
10th percentile: 1.7637567043304443
20th percentile: 1.7644745349884032
30th percentile: 1.7673083305358888
40th percentile: 1.7722580909729004
50th percentile: 1.777207851409912
60th percentile: 1.7781989097595214
70th percentile: 1.7791899681091308
80th percentile: 1.969270706176758
90th percentile: 2.3484411239624023
95th percentile: 2.5380263328552246
99th percentile: 2.6896944999694825
mean time: 1.9624754428863525
Pipeline stage StressChecker completed in 10.73s
Running pipeline stage DaemonicModelEvalScorer
Pipeline stage DaemonicModelEvalScorer completed in 0.04s
Running pipeline stage DaemonicSafetyScorer
Running M-Eval for topic stay_in_character
Pipeline stage DaemonicSafetyScorer completed in 0.04s
M-Eval Dataset for topic stay_in_character is loaded
AUTO_DEACTIVATION: submission %s deactivated %s
thanhdaonguyen-once-upon-a-t_v35 status is now inactive due to auto deactivation removed underperforming models
AUTO_DEACTIVATION: submission %s deactivated %s
thanhdaonguyen-once-upon-a-t_v35 status is now deployed due to admin request

Usage Metrics

Latency Metrics