submission_id: taozi555-wl-7b_v3
developer_uid: taozi555
status: inactive
model_repo: taozi555/wl-7b
reward_repo: rirv938/reward_gpt2_medium_preference_24m_e2
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['<|'], 'max_input_tokens': 512, 'best_of': 16, 'max_output_tokens': 64}
formatter: {'memory_template': "<|system|>\nPlay the role of {bot_name}. Engage in a chat with {user_name} while stay in character. Do not write dialogues and narration for {user_name}. {bot_name} should response with engaging messages of medium length that encourage responses.\n{bot_name}'s Persona: {memory}\n", 'prompt_template': '{prompt}\n\n', 'bot_template': '<|assistant|>\n{bot_name}: {message}\n', 'user_template': '<|user|>\n{user_name}: {message}\n', 'response_template': '<|assistant|>\n{bot_name}:'}
reward_formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:'}
timestamp: 2024-03-20T06:38:05+00:00
model_name: wl-7b
model_eval_status: success
safety_score: 0.99
entertaining: 7.22
stay_in_character: 8.77
user_preference: 7.58
double_thumbs_up: 714
thumbs_up: 1147
thumbs_down: 659
num_battles: 67746
num_wins: 33386
win_ratio: 0.4928113836979305
celo_rating: 1152.35
Resubmit model
Running pipeline stage MKMLizer
Starting job with name taozi555-wl-7b-v3-mkmlizer
Waiting for job on taozi555-wl-7b-v3-mkmlizer to finish
Failed to get response for submission thanhdaonguyen-once-upon-a-t_v37: HTTPConnectionPool(host='thanhdaonguyen-once-upon-a-t-v37-predictor-default.tenant-chaiml-guanaco.knative.ord1.coreweave.cloud', port=80): Read timed out. (read timeout=5.5)
taozi555-wl-7b-v3-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
taozi555-wl-7b-v3-mkmlizer: ║ _____ __ __ ║
taozi555-wl-7b-v3-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
taozi555-wl-7b-v3-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
taozi555-wl-7b-v3-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
taozi555-wl-7b-v3-mkmlizer: ║ /___/ ║
taozi555-wl-7b-v3-mkmlizer: ║ ║
taozi555-wl-7b-v3-mkmlizer: ║ Version: 0.6.11 ║
taozi555-wl-7b-v3-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
taozi555-wl-7b-v3-mkmlizer: ║ ║
taozi555-wl-7b-v3-mkmlizer: ║ The license key for the current software has been verified as ║
taozi555-wl-7b-v3-mkmlizer: ║ belonging to: ║
taozi555-wl-7b-v3-mkmlizer: ║ ║
taozi555-wl-7b-v3-mkmlizer: ║ Chai Research Corp. ║
taozi555-wl-7b-v3-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
taozi555-wl-7b-v3-mkmlizer: ║ Expiration: 2024-04-15 23:59:59 ║
taozi555-wl-7b-v3-mkmlizer: ║ ║
taozi555-wl-7b-v3-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
taozi555-wl-7b-v3-mkmlizer: .gitattributes: 0%| | 0.00/1.52k [00:00<?, ?B/s] .gitattributes: 100%|██████████| 1.52k/1.52k [00:00<00:00, 20.0MB/s]
taozi555-wl-7b-v3-mkmlizer: (…)ipynb_checkpoints/config-checkpoint.json: 0%| | 0.00/673 [00:00<?, ?B/s] (…)ipynb_checkpoints/config-checkpoint.json: 100%|██████████| 673/673 [00:00<00:00, 5.46MB/s]
taozi555-wl-7b-v3-mkmlizer: generation_config.json: 0%| | 0.00/111 [00:00<?, ?B/s] generation_config.json: 100%|██████████| 111/111 [00:00<00:00, 897kB/s]
taozi555-wl-7b-v3-mkmlizer: model-00001-of-00003.safetensors: 0%| | 0.00/4.94G [00:00<?, ?B/s] model-00001-of-00003.safetensors: 0%| | 10.5M/4.94G [00:00<01:29, 55.2MB/s] model-00001-of-00003.safetensors: 1%|▏ | 62.9M/4.94G [00:00<00:48, 101MB/s] model-00001-of-00003.safetensors: 3%|▎ | 126M/4.94G [00:00<00:31, 152MB/s] model-00001-of-00003.safetensors: 3%|▎ | 147M/4.94G [00:01<00:32, 146MB/s] model-00001-of-00003.safetensors: 5%|▌ | 262M/4.94G [00:01<00:14, 329MB/s] model-00001-of-00003.safetensors: 11%|█ | 545M/4.94G [00:01<00:05, 841MB/s] model-00001-of-00003.safetensors: 24%|██▍ | 1.17G/4.94G [00:01<00:01, 2.08GB/s] model-00001-of-00003.safetensors: 30%|██▉ | 1.47G/4.94G [00:01<00:01, 2.27GB/s] model-00001-of-00003.safetensors: 36%|███▌ | 1.76G/4.94G [00:01<00:01, 2.13GB/s] model-00001-of-00003.safetensors: 41%|████ | 2.02G/4.94G [00:01<00:01, 1.66GB/s] model-00001-of-00003.safetensors: 45%|████▌ | 2.24G/4.94G [00:02<00:01, 1.74GB/s] model-00001-of-00003.safetensors: 50%|████▉ | 2.45G/4.94G [00:02<00:01, 1.41GB/s] model-00001-of-00003.safetensors: 55%|█████▍ | 2.72G/4.94G [00:02<00:01, 1.62GB/s] model-00001-of-00003.safetensors: 59%|█████▉ | 2.92G/4.94G [00:02<00:01, 1.68GB/s] model-00001-of-00003.safetensors: 63%|██████▎ | 3.11G/4.94G [00:02<00:01, 1.69GB/s] model-00001-of-00003.safetensors: 67%|██████▋ | 3.33G/4.94G [00:02<00:00, 1.79GB/s] model-00001-of-00003.safetensors: 71%|███████▏ | 3.53G/4.94G [00:02<00:00, 1.58GB/s] model-00001-of-00003.safetensors: 77%|███████▋ | 3.80G/4.94G [00:02<00:00, 1.83GB/s] model-00001-of-00003.safetensors: 82%|████████▏ | 4.04G/4.94G [00:03<00:00, 1.96GB/s] model-00001-of-00003.safetensors: 87%|████████▋ | 4.31G/4.94G [00:03<00:00, 2.13GB/s] model-00001-of-00003.safetensors: 94%|█████████▍| 4.67G/4.94G [00:03<00:00, 2.53GB/s] model-00001-of-00003.safetensors: 100%|█████████▉| 4.94G/4.94G [00:03<00:00, 1.50GB/s]
taozi555-wl-7b-v3-mkmlizer: model-00002-of-00003.safetensors: 0%| | 0.00/5.00G [00:00<?, ?B/s] model-00002-of-00003.safetensors: 0%| | 10.5M/5.00G [00:00<03:25, 24.2MB/s] model-00002-of-00003.safetensors: 0%| | 21.0M/5.00G [00:00<02:00, 41.4MB/s] model-00002-of-00003.safetensors: 2%|▏ | 105M/5.00G [00:00<00:22, 214MB/s] model-00002-of-00003.safetensors: 3%|▎ | 136M/5.00G [00:00<00:23, 208MB/s] model-00002-of-00003.safetensors: 4%|▍ | 199M/5.00G [00:00<00:15, 301MB/s] model-00002-of-00003.safetensors: 15%|█▌ | 765M/5.00G [00:01<00:02, 1.61GB/s] model-00002-of-00003.safetensors: 27%|██▋ | 1.33G/5.00G [00:01<00:01, 2.66GB/s] model-00002-of-00003.safetensors: 33%|███▎ | 1.67G/5.00G [00:01<00:01, 1.81GB/s] model-00002-of-00003.safetensors: 39%|███▊ | 1.93G/5.00G [00:01<00:02, 1.28GB/s] model-00002-of-00003.safetensors: 44%|████▍ | 2.21G/5.00G [00:01<00:01, 1.51GB/s] model-00002-of-00003.safetensors: 50%|████▉ | 2.49G/5.00G [00:02<00:01, 1.73GB/s] model-00002-of-00003.safetensors: 57%|█████▋ | 2.85G/5.00G [00:02<00:01, 2.06GB/s] model-00002-of-00003.safetensors: 62%|██████▏ | 3.12G/5.00G [00:02<00:00, 2.19GB/s] model-00002-of-00003.safetensors: 69%|██████▉ | 3.46G/5.00G [00:02<00:00, 2.41GB/s] model-00002-of-00003.safetensors: 75%|███████▍ | 3.74G/5.00G [00:02<00:00, 1.50GB/s] model-00002-of-00003.safetensors: 84%|████████▎ | 4.18G/5.00G [00:02<00:00, 2.00GB/s] model-00002-of-00003.safetensors: 92%|█████████▏| 4.61G/5.00G [00:02<00:00, 2.44GB/s] model-00002-of-00003.safetensors: 100%|█████████▉| 5.00G/5.00G [00:03<00:00, 1.61GB/s]
taozi555-wl-7b-v3-mkmlizer: model-00003-of-00003.safetensors: 0%| | 0.00/4.54G [00:00<?, ?B/s] model-00003-of-00003.safetensors: 0%| | 10.5M/4.54G [00:00<01:25, 53.0MB/s] model-00003-of-00003.safetensors: 2%|▏ | 105M/4.54G [00:00<00:10, 419MB/s] model-00003-of-00003.safetensors: 5%|▍ | 210M/4.54G [00:00<00:06, 641MB/s] model-00003-of-00003.safetensors: 6%|▋ | 294M/4.54G [00:00<00:06, 670MB/s] model-00003-of-00003.safetensors: 10%|█ | 461M/4.54G [00:00<00:04, 965MB/s] model-00003-of-00003.safetensors: 12%|█▏ | 566M/4.54G [00:00<00:04, 930MB/s] model-00003-of-00003.safetensors: 16%|█▋ | 744M/4.54G [00:00<00:03, 1.15GB/s] model-00003-of-00003.safetensors: 26%|██▌ | 1.16G/4.54G [00:00<00:01, 2.02GB/s] model-00003-of-00003.safetensors: 30%|███ | 1.38G/4.54G [00:01<00:01, 1.98GB/s] model-00003-of-00003.safetensors: 36%|███▌ | 1.61G/4.54G [00:01<00:01, 2.06GB/s] model-00003-of-00003.safetensors: 40%|████ | 1.84G/4.54G [00:01<00:01, 1.92GB/s] model-00003-of-00003.safetensors: 45%|████▍ | 2.03G/4.54G [00:01<00:01, 1.79GB/s] model-00003-of-00003.safetensors: 51%|█████ | 2.32G/4.54G [00:01<00:01, 2.05GB/s] model-00003-of-00003.safetensors: 58%|█████▊ | 2.62G/4.54G [00:01<00:00, 2.31GB/s] model-00003-of-00003.safetensors: 63%|██████▎ | 2.86G/4.54G [00:01<00:00, 2.07GB/s] model-00003-of-00003.safetensors: 68%|██████▊ | 3.08G/4.54G [00:01<00:00, 2.02GB/s] model-00003-of-00003.safetensors: 74%|███████▎ | 3.34G/4.54G [00:02<00:00, 2.05GB/s] model-00003-of-00003.safetensors: 79%|███████▊ | 3.57G/4.54G [00:02<00:00, 2.08GB/s] model-00003-of-00003.safetensors: 86%|████████▌ | 3.89G/4.54G [00:02<00:00, 2.39GB/s] model-00003-of-00003.safetensors: 95%|█████████▍| 4.31G/4.54G [00:02<00:00, 2.89GB/s] model-00003-of-00003.safetensors: 100%|█████████▉| 4.54G/4.54G [00:02<00:00, 1.90GB/s]
taozi555-wl-7b-v3-mkmlizer: model.safetensors.index.json: 0%| | 0.00/23.9k [00:00<?, ?B/s] model.safetensors.index.json: 100%|██████████| 23.9k/23.9k [00:00<00:00, 112MB/s]
taozi555-wl-7b-v3-mkmlizer: special_tokens_map.json: 0%| | 0.00/511 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 511/511 [00:00<00:00, 5.72MB/s]
taozi555-wl-7b-v3-mkmlizer: tokenizer.json: 0%| | 0.00/1.80M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 1.80M/1.80M [00:00<00:00, 25.2MB/s]
taozi555-wl-7b-v3-mkmlizer: tokenizer.model: 0%| | 0.00/493k [00:00<?, ?B/s] tokenizer.model: 100%|██████████| 493k/493k [00:00<00:00, 60.8MB/s]
taozi555-wl-7b-v3-mkmlizer: tokenizer_config.json: 0%| | 0.00/1.01k [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 1.01k/1.01k [00:00<00:00, 7.99MB/s]
taozi555-wl-7b-v3-mkmlizer: Downloaded to shared memory in 10.863s
taozi555-wl-7b-v3-mkmlizer: quantizing model to /dev/shm/model_cache
taozi555-wl-7b-v3-mkmlizer: Saving mkml model at /dev/shm/model_cache
taozi555-wl-7b-v3-mkmlizer: Reading /tmp/tmpxzze8o37/model.safetensors.index.json
taozi555-wl-7b-v3-mkmlizer: Profiling: 0%| | 0/291 [00:00<?, ?it/s] Profiling: 0%| | 1/291 [00:01<06:04, 1.26s/it] Profiling: 5%|▍ | 14/291 [00:01<00:19, 13.87it/s] Profiling: 11%|█▏ | 33/291 [00:01<00:07, 36.05it/s] Profiling: 19%|█▊ | 54/291 [00:01<00:03, 61.83it/s] Profiling: 24%|██▍ | 70/291 [00:01<00:02, 76.71it/s] Profiling: 31%|███ | 90/291 [00:01<00:02, 96.22it/s] Profiling: 36%|███▋ | 106/291 [00:02<00:02, 69.06it/s] Profiling: 43%|████▎ | 124/291 [00:02<00:01, 86.72it/s] Profiling: 49%|████▉ | 143/291 [00:02<00:01, 105.40it/s] Profiling: 55%|█████▌ | 161/291 [00:02<00:01, 120.60it/s] Profiling: 62%|██████▏ | 179/291 [00:02<00:00, 133.79it/s] Profiling: 68%|██████▊ | 199/291 [00:02<00:00, 149.88it/s] Profiling: 75%|███████▍ | 217/291 [00:04<00:02, 33.66it/s] Profiling: 79%|███████▉ | 231/291 [00:04<00:01, 41.26it/s] Profiling: 87%|████████▋ | 253/291 [00:04<00:00, 58.05it/s] Profiling: 94%|█████████▍| 274/291 [00:04<00:00, 75.94it/s] Profiling: 100%|██████████| 291/291 [00:04<00:00, 60.40it/s]
taozi555-wl-7b-v3-mkmlizer: quantized model in 15.074s
taozi555-wl-7b-v3-mkmlizer: Processed model taozi555/wl-7b in 26.907s
taozi555-wl-7b-v3-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
taozi555-wl-7b-v3-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/taozi555-wl-7b-v3
taozi555-wl-7b-v3-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/taozi555-wl-7b-v3/config.json
taozi555-wl-7b-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer.model s3://guanaco-mkml-models/taozi555-wl-7b-v3/tokenizer.model
taozi555-wl-7b-v3-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/taozi555-wl-7b-v3/special_tokens_map.json
taozi555-wl-7b-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/taozi555-wl-7b-v3/tokenizer_config.json
taozi555-wl-7b-v3-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/taozi555-wl-7b-v3/tokenizer.json
taozi555-wl-7b-v3-mkmlizer: cp /dev/shm/model_cache/mkml_model.tensors s3://guanaco-mkml-models/taozi555-wl-7b-v3/mkml_model.tensors
taozi555-wl-7b-v3-mkmlizer: loading reward model from rirv938/reward_gpt2_medium_preference_24m_e2
taozi555-wl-7b-v3-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:1067: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
taozi555-wl-7b-v3-mkmlizer: warnings.warn(
taozi555-wl-7b-v3-mkmlizer: config.json: 0%| | 0.00/1.05k [00:00<?, ?B/s] config.json: 100%|██████████| 1.05k/1.05k [00:00<00:00, 12.5MB/s]
taozi555-wl-7b-v3-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:690: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
taozi555-wl-7b-v3-mkmlizer: warnings.warn(
taozi555-wl-7b-v3-mkmlizer: tokenizer_config.json: 0%| | 0.00/234 [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 234/234 [00:00<00:00, 2.69MB/s]
taozi555-wl-7b-v3-mkmlizer: vocab.json: 0%| | 0.00/1.04M [00:00<?, ?B/s] vocab.json: 100%|██████████| 1.04M/1.04M [00:00<00:00, 11.1MB/s]
taozi555-wl-7b-v3-mkmlizer: tokenizer.json: 0%| | 0.00/2.11M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 2.11M/2.11M [00:00<00:00, 43.9MB/s]
taozi555-wl-7b-v3-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:472: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
taozi555-wl-7b-v3-mkmlizer: warnings.warn(
taozi555-wl-7b-v3-mkmlizer: pytorch_model.bin: 0%| | 0.00/1.44G [00:00<?, ?B/s] pytorch_model.bin: 1%| | 10.5M/1.44G [00:00<00:23, 60.2MB/s] pytorch_model.bin: 2%|▏ | 31.5M/1.44G [00:00<00:16, 86.9MB/s] pytorch_model.bin: 9%|▊ | 126M/1.44G [00:00<00:04, 311MB/s] pytorch_model.bin: 15%|█▌ | 220M/1.44G [00:00<00:02, 469MB/s] pytorch_model.bin: 24%|██▍ | 346M/1.44G [00:01<00:02, 378MB/s] pytorch_model.bin: 28%|██▊ | 398M/1.44G [00:01<00:02, 359MB/s] pytorch_model.bin: 33%|███▎ | 482M/1.44G [00:01<00:02, 445MB/s] pytorch_model.bin: 50%|█████ | 724M/1.44G [00:01<00:00, 854MB/s] pytorch_model.bin: 100%|█████████▉| 1.44G/1.44G [00:01<00:00, 944MB/s]
taozi555-wl-7b-v3-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
taozi555-wl-7b-v3-mkmlizer: Saving duration: 0.234s
taozi555-wl-7b-v3-mkmlizer: Processed model rirv938/reward_gpt2_medium_preference_24m_e2 in 5.151s
taozi555-wl-7b-v3-mkmlizer: creating bucket guanaco-reward-models
taozi555-wl-7b-v3-mkmlizer: Bucket 's3://guanaco-reward-models/' created
taozi555-wl-7b-v3-mkmlizer: uploading /tmp/reward_cache to s3://guanaco-reward-models/taozi555-wl-7b-v3_reward
taozi555-wl-7b-v3-mkmlizer: cp /tmp/reward_cache/config.json s3://guanaco-reward-models/taozi555-wl-7b-v3_reward/config.json
taozi555-wl-7b-v3-mkmlizer: cp /tmp/reward_cache/special_tokens_map.json s3://guanaco-reward-models/taozi555-wl-7b-v3_reward/special_tokens_map.json
taozi555-wl-7b-v3-mkmlizer: cp /tmp/reward_cache/tokenizer_config.json s3://guanaco-reward-models/taozi555-wl-7b-v3_reward/tokenizer_config.json
taozi555-wl-7b-v3-mkmlizer: cp /tmp/reward_cache/merges.txt s3://guanaco-reward-models/taozi555-wl-7b-v3_reward/merges.txt
taozi555-wl-7b-v3-mkmlizer: cp /tmp/reward_cache/vocab.json s3://guanaco-reward-models/taozi555-wl-7b-v3_reward/vocab.json
taozi555-wl-7b-v3-mkmlizer: cp /tmp/reward_cache/tokenizer.json s3://guanaco-reward-models/taozi555-wl-7b-v3_reward/tokenizer.json
taozi555-wl-7b-v3-mkmlizer: cp /tmp/reward_cache/reward.tensors s3://guanaco-reward-models/taozi555-wl-7b-v3_reward/reward.tensors
Job taozi555-wl-7b-v3-mkmlizer completed after 53.53s with status: succeeded
Stopping job with name taozi555-wl-7b-v3-mkmlizer
Pipeline stage MKMLizer completed in 57.31s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.11s
Running pipeline stage ISVCDeployer
Creating inference service taozi555-wl-7b-v3
Waiting for inference service taozi555-wl-7b-v3 to be ready
Inference service taozi555-wl-7b-v3 ready after 30.17610192298889s
Pipeline stage ISVCDeployer completed in 37.62s
Running pipeline stage StressChecker
Received healthy response to inference request in 1.7055344581604004s
Received healthy response to inference request in 1.2150797843933105s
Received healthy response to inference request in 1.2160687446594238s
Received healthy response to inference request in 1.2332100868225098s
Received healthy response to inference request in 1.2156097888946533s
5 requests
0 failed requests
5th percentile: 1.215185785293579
10th percentile: 1.2152917861938477
20th percentile: 1.2155037879943849
30th percentile: 1.2157015800476074
40th percentile: 1.2158851623535156
50th percentile: 1.2160687446594238
60th percentile: 1.2229252815246583
70th percentile: 1.2297818183898925
80th percentile: 1.327674961090088
90th percentile: 1.5166047096252442
95th percentile: 1.6110695838928222
99th percentile: 1.6866414833068848
mean time: 1.3171005725860596
Pipeline stage StressChecker completed in 7.50s
Running pipeline stage DaemonicModelEvalScorer
Pipeline stage DaemonicModelEvalScorer completed in 0.04s
Running pipeline stage DaemonicSafetyScorer
Running M-Eval for topic stay_in_character
Pipeline stage DaemonicSafetyScorer completed in 0.06s
M-Eval Dataset for topic stay_in_character is loaded
taozi555-wl-7b_v3 status is now inactive due to auto deactivation removed underperforming models

Usage Metrics

Latency Metrics