submission_id: bbchicago-llama3-8b-lora_3778_v1
developer_uid: Bbbrun0
best_of: 4
celo_rating: 1141.67
display_name: bbchicago-llama3-8b-lora_3778_v1
family_friendly_score: 0.0
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 512, 'best_of': 4, 'max_output_tokens': 64}
is_internal_developer: False
language_model: BBChicago/llama3-8b-lora-v0.0.0
max_input_tokens: 512
max_output_tokens: 64
model_architecture: LlamaForCausalLM
model_eval_status: success
model_group: BBChicago/llama3-8b-lora
model_name: bbchicago-llama3-8b-lora_3778_v1
model_num_parameters: 8030261248.0
model_repo: BBChicago/llama3-8b-lora-v0.0.0
model_size: 8B
num_battles: 10209
num_wins: 4602
ranking_group: single
reward_formatter: {'bot_template': '{bot_name}: {message}\n', 'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'response_template': '{bot_name}:', 'truncate_by_message': False, 'user_template': '{user_name}: {message}\n'}
reward_repo: ChaiML/reward_gpt2_medium_preference_24m_e2
status: torndown
submission_type: basic
timestamp: 2024-05-30T08:42:14+00:00
us_pacific_date: 2024-05-30
win_ratio: 0.4507787246547164
Resubmit model
Running pipeline stage MKMLizer
Starting job with name bbchicago-llama3-8b-lora-3778-v1-mkmlizer
Waiting for job on bbchicago-llama3-8b-lora-3778-v1-mkmlizer to finish
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ _____ __ __ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ /___/ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Version: 0.8.14 ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ https://mk1.ai ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ The license key for the current software has been verified as ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ belonging to: ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Chai Research Corp. ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Expiration: 2024-07-15 23:59:59 ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_deprecation.py:131: FutureWarning: 'list_files_info' (from 'huggingface_hub.hf_api') is deprecated and will be removed from version '0.23'. Use `list_repo_tree` and `get_paths_info` instead.
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: warnings.warn(warning_message, FutureWarning)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Traceback (most recent call last):
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 509, in http_get
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: hf_transfer.download(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Exception: Error while downloading: reqwest::Error { kind: Request, url: Url { scheme: "https", cannot_be_a_base: false, username: "", password: None, host: Some(Domain("cdn-lfs-us-1.huggingface.co")), port: None, path: "/repos/06/3d/c283c0a307186c4fdbe35/5f8265ecf54f236743ffff1c2cc0600d34a00e7e75ee7801ed6915c9544a36d2", query: Some("response-content-disposition=attachment%3B+filename*%3DUTF-8%27%27model-00004-of-00017.safetensors%3B+filename%3D%22model-00004-of-00017.safetensors%22%3B&Expires=1717317794&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTcxNzMxNzc5NH19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2RuLWxmcy11cy0xLmh1Z2dpbmdmYWNlLmNvL3JlcG9zLzA2LzNkLzA2M2Q2YzYxYjg2NDNiZDUzMWM0ODBkYzRjMzdhNDEzZDgzY2U3OTZlNWZjMjgzYzBhMzA3MTg2YzRmZGJlMzUvNWY4MjY1ZWNmNTRmMjM2NzQzZmZmZjFjMmNjMDYwMGQzNGEwMGU3ZTc1ZWU3ODAxZWQ2OTE1Yzk1NDRhMzZkMj9yZXNwb25zZS1jb250ZW50LWRpc3Bvc2l0aW9uPSoifV19&Signature=JmflDDfkL5Sl0hnB7uJc3qbb23NB4euFDPGQt6mzRU6pO3C6xfI6WZgCviNcRPe%7EOpQmdY9EhETkCCnDpLG8r1tHVRr%7E2XKuWMBjbcM6RNw6VpY8OzKg85C7haoonDCvtudwHr7qtJzKQ07Wy9FTLtFoXD12DppUFHpjSH%7ENBgAw2WfO8%7EYucf3jlOBm8sNhWkP-gjDG12FYu2cs9j3tSdMoyjoYR5viS%7EQOGtepRcdnbX5tqc0KpnAJLELSQmi2RKyTUOjoLXwvhvYE5m%7EUirASz%7ECoWDssutE%7Eyi%7EV7Q5I%7EuR3oRZoJuEhJJPyiiZPs67SA6gN7VEknSrHcHZItg__&Key-Pair-Id=KCD77M1F0VK2B"), fragment: None }, source: hyper::Error(Connect, ConnectError("dns error", Custom { kind: Uncategorized, error: "failed to lookup address information: Temporary failure in name resolution" })) }
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: The above exception was the direct cause of the following exception:
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Traceback (most recent call last):
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/code/uploading/mkmlize.py", line 151, in <module>
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cli()
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1128, in __call__
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return self.main(*args, **kwargs)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1053, in main
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: rv = self.invoke(ctx)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1659, in invoke
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return _process_result(sub_ctx.command.invoke(sub_ctx))
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1395, in invoke
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return ctx.invoke(self.callback, **ctx.params)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 754, in invoke
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return __callback(*args, **kwargs)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/code/uploading/mkmlize.py", line 38, in quantize
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: temp_folder = download_to_shared_memory(repo_id, revision, hf_auth_token)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/code/uploading/mkmlize.py", line 65, in download_to_shared_memory
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: snapshot_download(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 119, in _inner_fn
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return fn(*args, **kwargs)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/_snapshot_download.py", line 314, in snapshot_download
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: _inner_hf_hub_download(file)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/_snapshot_download.py", line 290, in _inner_hf_hub_download
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return hf_hub_download(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 119, in _inner_fn
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return fn(*args, **kwargs)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1492, in hf_hub_download
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: http_get(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 520, in http_get
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: raise RuntimeError(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: RuntimeError: An error occurred while downloading using `hf_transfer`. Consider disabling HF_HUB_ENABLE_HF_TRANSFER for better error handling.
Job bbchicago-llama3-8b-lora-3778-v1-mkmlizer completed after 72.66s with status: failed
Stopping job with name bbchicago-llama3-8b-lora-3778-v1-mkmlizer
%s, retrying in %s seconds...
Starting job with name bbchicago-llama3-8b-lora-3778-v1-mkmlizer
Waiting for job on bbchicago-llama3-8b-lora-3778-v1-mkmlizer to finish
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ _____ __ __ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ /___/ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Version: 0.8.14 ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ https://mk1.ai ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ The license key for the current software has been verified as ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ belonging to: ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Chai Research Corp. ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ Expiration: 2024-07-15 23:59:59 ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ║ ║
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_deprecation.py:131: FutureWarning: 'list_files_info' (from 'huggingface_hub.hf_api') is deprecated and will be removed from version '0.23'. Use `list_repo_tree` and `get_paths_info` instead.
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: warnings.warn(warning_message, FutureWarning)
Retrying (%r) after connection broken by '%r': %s
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_deprecation.py:131: FutureWarning: 'list_files_info' (from 'huggingface_hub.hf_api') is deprecated and will be removed from version '0.23'. Use `list_repo_tree` and `get_paths_info` instead.
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: warnings.warn(warning_message, FutureWarning)
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Downloaded to shared memory in 195.724s
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: quantizing model to /dev/shm/model_cache
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Saving flywheel model at /dev/shm/model_cache
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 1%| | 2/291 [00:00<00:18, 15.83it/s] Loading 0: 3%|▎ | 9/291 [00:00<00:06, 43.52it/s] Loading 0: 7%|▋ | 21/291 [00:00<00:03, 75.42it/s] Loading 0: 10%|▉ | 29/291 [00:00<00:04, 58.77it/s] Loading 0: 12%|█▏ | 36/291 [00:00<00:04, 62.01it/s] Loading 0: 15%|█▍ | 43/291 [00:00<00:04, 57.40it/s] Loading 0: 17%|█▋ | 50/291 [00:00<00:04, 54.89it/s] Loading 0: 21%|██▏ | 62/291 [00:01<00:03, 58.26it/s] Loading 0: 23%|██▎ | 68/291 [00:01<00:03, 56.89it/s] Loading 0: 27%|██▋ | 80/291 [00:01<00:02, 72.01it/s] Loading 0: 30%|███ | 88/291 [00:01<00:03, 60.48it/s] Loading 0: 34%|███▍ | 99/291 [00:01<00:02, 69.06it/s] Loading 0: 37%|███▋ | 107/291 [00:01<00:02, 61.41it/s] Loading 0: 39%|███▉ | 114/291 [00:01<00:02, 62.11it/s] Loading 0: 43%|████▎ | 124/291 [00:02<00:02, 64.28it/s] Loading 0: 45%|████▌ | 131/291 [00:02<00:02, 62.99it/s] Loading 0: 49%|████▉ | 143/291 [00:02<00:02, 66.09it/s] Loading 0: 52%|█████▏ | 150/291 [00:02<00:02, 64.36it/s] Loading 0: 54%|█████▍ | 158/291 [00:02<00:01, 67.71it/s] Loading 0: 57%|█████▋ | 165/291 [00:02<00:02, 61.91it/s] Loading 0: 60%|██████ | 175/291 [00:02<00:01, 69.18it/s] Loading 0: 64%|██████▍ | 186/291 [00:02<00:01, 72.09it/s] Loading 0: 67%|██████▋ | 194/291 [00:03<00:01, 64.69it/s] Loading 0: 70%|███████ | 205/291 [00:03<00:01, 67.94it/s] Loading 0: 73%|███████▎ | 212/291 [00:03<00:01, 61.57it/s] Loading 0: 77%|███████▋ | 224/291 [00:03<00:01, 61.37it/s] Loading 0: 79%|███████▉ | 231/291 [00:03<00:00, 62.60it/s] Loading 0: 84%|████████▍ | 245/291 [00:03<00:00, 67.41it/s] Loading 0: 87%|████████▋ | 252/291 [00:03<00:00, 67.59it/s] Loading 0: 90%|████████▉ | 261/291 [00:04<00:00, 72.18it/s] Loading 0: 92%|█████████▏| 269/291 [00:04<00:00, 65.51it/s] Loading 0: 97%|█████████▋| 281/291 [00:04<00:00, 75.31it/s] Loading 0: 99%|█████████▉| 289/291 [00:04<00:00, 67.22it/s] Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: quantized model in 25.802s
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Processed model BBChicago/llama3-8b-lora-v0.0.0 in 224.130s
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/bbchicago-llama3-8b-lora-3778-v1
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/bbchicago-llama3-8b-lora-3778-v1/config.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/bbchicago-llama3-8b-lora-3778-v1/special_tokens_map.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/bbchicago-llama3-8b-lora-3778-v1/tokenizer_config.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/bbchicago-llama3-8b-lora-3778-v1/tokenizer.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/bbchicago-llama3-8b-lora-3778-v1/flywheel_model.0.safetensors
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: loading reward model from ChaiML/reward_gpt2_medium_preference_24m_e2
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:913: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: warnings.warn(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:757: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: warnings.warn(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:468: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: warnings.warn(
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: /opt/conda/lib/python3.10/site-packages/torch/_utils.py:831: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage()
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: return self.fget.__get__(instance, owner)()
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Saving duration: 0.421s
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Processed model ChaiML/reward_gpt2_medium_preference_24m_e2 in 22.058s
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: creating bucket guanaco-reward-models
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: Bucket 's3://guanaco-reward-models/' created
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: uploading /tmp/reward_cache to s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /tmp/reward_cache/config.json s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward/config.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /tmp/reward_cache/tokenizer_config.json s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward/tokenizer_config.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /tmp/reward_cache/special_tokens_map.json s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward/special_tokens_map.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /tmp/reward_cache/merges.txt s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward/merges.txt
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /tmp/reward_cache/vocab.json s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward/vocab.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /tmp/reward_cache/tokenizer.json s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward/tokenizer.json
bbchicago-llama3-8b-lora-3778-v1-mkmlizer: cp /tmp/reward_cache/reward.tensors s3://guanaco-reward-models/bbchicago-llama3-8b-lora-3778-v1_reward/reward.tensors
Job bbchicago-llama3-8b-lora-3778-v1-mkmlizer completed after 409.57s with status: succeeded
Stopping job with name bbchicago-llama3-8b-lora-3778-v1-mkmlizer
Pipeline stage MKMLizer completed in 485.97s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.09s
Running pipeline stage ISVCDeployer
Creating inference service bbchicago-llama3-8b-lora-3778-v1
Waiting for inference service bbchicago-llama3-8b-lora-3778-v1 to be ready
Inference service bbchicago-llama3-8b-lora-3778-v1 ready after 40.17700386047363s
Pipeline stage ISVCDeployer completed in 47.77s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.024200916290283s
Received healthy response to inference request in 1.111755609512329s
Received healthy response to inference request in 1.1441864967346191s
Received healthy response to inference request in 1.142338752746582s
Received healthy response to inference request in 1.118805170059204s
5 requests
0 failed requests
5th percentile: 1.113165521621704
10th percentile: 1.1145754337310791
20th percentile: 1.1173952579498292
30th percentile: 1.1235118865966798
40th percentile: 1.132925319671631
50th percentile: 1.142338752746582
60th percentile: 1.1430778503417969
70th percentile: 1.1438169479370117
80th percentile: 1.320189380645752
90th percentile: 1.6721951484680178
95th percentile: 1.8481980323791503
99th percentile: 1.9890003395080567
mean time: 1.3082573890686036
Pipeline stage StressChecker completed in 7.38s
Running pipeline stage DaemonicModelEvalScorer
Pipeline stage DaemonicModelEvalScorer completed in 0.05s
Running pipeline stage DaemonicSafetyScorer
Running M-Eval for topic stay_in_character
Pipeline stage DaemonicSafetyScorer completed in 0.04s
M-Eval Dataset for topic stay_in_character is loaded
bbchicago-llama3-8b-lora_3778_v1 status is now deployed due to DeploymentManager action
bbchicago-llama3-8b-lora_3778_v1 status is now inactive due to auto deactivation removed underperforming models
admin requested tearing down of bbchicago-llama3-8b-lora_3778_v1
Running pipeline stage ISVCDeleter
Checking if service bbchicago-llama3-8b-lora-3778-v1 is running
Skipping teardown as no inference service was found
Pipeline stage ISVCDeleter completed in 2.93s
Running pipeline stage MKMLModelDeleter
Cleaning model data from S3
Cleaning model data from model cache
Deleting key bbchicago-llama3-8b-lora-3778-v1/config.json from bucket guanaco-mkml-models
Deleting key bbchicago-llama3-8b-lora-3778-v1/flywheel_model.0.safetensors from bucket guanaco-mkml-models
Deleting key bbchicago-llama3-8b-lora-3778-v1/special_tokens_map.json from bucket guanaco-mkml-models
Deleting key bbchicago-llama3-8b-lora-3778-v1/tokenizer.json from bucket guanaco-mkml-models
Deleting key bbchicago-llama3-8b-lora-3778-v1/tokenizer_config.json from bucket guanaco-mkml-models
Cleaning model data from model cache
Deleting key bbchicago-llama3-8b-lora-3778-v1_reward/config.json from bucket guanaco-reward-models
Deleting key bbchicago-llama3-8b-lora-3778-v1_reward/merges.txt from bucket guanaco-reward-models
Deleting key bbchicago-llama3-8b-lora-3778-v1_reward/reward.tensors from bucket guanaco-reward-models
Deleting key bbchicago-llama3-8b-lora-3778-v1_reward/special_tokens_map.json from bucket guanaco-reward-models
Deleting key bbchicago-llama3-8b-lora-3778-v1_reward/tokenizer.json from bucket guanaco-reward-models
Deleting key bbchicago-llama3-8b-lora-3778-v1_reward/tokenizer_config.json from bucket guanaco-reward-models
Deleting key bbchicago-llama3-8b-lora-3778-v1_reward/vocab.json from bucket guanaco-reward-models
Pipeline stage MKMLModelDeleter completed in 4.03s
bbchicago-llama3-8b-lora_3778_v1 status is now torndown due to DeploymentManager action