developer_uid: Trace2333
submission_id: trace2333-joint-filtered_3791_v5
model_name: trace2333-joint-filtered_3791_v5
model_group: Trace2333/joint_filtered
status: torndown
timestamp: 2024-07-04T18:23:21+00:00
num_battles: 13555
num_wins: 5273
celo_rating: 1103.41
family_friendly_score: 0.0
submission_type: basic
model_repo: Trace2333/joint_filtered_llama2_v0
model_architecture: LlamaForCausalLM
reward_repo: ChaiML/reward_gpt2_medium_preference_24m_e2
model_num_parameters: 6738415616.0
best_of: 16
max_input_tokens: 512
max_output_tokens: 64
display_name: trace2333-joint-filtered_3791_v5
is_internal_developer: False
language_model: Trace2333/joint_filtered_llama2_v0
model_size: 7B
ranking_group: single
us_pacific_date: 2024-07-04
win_ratio: 0.38900774621910733
generation_params: {'temperature': 1.25, 'top_p': 1.0, 'min_p': 0.12, 'top_k': 200, 'presence_penalty': 0.2, 'frequency_penalty': 0.2, 'stopping_words': ['</s>'], 'max_input_tokens': 512, 'best_of': 16, 'max_output_tokens': 64}
formatter: {'memory_template': "{bot_name}'s Persona####: {memory}", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': 'You: {message}\\\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
reward_formatter: {'bot_template': '<|im_start|>assistant\n{bot_name}: {message}<|im_end|>\n', 'memory_template': '<|im_start|>system\n{memory}<|im_end|>\n', 'prompt_template': '<|im_start|>user\n{prompt}<|im_end|>\n', 'response_template': '{bot_name}:', 'truncate_by_message': True, 'user_template': '<|im_start|>user\n{user_name}: {message}<|im_end|>\n'}
Resubmit model
Running pipeline stage MKMLizer
Starting job with name trace2333-joint-filtered-3791-v5-mkmlizer
Waiting for job on trace2333-joint-filtered-3791-v5-mkmlizer to finish
trace2333-joint-filtered-3791-v5-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
trace2333-joint-filtered-3791-v5-mkmlizer: ║ _____ __ __ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ /___/ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Version: 0.8.14 ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ https://mk1.ai ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ The license key for the current software has been verified as ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ belonging to: ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Chai Research Corp. ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Expiration: 2024-07-15 23:59:59 ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
trace2333-joint-filtered-3791-v5-mkmlizer: Downloaded to shared memory in 73.564s
trace2333-joint-filtered-3791-v5-mkmlizer: quantizing model to /dev/shm/model_cache
trace2333-joint-filtered-3791-v5-mkmlizer: Saving flywheel model at /dev/shm/model_cache
trace2333-joint-filtered-3791-v5-mkmlizer: quantized model in 38.842s
trace2333-joint-filtered-3791-v5-mkmlizer: Processed model Trace2333/joint_filtered_llama2_v0 in 112.406s
trace2333-joint-filtered-3791-v5-mkmlizer: creating bucket guanaco-mkml-models
trace2333-joint-filtered-3791-v5-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
trace2333-joint-filtered-3791-v5-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/special_tokens_map.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/tokenizer_config.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/config.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/tokenizer.model s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/tokenizer.model
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/tokenizer.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/flywheel_model.0.safetensors
trace2333-joint-filtered-3791-v5-mkmlizer: loading reward model from ChaiML/reward_gpt2_medium_preference_24m_e2
trace2333-joint-filtered-3791-v5-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 1%| | 3/291 [00:00<00:33, 8.67it/s] Loading 0: 1%|▏ | 4/291 [00:00<00:57, 4.98it/s] Loading 0: 2%|▏ | 5/291 [00:01<01:09, 4.14it/s] Loading 0: 2%|▏ | 7/291 [00:01<00:44, 6.36it/s] Loading 0: 3%|▎ | 8/291 [00:01<00:41, 6.82it/s] Loading 0: 3%|▎ | 9/291 [00:01<00:38, 7.24it/s] Loading 0: 3%|▎ | 10/291 [00:01<00:37, 7.45it/s] Loading 0: 4%|▍ | 12/291 [00:01<00:39, 7.03it/s] Loading 0: 4%|▍ | 13/291 [00:02<00:46, 5.94it/s] Loading 0: 5%|▍ | 14/291 [00:02<00:52, 5.24it/s] Loading 0: 6%|▌ | 17/291 [00:02<00:34, 8.05it/s] Loading 0: 7%|▋ | 19/291 [00:02<00:27, 9.72it/s] Loading 0: 7%|▋ | 21/291 [00:02<00:25, 10.41it/s] Loading 0: 8%|▊ | 23/291 [00:03<00:25, 10.47it/s] Loading 0: 11%|█ | 31/291 [00:03<00:11, 22.73it/s] Loading 0: 13%|█▎ | 39/291 [00:03<00:07, 33.33it/s] Loading 0: 15%|█▌ | 45/291 [00:03<00:06, 38.76it/s] Loading 0: 18%|█▊ | 51/291 [00:03<00:09, 26.54it/s] Loading 0: 19%|█▉ | 55/291 [00:03<00:08, 28.21it/s] Loading 0: 22%|██▏ | 63/291 [00:03<00:06, 37.70it/s] Loading 0: 23%|██▎ | 68/291 [00:04<00:05, 39.63it/s] Loading 0: 25%|██▌ | 73/291 [00:04<00:05, 41.13it/s] Loading 0: 27%|██▋ | 80/291 [00:04<00:04, 45.35it/s] Loading 0: 30%|███ | 88/291 [00:04<00:03, 50.84it/s] Loading 0: 32%|███▏ | 94/291 [00:04<00:03, 51.70it/s] Loading 0: 34%|███▍ | 100/291 [00:04<00:03, 50.94it/s] Loading 0: 36%|███▋ | 106/291 [00:05<00:05, 31.30it/s] Loading 0: 38%|███▊ | 112/291 [00:05<00:05, 35.05it/s] Loading 0: 41%|████ | 120/291 [00:05<00:04, 41.83it/s] Loading 0: 44%|████▎ | 127/291 [00:05<00:03, 46.83it/s] Loading 0: 46%|████▌ | 134/291 [00:05<00:03, 51.40it/s] Loading 0: 49%|████▉ | 142/291 [00:05<00:02, 57.10it/s] Loading 0: 51%|█████ | 149/291 [00:05<00:02, 57.47it/s] Loading 0: 54%|█████▍ | 158/291 [00:05<00:02, 64.44it/s] Loading 0: 57%|█████▋ | 165/291 [00:06<00:03, 35.34it/s] Loading 0: 59%|█████▉ | 171/291 [00:06<00:03, 38.78it/s] Loading 0: 61%|██████ | 177/291 [00:06<00:02, 41.12it/s] Loading 0: 63%|██████▎ | 184/291 [00:06<00:02, 44.80it/s] Loading 0: 66%|██████▌ | 192/291 [00:06<00:01, 49.81it/s] Loading 0: 68%|██████▊ | 198/291 [00:06<00:01, 51.34it/s] Loading 0: 70%|███████ | 204/291 [00:06<00:01, 50.76it/s] Loading 0: 73%|███████▎ | 212/291 [00:07<00:01, 56.69it/s] Loading 0: 75%|███████▍ | 218/291 [00:07<00:02, 31.38it/s] Loading 0: 77%|███████▋ | 223/291 [00:07<00:02, 31.01it/s] Loading 0: 79%|███████▊ | 229/291 [00:07<00:01, 35.44it/s] Loading 0: 81%|████████▏ | 237/291 [00:07<00:01, 40.47it/s] Loading 0: 83%|████████▎ | 242/291 [00:08<00:01, 41.87it/s] Loading 0: 85%|████████▍ | 247/291 [00:08<00:01, 43.54it/s] Loading 0: 88%|████████▊ | 255/291 [00:08<00:00, 49.13it/s] Loading 0: 90%|████████▉ | 261/291 [00:08<00:00, 50.19it/s] Loading 0: 92%|█████████▏| 267/291 [00:10<00:02, 9.45it/s] Loading 0: 94%|█████████▍| 274/291 [00:10<00:01, 12.92it/s] Loading 0: 96%|█████████▌| 279/291 [00:10<00:00, 15.83it/s] Loading 0: 98%|█████████▊| 284/291 [00:10<00:00, 19.26it/s] /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:919: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
trace2333-joint-filtered-3791-v5-mkmlizer: warnings.warn(
trace2333-joint-filtered-3791-v5-mkmlizer: /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
trace2333-joint-filtered-3791-v5-mkmlizer: warnings.warn(
trace2333-joint-filtered-3791-v5-mkmlizer: Traceback (most recent call last):
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/connection.py", line 174, in _new_conn
trace2333-joint-filtered-3791-v5-mkmlizer: conn = connection.create_connection(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/util/connection.py", line 72, in create_connection
trace2333-joint-filtered-3791-v5-mkmlizer: for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/socket.py", line 955, in getaddrinfo
trace2333-joint-filtered-3791-v5-mkmlizer: for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
trace2333-joint-filtered-3791-v5-mkmlizer: socket.gaierror: [Errno -3] Temporary failure in name resolution
trace2333-joint-filtered-3791-v5-mkmlizer: During handling of the above exception, another exception occurred:
trace2333-joint-filtered-3791-v5-mkmlizer: Traceback (most recent call last):
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/connectionpool.py", line 714, in urlopen
trace2333-joint-filtered-3791-v5-mkmlizer: httplib_response = self._make_request(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/connectionpool.py", line 403, in _make_request
trace2333-joint-filtered-3791-v5-mkmlizer: self._validate_conn(conn)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/connectionpool.py", line 1053, in _validate_conn
trace2333-joint-filtered-3791-v5-mkmlizer: conn.connect()
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/connection.py", line 363, in connect
trace2333-joint-filtered-3791-v5-mkmlizer: self.sock = conn = self._new_conn()
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/connection.py", line 186, in _new_conn
trace2333-joint-filtered-3791-v5-mkmlizer: raise NewConnectionError(
trace2333-joint-filtered-3791-v5-mkmlizer: urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7fbb4b906290>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution
trace2333-joint-filtered-3791-v5-mkmlizer: During handling of the above exception, another exception occurred:
trace2333-joint-filtered-3791-v5-mkmlizer: Traceback (most recent call last):
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/requests/adapters.py", line 486, in send
trace2333-joint-filtered-3791-v5-mkmlizer: resp = conn.urlopen(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/connectionpool.py", line 798, in urlopen
trace2333-joint-filtered-3791-v5-mkmlizer: retries = retries.increment(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/urllib3/util/retry.py", line 592, in increment
trace2333-joint-filtered-3791-v5-mkmlizer: raise MaxRetryError(_pool, url, error or ResponseError(cause))
trace2333-joint-filtered-3791-v5-mkmlizer: urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /ChaiML/reward_gpt2_medium_preference_24m_e2/resolve/main/config.json (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7fbb4b906290>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution'))
trace2333-joint-filtered-3791-v5-mkmlizer: During handling of the above exception, another exception occurred:
trace2333-joint-filtered-3791-v5-mkmlizer: Traceback (most recent call last):
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1722, in _get_metadata_or_catch_error
trace2333-joint-filtered-3791-v5-mkmlizer: metadata = get_hf_file_metadata(url=url, proxies=proxies, timeout=etag_timeout, headers=headers)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
trace2333-joint-filtered-3791-v5-mkmlizer: return fn(*args, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1645, in get_hf_file_metadata
trace2333-joint-filtered-3791-v5-mkmlizer: r = _request_wrapper(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 372, in _request_wrapper
trace2333-joint-filtered-3791-v5-mkmlizer: response = _request_wrapper(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 395, in _request_wrapper
trace2333-joint-filtered-3791-v5-mkmlizer: response = get_session().request(method=method, url=url, **params)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/requests/sessions.py", line 589, in request
trace2333-joint-filtered-3791-v5-mkmlizer: resp = self.send(prep, **send_kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/requests/sessions.py", line 703, in send
trace2333-joint-filtered-3791-v5-mkmlizer: r = adapter.send(request, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_http.py", line 66, in send
trace2333-joint-filtered-3791-v5-mkmlizer: return super().send(request, *args, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/requests/adapters.py", line 519, in send
trace2333-joint-filtered-3791-v5-mkmlizer: raise ConnectionError(e, request=request)
trace2333-joint-filtered-3791-v5-mkmlizer: requests.exceptions.ConnectionError: (MaxRetryError("HTTPSConnectionPool(host='huggingface.co', port=443): Max retries exceeded with url: /ChaiML/reward_gpt2_medium_preference_24m_e2/resolve/main/config.json (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7fbb4b906290>: Failed to establish a new connection: [Errno -3] Temporary failure in name resolution'))"), '(Request ID: b7c19126-c56e-4dab-9e7b-13d3477355b3)')
trace2333-joint-filtered-3791-v5-mkmlizer: The above exception was the direct cause of the following exception:
trace2333-joint-filtered-3791-v5-mkmlizer: Traceback (most recent call last):
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/transformers/utils/hub.py", line 399, in cached_file
trace2333-joint-filtered-3791-v5-mkmlizer: resolved_file = hf_hub_download(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
trace2333-joint-filtered-3791-v5-mkmlizer: return fn(*args, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1221, in hf_hub_download
trace2333-joint-filtered-3791-v5-mkmlizer: return _hf_hub_download_to_cache_dir(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1325, in _hf_hub_download_to_cache_dir
trace2333-joint-filtered-3791-v5-mkmlizer: _raise_on_head_call_error(head_call_error, force_download, local_files_only)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1826, in _raise_on_head_call_error
trace2333-joint-filtered-3791-v5-mkmlizer: raise LocalEntryNotFoundError(
trace2333-joint-filtered-3791-v5-mkmlizer: huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.
trace2333-joint-filtered-3791-v5-mkmlizer: The above exception was the direct cause of the following exception:
trace2333-joint-filtered-3791-v5-mkmlizer: Traceback (most recent call last):
trace2333-joint-filtered-3791-v5-mkmlizer: File "/code/uploading/reward.py", line 66, in <module>
trace2333-joint-filtered-3791-v5-mkmlizer: cli()
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1128, in __call__
trace2333-joint-filtered-3791-v5-mkmlizer: return self.main(*args, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1053, in main
trace2333-joint-filtered-3791-v5-mkmlizer: rv = self.invoke(ctx)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1659, in invoke
trace2333-joint-filtered-3791-v5-mkmlizer: return _process_result(sub_ctx.command.invoke(sub_ctx))
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1395, in invoke
trace2333-joint-filtered-3791-v5-mkmlizer: return ctx.invoke(self.callback, **ctx.params)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 754, in invoke
trace2333-joint-filtered-3791-v5-mkmlizer: return __callback(*args, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/code/uploading/reward.py", line 29, in tensorize_reward_model
trace2333-joint-filtered-3791-v5-mkmlizer: model, config, tokenizer = download_from_huggingface(repo_id, revision, hf_auth_token)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/code/uploading/reward.py", line 36, in download_from_huggingface
trace2333-joint-filtered-3791-v5-mkmlizer: config = AutoConfig.from_pretrained(model, revision=revision, use_auth_token=token)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained
trace2333-joint-filtered-3791-v5-mkmlizer: config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/transformers/configuration_utils.py", line 632, in get_config_dict
trace2333-joint-filtered-3791-v5-mkmlizer: config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/transformers/configuration_utils.py", line 689, in _get_config_dict
trace2333-joint-filtered-3791-v5-mkmlizer: resolved_config_file = cached_file(
trace2333-joint-filtered-3791-v5-mkmlizer: File "/opt/conda/lib/python3.10/site-packages/transformers/utils/hub.py", line 442, in cached_file
trace2333-joint-filtered-3791-v5-mkmlizer: raise EnvironmentError(
trace2333-joint-filtered-3791-v5-mkmlizer: OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like ChaiML/reward_gpt2_medium_preference_24m_e2 is not the path to a directory containing a file named config.json.
trace2333-joint-filtered-3791-v5-mkmlizer: Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.
Job trace2333-joint-filtered-3791-v5-mkmlizer completed after 148.32s with status: failed
Stopping job with name trace2333-joint-filtered-3791-v5-mkmlizer
%s, retrying in %s seconds...
Starting job with name trace2333-joint-filtered-3791-v5-mkmlizer
Waiting for job on trace2333-joint-filtered-3791-v5-mkmlizer to finish
trace2333-joint-filtered-3791-v5-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
trace2333-joint-filtered-3791-v5-mkmlizer: ║ _____ __ __ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ /___/ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Version: 0.8.14 ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ https://mk1.ai ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ The license key for the current software has been verified as ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ belonging to: ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Chai Research Corp. ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ Expiration: 2024-07-15 23:59:59 ║
trace2333-joint-filtered-3791-v5-mkmlizer: ║ ║
trace2333-joint-filtered-3791-v5-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
trace2333-joint-filtered-3791-v5-mkmlizer: Downloaded to shared memory in 62.888s
trace2333-joint-filtered-3791-v5-mkmlizer: quantizing model to /dev/shm/model_cache
trace2333-joint-filtered-3791-v5-mkmlizer: Saving flywheel model at /dev/shm/model_cache
trace2333-joint-filtered-3791-v5-mkmlizer: quantized model in 17.313s
trace2333-joint-filtered-3791-v5-mkmlizer: Processed model Trace2333/joint_filtered_llama2_v0 in 80.202s
trace2333-joint-filtered-3791-v5-mkmlizer: creating bucket guanaco-mkml-models
trace2333-joint-filtered-3791-v5-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 2%|▏ | 7/291 [00:00<00:04, 63.57it/s] Loading 0: 5%|▍ | 14/291 [00:00<00:05, 53.02it/s] Loading 0: 8%|▊ | 22/291 [00:00<00:04, 57.56it/s] Loading 0: 11%|█ | 31/291 [00:00<00:04, 63.05it/s] Loading 0: 13%|█▎ | 39/291 [00:00<00:03, 67.74it/s] Loading 0: 16%|█▌ | 46/291 [00:00<00:03, 67.89it/s] Loading 0: 18%|█▊ | 53/291 [00:01<00:06, 34.68it/s] Loading 0: 21%|██ | 61/291 [00:01<00:05, 42.47it/s] Loading 0: 23%|██▎ | 67/291 [00:01<00:05, 44.41it/s] Loading 0: 25%|██▌ | 74/291 [00:01<00:04, 49.40it/s] Loading 0: 28%|██▊ | 81/291 [00:01<00:04, 51.33it/s] Loading 0: 31%|███ | 89/291 [00:01<00:03, 56.75it/s] Loading 0: 33%|███▎ | 97/291 [00:01<00:03, 61.68it/s] Loading 0: 36%|███▌ | 104/291 [00:01<00:02, 63.17it/s] Loading 0: 38%|███▊ | 111/291 [00:02<00:05, 33.00it/s] Loading 0: 41%|████ | 118/291 [00:02<00:04, 38.76it/s] Loading 0: 43%|████▎ | 124/291 [00:02<00:03, 42.62it/s] Loading 0: 45%|████▌ | 131/291 [00:02<00:03, 46.27it/s] Loading 0: 48%|████▊ | 139/291 [00:02<00:02, 53.33it/s] Loading 0: 51%|█████ | 147/291 [00:02<00:02, 59.34it/s] Loading 0: 53%|█████▎ | 154/291 [00:03<00:02, 61.57it/s] Loading 0: 55%|█████▌ | 161/291 [00:03<00:03, 36.07it/s] Loading 0: 57%|█████▋ | 167/291 [00:03<00:03, 39.54it/s] Loading 0: 60%|██████ | 176/291 [00:03<00:02, 47.27it/s] Loading 0: 63%|██████▎ | 184/291 [00:03<00:01, 53.72it/s] Loading 0: 66%|██████▌ | 192/291 [00:03<00:01, 59.52it/s] Loading 0: 69%|██████▉ | 201/291 [00:03<00:01, 63.45it/s] Loading 0: 72%|███████▏ | 209/291 [00:04<00:01, 66.11it/s] Loading 0: 75%|███████▍ | 217/291 [00:04<00:01, 39.18it/s] Loading 0: 77%|███████▋ | 224/291 [00:04<00:01, 43.74it/s] Loading 0: 79%|███████▉ | 230/291 [00:04<00:01, 46.06it/s] Loading 0: 82%|████████▏ | 238/291 [00:04<00:00, 53.01it/s] Loading 0: 85%|████████▍ | 246/291 [00:04<00:00, 58.78it/s] Loading 0: 87%|████████▋ | 253/291 [00:05<00:00, 60.57it/s] Loading 0: 89%|████████▉ | 260/291 [00:05<00:00, 61.00it/s] Loading 0: 92%|█████████▏| 267/291 [00:06<00:02, 11.26it/s] Loading 0: 94%|█████████▍| 274/291 [00:07<00:01, 14.75it/s] Loading 0: 97%|█████████▋| 281/291 [00:07<00:00, 19.17it/s] Loading 0: 99%|█████████▊| 287/291 [00:07<00:00, 23.39it/s] ERROR: [Errno -3] Temporary failure in name resolution
trace2333-joint-filtered-3791-v5-mkmlizer: ERROR: Connection Error: Error resolving a server hostname.
trace2333-joint-filtered-3791-v5-mkmlizer: Please check the servers address specified in 'host_base', 'host_bucket', 'cloudfront_host', 'website_endpoint'
Job trace2333-joint-filtered-3791-v5-mkmlizer completed after 106.73s with status: failed
Stopping job with name trace2333-joint-filtered-3791-v5-mkmlizer
%s, retrying in %s seconds...
Starting job with name trace2333-joint-filtered-3791-v5-mkmlizer
Waiting for job on trace2333-joint-filtered-3791-v5-mkmlizer to finish
Retrying (%r) after connection broken by '%r': %s
trace2333-joint-filtered-3791-v5-mkmlizer: Downloaded to shared memory in 63.063s
trace2333-joint-filtered-3791-v5-mkmlizer: quantizing model to /dev/shm/model_cache
trace2333-joint-filtered-3791-v5-mkmlizer: Saving flywheel model at /dev/shm/model_cache
trace2333-joint-filtered-3791-v5-mkmlizer: quantized model in 18.771s
trace2333-joint-filtered-3791-v5-mkmlizer: Processed model Trace2333/joint_filtered_llama2_v0 in 81.834s
trace2333-joint-filtered-3791-v5-mkmlizer: creating bucket guanaco-mkml-models
trace2333-joint-filtered-3791-v5-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
trace2333-joint-filtered-3791-v5-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/config.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/special_tokens_map.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/tokenizer_config.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/tokenizer.model s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/tokenizer.model
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/tokenizer.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/trace2333-joint-filtered-3791-v5/flywheel_model.0.safetensors
trace2333-joint-filtered-3791-v5-mkmlizer: loading reward model from ChaiML/reward_gpt2_medium_preference_24m_e2
trace2333-joint-filtered-3791-v5-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 2%|▏ | 5/291 [00:00<00:06, 43.29it/s] Loading 0: 4%|▍ | 13/291 [00:00<00:04, 61.93it/s] Loading 0: 7%|▋ | 21/291 [00:00<00:03, 67.98it/s] Loading 0: 10%|▉ | 28/291 [00:00<00:03, 67.53it/s] Loading 0: 12%|█▏ | 35/291 [00:00<00:03, 65.46it/s] Loading 0: 14%|█▍ | 42/291 [00:00<00:03, 63.88it/s] Loading 0: 18%|█▊ | 51/291 [00:01<00:06, 37.98it/s] Loading 0: 20%|█▉ | 57/291 [00:01<00:05, 40.61it/s] Loading 0: 23%|██▎ | 66/291 [00:01<00:04, 48.05it/s] Loading 0: 25%|██▍ | 72/291 [00:01<00:04, 47.78it/s] Loading 0: 27%|██▋ | 80/291 [00:01<00:03, 54.44it/s] Loading 0: 30%|███ | 88/291 [00:01<00:03, 60.04it/s] Loading 0: 33%|███▎ | 95/291 [00:01<00:03, 61.92it/s] Loading 0: 35%|███▌ | 102/291 [00:01<00:03, 62.05it/s] Loading 0: 37%|███▋ | 109/291 [00:02<00:05, 32.60it/s] Loading 0: 39%|███▉ | 114/291 [00:02<00:05, 31.71it/s] Loading 0: 41%|████ | 120/291 [00:02<00:05, 34.10it/s] Loading 0: 43%|████▎ | 125/291 [00:02<00:04, 33.62it/s] Loading 0: 45%|████▍ | 130/291 [00:02<00:04, 35.32it/s] Loading 0: 47%|████▋ | 138/291 [00:03<00:03, 43.29it/s] Loading 0: 49%|████▉ | 144/291 [00:03<00:03, 46.82it/s] Loading 0: 52%|█████▏ | 150/291 [00:03<00:02, 48.71it/s] Loading 0: 55%|█████▍ | 159/291 [00:03<00:04, 29.61it/s] Loading 0: 56%|█████▋ | 164/291 [00:03<00:04, 31.51it/s] Loading 0: 58%|█████▊ | 169/291 [00:03<00:03, 34.02it/s] Loading 0: 60%|██████ | 176/291 [00:04<00:02, 38.70it/s] Loading 0: 63%|██████▎ | 184/291 [00:04<00:02, 46.25it/s] Loading 0: 66%|██████▌ | 192/291 [00:04<00:01, 52.56it/s] Loading 0: 68%|██████▊ | 199/291 [00:04<00:01, 55.27it/s] Loading 0: 71%|███████ | 206/291 [00:04<00:01, 56.32it/s] Loading 0: 73%|███████▎ | 213/291 [00:04<00:02, 33.15it/s] Loading 0: 75%|███████▍ | 218/291 [00:05<00:02, 33.03it/s] Loading 0: 77%|███████▋ | 223/291 [00:05<00:01, 35.32it/s] Loading 0: 78%|███████▊ | 228/291 [00:05<00:01, 37.77it/s] Loading 0: 81%|████████ | 235/291 [00:05<00:01, 43.36it/s] Loading 0: 82%|████████▏ | 240/291 [00:05<00:01, 39.54it/s] Loading 0: 85%|████████▍ | 247/291 [00:05<00:00, 45.06it/s] Loading 0: 88%|████████▊ | 255/291 [00:05<00:00, 52.43it/s] Loading 0: 90%|████████▉ | 261/291 [00:05<00:00, 53.08it/s] Loading 0: 92%|█████████▏| 267/291 [00:07<00:02, 9.62it/s] Loading 0: 93%|█████████▎| 272/291 [00:07<00:01, 11.80it/s] Loading 0: 96%|█████████▌| 278/291 [00:08<00:00, 15.41it/s] Loading 0: 97%|█████████▋| 283/291 [00:08<00:00, 18.72it/s] Loading 0: 99%|█████████▉| 288/291 [00:08<00:00, 21.44it/s] /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:919: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
trace2333-joint-filtered-3791-v5-mkmlizer: warnings.warn(
trace2333-joint-filtered-3791-v5-mkmlizer: /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
trace2333-joint-filtered-3791-v5-mkmlizer: warnings.warn(
trace2333-joint-filtered-3791-v5-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:769: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
trace2333-joint-filtered-3791-v5-mkmlizer: warnings.warn(
trace2333-joint-filtered-3791-v5-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:468: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
trace2333-joint-filtered-3791-v5-mkmlizer: warnings.warn(
trace2333-joint-filtered-3791-v5-mkmlizer: /opt/conda/lib/python3.10/site-packages/torch/_utils.py:831: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage()
trace2333-joint-filtered-3791-v5-mkmlizer: return self.fget.__get__(instance, owner)()
trace2333-joint-filtered-3791-v5-mkmlizer: Saving model to /tmp/reward_cache/reward.tensors
trace2333-joint-filtered-3791-v5-mkmlizer: Saving duration: 0.429s
trace2333-joint-filtered-3791-v5-mkmlizer: Processed model ChaiML/reward_gpt2_medium_preference_24m_e2 in 12.837s
trace2333-joint-filtered-3791-v5-mkmlizer: creating bucket guanaco-reward-models
trace2333-joint-filtered-3791-v5-mkmlizer: Bucket 's3://guanaco-reward-models/' created
trace2333-joint-filtered-3791-v5-mkmlizer: uploading /tmp/reward_cache to s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward
trace2333-joint-filtered-3791-v5-mkmlizer: cp /tmp/reward_cache/config.json s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward/config.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /tmp/reward_cache/special_tokens_map.json s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward/special_tokens_map.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /tmp/reward_cache/tokenizer_config.json s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward/tokenizer_config.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /tmp/reward_cache/merges.txt s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward/merges.txt
trace2333-joint-filtered-3791-v5-mkmlizer: cp /tmp/reward_cache/vocab.json s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward/vocab.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /tmp/reward_cache/tokenizer.json s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward/tokenizer.json
trace2333-joint-filtered-3791-v5-mkmlizer: cp /tmp/reward_cache/reward.tensors s3://guanaco-reward-models/trace2333-joint-filtered-3791-v5_reward/reward.tensors
Job trace2333-joint-filtered-3791-v5-mkmlizer completed after 127.06s with status: succeeded
Stopping job with name trace2333-joint-filtered-3791-v5-mkmlizer
Pipeline stage MKMLizer completed in 385.47s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.16s
Running pipeline stage ISVCDeployer
Creating inference service trace2333-joint-filtered-3791-v5
Waiting for inference service trace2333-joint-filtered-3791-v5 to be ready
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
Connection pool is full, discarding connection: %s
Inference service trace2333-joint-filtered-3791-v5 ready after 50.42999315261841s
Pipeline stage ISVCDeployer completed in 57.44s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.1167852878570557s
Received healthy response to inference request in 1.292888879776001s
Received healthy response to inference request in 6.876341819763184s
Received healthy response to inference request in 1.2842001914978027s
Received healthy response to inference request in 1.9618785381317139s
5 requests
0 failed requests
5th percentile: 1.2859379291534423
10th percentile: 1.287675666809082
20th percentile: 1.2911511421203614
30th percentile: 1.4266868114471436
40th percentile: 1.6942826747894288
50th percentile: 1.9618785381317139
60th percentile: 2.0238412380218507
70th percentile: 2.0858039379119875
80th percentile: 3.0686965942382822
90th percentile: 4.972519207000733
95th percentile: 5.9244305133819575
99th percentile: 6.685959558486938
mean time: 2.7064189434051515
Pipeline stage StressChecker completed in 14.88s
trace2333-joint-filtered_3791_v5 status is now deployed due to DeploymentManager action
trace2333-joint-filtered_3791_v5 status is now inactive due to auto deactivation removed underperforming models
admin requested tearing down of trace2333-joint-filtered_3791_v5
Running pipeline stage ISVCDeleter
Checking if service trace2333-joint-filtered-3791-v5 is running
Skipping teardown as no inference service was found
Pipeline stage ISVCDeleter completed in 5.18s
Running pipeline stage MKMLModelDeleter
Cleaning model data from S3
Cleaning model data from model cache
Deleting key trace2333-joint-filtered-3791-v5/config.json from bucket guanaco-mkml-models
Deleting key trace2333-joint-filtered-3791-v5/flywheel_model.0.safetensors from bucket guanaco-mkml-models
Deleting key trace2333-joint-filtered-3791-v5/special_tokens_map.json from bucket guanaco-mkml-models
Deleting key trace2333-joint-filtered-3791-v5/tokenizer.json from bucket guanaco-mkml-models
Deleting key trace2333-joint-filtered-3791-v5/tokenizer.model from bucket guanaco-mkml-models
Deleting key trace2333-joint-filtered-3791-v5/tokenizer_config.json from bucket guanaco-mkml-models
Cleaning model data from model cache
Deleting key trace2333-joint-filtered-3791-v5_reward/config.json from bucket guanaco-reward-models
Deleting key trace2333-joint-filtered-3791-v5_reward/merges.txt from bucket guanaco-reward-models
Deleting key trace2333-joint-filtered-3791-v5_reward/reward.tensors from bucket guanaco-reward-models
Deleting key trace2333-joint-filtered-3791-v5_reward/special_tokens_map.json from bucket guanaco-reward-models
Deleting key trace2333-joint-filtered-3791-v5_reward/tokenizer.json from bucket guanaco-reward-models
Deleting key trace2333-joint-filtered-3791-v5_reward/tokenizer_config.json from bucket guanaco-reward-models
Deleting key trace2333-joint-filtered-3791-v5_reward/vocab.json from bucket guanaco-reward-models
Pipeline stage MKMLModelDeleter completed in 6.97s
trace2333-joint-filtered_3791_v5 status is now torndown due to DeploymentManager action