submission_id: nousresearch-meta-llama_4941_v54
developer_uid: chai_backend_admin
status: torndown
model_repo: NousResearch/Meta-Llama-3-8B-Instruct
reward_repo: ChaiML/reward_gpt2_medium_preference_24m_e2
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.0, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n'], 'max_input_tokens': 1024, 'best_of': 4, 'max_output_tokens': 300}
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
reward_formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
timestamp: 2024-05-14T23:59:40+00:00
model_name: nousresearch-meta-llama_4941_v54
model_eval_status: success
model_group: NousResearch/Meta-Llama-
num_battles: 16912
num_wins: 8346
celo_rating: 1184.54
propriety_score: 0.0
propriety_total_count: 0.0
submission_type: basic
model_architecture: LlamaForCausalLM
model_num_parameters: 8030261248.0
best_of: 4
max_input_tokens: 1024
max_output_tokens: 300
display_name: nousresearch-meta-llama_4941_v54
ineligible_reason: max_output_tokens!=64
language_model: NousResearch/Meta-Llama-3-8B-Instruct
model_size: 8B
reward_model: ChaiML/reward_gpt2_medium_preference_24m_e2
us_pacific_date: 2024-05-14
win_ratio: 0.4934957426679281
preference_data_url: None
Resubmit model
Running pipeline stage MKMLizer
Starting job with name nousresearch-meta-llama-4941-v54-mkmlizer
Waiting for job on nousresearch-meta-llama-4941-v54-mkmlizer to finish
nousresearch-meta-llama-4941-v54-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
nousresearch-meta-llama-4941-v54-mkmlizer: ║ _____ __ __ ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ /___/ ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ Version: 0.8.14 ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ https://mk1.ai ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ The license key for the current software has been verified as ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ belonging to: ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ Chai Research Corp. ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ Expiration: 2024-07-15 23:59:59 ║
nousresearch-meta-llama-4941-v54-mkmlizer: ║ ║
nousresearch-meta-llama-4941-v54-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
nousresearch-meta-llama-4941-v54-mkmlizer: /opt/conda/lib/python3.10/site-packages/huggingface_hub/utils/_deprecation.py:131: FutureWarning: 'list_files_info' (from 'huggingface_hub.hf_api') is deprecated and will be removed from version '0.23'. Use `list_repo_tree` and `get_paths_info` instead.
nousresearch-meta-llama-4941-v54-mkmlizer: warnings.warn(warning_message, FutureWarning)
nousresearch-meta-llama-4941-v54-mkmlizer: Downloaded to shared memory in 22.348s
nousresearch-meta-llama-4941-v54-mkmlizer: quantizing model to /dev/shm/model_cache
nousresearch-meta-llama-4941-v54-mkmlizer: Saving flywheel model at /dev/shm/model_cache
nousresearch-meta-llama-4941-v54-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 8%|▊ | 22/291 [00:00<00:01, 201.50it/s] Loading 0: 15%|█▌ | 44/291 [00:00<00:01, 210.00it/s] Loading 0: 23%|██▎ | 66/291 [00:00<00:01, 209.43it/s] Loading 0: 30%|██▉ | 87/291 [00:00<00:02, 92.22it/s] Loading 0: 35%|███▌ | 102/291 [00:00<00:01, 98.98it/s] Loading 0: 40%|███▉ | 116/291 [00:00<00:01, 106.96it/s] Loading 0: 45%|████▍ | 130/291 [00:01<00:01, 113.08it/s] Loading 0: 51%|█████ | 148/291 [00:01<00:01, 128.75it/s] Loading 0: 57%|█████▋ | 167/291 [00:01<00:00, 142.97it/s] Loading 0: 64%|██████▍ | 187/291 [00:01<00:01, 95.04it/s] Loading 0: 70%|██████▉ | 203/291 [00:01<00:00, 105.92it/s] Loading 0: 78%|███████▊ | 228/291 [00:01<00:00, 133.88it/s] Loading 0: 85%|████████▌ | 248/291 [00:01<00:00, 146.63it/s] Loading 0: 93%|█████████▎| 270/291 [00:02<00:00, 164.30it/s] Loading 0: 99%|█████████▉| 289/291 [00:06<00:00, 12.89it/s] Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
nousresearch-meta-llama-4941-v54-mkmlizer: quantized model in 17.471s
nousresearch-meta-llama-4941-v54-mkmlizer: Processed model NousResearch/Meta-Llama-3-8B-Instruct in 40.798s
nousresearch-meta-llama-4941-v54-mkmlizer: creating bucket guanaco-mkml-models
nousresearch-meta-llama-4941-v54-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
nousresearch-meta-llama-4941-v54-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/nousresearch-meta-llama-4941-v54
nousresearch-meta-llama-4941-v54-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/nousresearch-meta-llama-4941-v54/tokenizer_config.json
nousresearch-meta-llama-4941-v54-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/nousresearch-meta-llama-4941-v54/special_tokens_map.json
nousresearch-meta-llama-4941-v54-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/nousresearch-meta-llama-4941-v54/config.json
nousresearch-meta-llama-4941-v54-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/nousresearch-meta-llama-4941-v54/tokenizer.json
nousresearch-meta-llama-4941-v54-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/nousresearch-meta-llama-4941-v54/flywheel_model.0.safetensors
nousresearch-meta-llama-4941-v54-mkmlizer: loading reward model from ChaiML/reward_gpt2_medium_preference_24m_e2
nousresearch-meta-llama-4941-v54-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:913: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
nousresearch-meta-llama-4941-v54-mkmlizer: warnings.warn(
nousresearch-meta-llama-4941-v54-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py:757: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
nousresearch-meta-llama-4941-v54-mkmlizer: warnings.warn(
nousresearch-meta-llama-4941-v54-mkmlizer: /opt/conda/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py:468: FutureWarning: The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.
nousresearch-meta-llama-4941-v54-mkmlizer: warnings.warn(
Job nousresearch-meta-llama-4941-v54-mkmlizer completed after 69.52s with status: succeeded
Stopping job with name nousresearch-meta-llama-4941-v54-mkmlizer
Pipeline stage MKMLizer completed in 71.10s
Running pipeline stage MKMLKubeTemplater
Pipeline stage MKMLKubeTemplater completed in 0.43s
Running pipeline stage ISVCDeployer
Creating inference service nousresearch-meta-llama-4941-v54
Waiting for inference service nousresearch-meta-llama-4941-v54 to be ready
Inference service nousresearch-meta-llama-4941-v54 ready after 40.63352823257446s
Pipeline stage ISVCDeployer completed in 47.13s
Running pipeline stage StressChecker
HTTP Request: %s %s "%s %d %s"
Received healthy response to inference request in 2.4288899898529053s
HTTP Request: %s %s "%s %d %s"
Received healthy response to inference request in 1.6834449768066406s
HTTP Request: %s %s "%s %d %s"
Received healthy response to inference request in 1.802610158920288s
HTTP Request: %s %s "%s %d %s"
Received healthy response to inference request in 1.6268060207366943s
HTTP Request: %s %s "%s %d %s"
Received healthy response to inference request in 1.775407075881958s
5 requests
0 failed requests
5th percentile: 1.6381338119506836
10th percentile: 1.649461603164673
20th percentile: 1.6721171855926513
30th percentile: 1.7018373966217042
40th percentile: 1.738622236251831
50th percentile: 1.775407075881958
60th percentile: 1.78628830909729
70th percentile: 1.7971695423126222
80th percentile: 1.9278661251068117
90th percentile: 2.1783780574798586
95th percentile: 2.303634023666382
99th percentile: 2.4038387966156005
mean time: 1.8634316444396972
Pipeline stage StressChecker completed in 12.74s
Running pipeline stage DaemonicModelEvalScorer
Pipeline stage DaemonicModelEvalScorer completed in 0.15s
Running pipeline stage DaemonicSafetyScorer
Pipeline stage DaemonicSafetyScorer completed in 0.15s
Running M-Eval for topic stay_in_character
%s, retrying in %s seconds...
nousresearch-meta-llama_4941_v54 status is now deployed due to DeploymentManager action
M-Eval Dataset for topic stay_in_character is loaded
%s, retrying in %s seconds...
nousresearch-meta-llama_4941_v54 status is now inactive due to auto deactivation removed underperforming models
admin requested tearing down of nousresearch-meta-llama_4941_v54
Running pipeline stage ISVCDeleter
Checking if service nousresearch-meta-llama-4941-v54 is running
Tearing down inference service nousresearch-meta-llama-4941-v54
Toredown service nousresearch-meta-llama-4941-v54
Pipeline stage ISVCDeleter completed in 3.16s
Running pipeline stage MKMLModelDeleter
Cleaning model data from S3
Cleaning model data from model cache
Deleting key nousresearch-meta-llama-4941-v54/config.json from bucket guanaco-mkml-models
Deleting key nousresearch-meta-llama-4941-v54/flywheel_model.0.safetensors from bucket guanaco-mkml-models
Deleting key nousresearch-meta-llama-4941-v54/special_tokens_map.json from bucket guanaco-mkml-models
Deleting key nousresearch-meta-llama-4941-v54/tokenizer.json from bucket guanaco-mkml-models
Deleting key nousresearch-meta-llama-4941-v54/tokenizer_config.json from bucket guanaco-mkml-models
Cleaning model data from model cache
Deleting key nousresearch-meta-llama-4941-v54_reward/config.json from bucket guanaco-reward-models
Deleting key nousresearch-meta-llama-4941-v54_reward/merges.txt from bucket guanaco-reward-models
Deleting key nousresearch-meta-llama-4941-v54_reward/reward.tensors from bucket guanaco-reward-models
Deleting key nousresearch-meta-llama-4941-v54_reward/special_tokens_map.json from bucket guanaco-reward-models
Deleting key nousresearch-meta-llama-4941-v54_reward/tokenizer.json from bucket guanaco-reward-models
Deleting key nousresearch-meta-llama-4941-v54_reward/tokenizer_config.json from bucket guanaco-reward-models
Deleting key nousresearch-meta-llama-4941-v54_reward/vocab.json from bucket guanaco-reward-models
Pipeline stage MKMLModelDeleter completed in 1.91s
nousresearch-meta-llama_4941_v54 status is now torndown due to DeploymentManager action

Usage Metrics

Latency Metrics