submission_id: sao10k-mn-12b-lyra-v4a1_v1
developer_uid: sao10k
alignment_samples: 10079
alignment_score: -0.4171370377119076
best_of: 8
celo_rating: 1258.72
display_name: lyra41
formatter: {'memory_template': '<|im_start|>system\n{memory}[/INST]\n', 'prompt_template': '<|im_start|>user\n{prompt}[/INST]\n', 'bot_template': '<|im_start|>assistant\n{bot_name}: {message}[/INST]\n', 'user_template': '<|im_start|>user\n{user_name}: {message}[/INST]\n', 'response_template': '[INST]assistant\n{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 0.75, 'top_p': 1.0, 'min_p': 0.1, 'top_k': 40, 'presence_penalty': 0.0, 'frequency_penalty': 0.0, 'stopping_words': ['\n', '\n\n', '\nYou:', '[/INST]', '<|im_end|>', '</s>'], 'max_input_tokens': 512, 'best_of': 8, 'max_output_tokens': 64}
gpu_counts: {'NVIDIA RTX A5000': 1}
is_internal_developer: False
language_model: Sao10K/MN-12B-Lyra-v4a1
latencies: [{'batch_size': 1, 'throughput': 0.6919619992049415, 'latency_mean': 1.445058172941208, 'latency_p50': 1.4457892179489136, 'latency_p90': 1.601093602180481}, {'batch_size': 3, 'throughput': 1.330256171230038, 'latency_mean': 2.247247760295868, 'latency_p50': 2.227228283882141, 'latency_p90': 2.504995322227478}, {'batch_size': 5, 'throughput': 1.5625558813529499, 'latency_mean': 3.1863015592098236, 'latency_p50': 3.169217348098755, 'latency_p90': 3.5948062658309934}, {'batch_size': 6, 'throughput': 1.597880025184406, 'latency_mean': 3.73540811419487, 'latency_p50': 3.7146109342575073, 'latency_p90': 4.2352792263031}, {'batch_size': 8, 'throughput': 1.5923812154336616, 'latency_mean': 4.987358438968658, 'latency_p50': 5.002313017845154, 'latency_p90': 5.6774814367294315}, {'batch_size': 10, 'throughput': 1.567853626823206, 'latency_mean': 6.342608182430268, 'latency_p50': 6.347144365310669, 'latency_p90': 7.311555075645447}]
max_input_tokens: 512
max_output_tokens: 64
model_architecture: MistralForCausalLM
model_group: Sao10K/MN-12B-Lyra-v4a1
model_name: lyra41
model_num_parameters: 12772070400.0
model_repo: Sao10K/MN-12B-Lyra-v4a1
model_size: 13B
num_battles: 10079
num_wins: 5287
propriety_score: 0.7510775862068966
propriety_total_count: 928.0
ranking_group: single
status: inactive
submission_type: basic
throughput_3p7s: 1.6
timestamp: 2024-09-05T15:45:51+00:00
us_pacific_date: 2024-09-05
win_ratio: 0.5245560075404306
Download Preference Data
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLizer
Starting job with name sao10k-mn-12b-lyra-v4a1-v1-mkmlizer
Waiting for job on sao10k-mn-12b-lyra-v4a1-v1-mkmlizer to finish
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ _____ __ __ ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ /___/ ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ Version: 0.10.1 ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ https://mk1.ai ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ The license key for the current software has been verified as ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ belonging to: ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ Chai Research Corp. ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ Expiration: 2024-10-15 23:59:59 ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ║ ║
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
Connection pool is full, discarding connection: %s. Connection pool size: %s
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: quantized model in 36.491s
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: Processed model Sao10K/MN-12B-Lyra-v4a1 in 84.412s
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: creating bucket guanaco-mkml-models
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/sao10k-mn-12b-lyra-v4a1-v1
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/sao10k-mn-12b-lyra-v4a1-v1/config.json
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/sao10k-mn-12b-lyra-v4a1-v1/special_tokens_map.json
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/sao10k-mn-12b-lyra-v4a1-v1/tokenizer_config.json
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/sao10k-mn-12b-lyra-v4a1-v1/tokenizer.json
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/sao10k-mn-12b-lyra-v4a1-v1/flywheel_model.0.safetensors
sao10k-mn-12b-lyra-v4a1-v1-mkmlizer: Loading 0: 0%| | 0/363 [00:00<?, ?it/s] Loading 0: 1%| | 2/363 [00:06<18:16, 3.04s/it] Loading 0: 2%|▏ | 6/363 [00:06<04:51, 1.22it/s] Loading 0: 3%|▎ | 11/363 [00:06<02:08, 2.75it/s] Loading 0: 4%|▍ | 15/363 [00:06<01:21, 4.29it/s] Loading 0: 6%|▌ | 22/363 [00:06<00:42, 8.01it/s] Loading 0: 7%|▋ | 27/363 [00:06<00:30, 11.04it/s] Loading 0: 9%|▉ | 32/363 [00:06<00:22, 14.73it/s] Loading 0: 10%|█ | 37/363 [00:06<00:17, 18.88it/s] Loading 0: 12%|█▏ | 42/363 [00:07<00:15, 20.57it/s] Loading 0: 13%|█▎ | 49/363 [00:07<00:11, 27.50it/s] Loading 0: 15%|█▍ | 54/363 [00:07<00:10, 30.07it/s] Loading 0: 16%|█▋ | 59/363 [00:07<00:12, 25.19it/s] Loading 0: 18%|█▊ | 64/363 [00:07<00:10, 29.32it/s] Loading 0: 19%|█▊ | 68/363 [00:07<00:09, 31.30it/s] Loading 0: 20%|██ | 74/363 [00:07<00:08, 34.50it/s] Loading 0: 22%|██▏ | 79/363 [00:08<00:07, 36.59it/s] Loading 0: 24%|██▎ | 86/363 [00:08<00:06, 43.28it/s] Loading 0: 25%|██▌ | 91/363 [00:08<00:06, 44.34it/s] Loading 0: 26%|██▋ | 96/363 [00:08<00:06, 38.52it/s] Loading 0: 29%|██▊ | 104/363 [00:08<00:05, 46.35it/s] Loading 0: 30%|███ | 109/363 [00:08<00:05, 46.98it/s] Loading 0: 31%|███▏ | 114/363 [00:08<00:06, 39.13it/s] Loading 0: 33%|███▎ | 121/363 [00:08<00:05, 45.65it/s] Loading 0: 35%|███▍ | 126/363 [00:09<00:05, 45.31it/s] Loading 0: 36%|███▌ | 131/363 [00:09<00:05, 45.71it/s] Loading 0: 37%|███▋ | 136/363 [00:09<00:05, 45.32it/s] Loading 0: 39%|███▉ | 141/363 [00:09<00:06, 35.88it/s] Loading 0: 41%|████ | 148/363 [00:09<00:05, 42.97it/s] Loading 0: 42%|████▏ | 153/363 [00:09<00:04, 42.72it/s] Loading 0: 44%|████▎ | 158/363 [00:10<00:06, 30.54it/s] Loading 0: 45%|████▍ | 163/363 [00:10<00:05, 33.95it/s] Loading 0: 46%|████▋ | 168/363 [00:10<00:06, 30.18it/s] Loading 0: 48%|████▊ | 175/363 [00:10<00:05, 37.03it/s] Loading 0: 50%|████▉ | 180/363 [00:10<00:04, 38.13it/s] Loading 0: 51%|█████ | 185/363 [00:10<00:04, 38.85it/s] Loading 0: 52%|█████▏ | 190/363 [00:10<00:04, 40.32it/s] Loading 0: 54%|█████▎ | 195/363 [00:11<00:05, 33.24it/s] Loading 0: 56%|█████▌ | 202/363 [00:11<00:03, 40.62it/s] Loading 0: 57%|█████▋ | 207/363 [00:11<00:03, 41.17it/s] Loading 0: 58%|█████▊ | 212/363 [00:11<00:03, 41.61it/s] Loading 0: 60%|█████▉ | 217/363 [00:11<00:03, 42.09it/s] Loading 0: 61%|██████ | 222/363 [00:11<00:04, 34.82it/s] Loading 0: 63%|██████▎ | 229/363 [00:11<00:03, 42.06it/s] Loading 0: 64%|██████▍ | 234/363 [00:11<00:03, 42.50it/s] Loading 0: 66%|██████▌ | 239/363 [00:12<00:02, 43.49it/s] Loading 0: 67%|██████▋ | 245/363 [00:12<00:02, 41.36it/s] Loading 0: 69%|██████▉ | 250/363 [00:12<00:02, 39.44it/s] Loading 0: 71%|███████ | 256/363 [00:12<00:03, 31.61it/s] Loading 0: 72%|███████▏ | 260/363 [00:12<00:03, 32.42it/s] Loading 0: 73%|███████▎ | 265/363 [00:12<00:02, 36.16it/s] Loading 0: 74%|███████▍ | 269/363 [00:12<00:02, 35.79it/s] Loading 0: 75%|███████▌ | 274/363 [00:13<00:02, 38.13it/s] Loading 0: 77%|███████▋ | 279/363 [00:13<00:02, 38.50it/s] Loading 0: 78%|███████▊ | 284/363 [00:13<00:01, 39.68it/s] Loading 0: 80%|███████▉ | 289/363 [00:13<00:01, 41.91it/s] Loading 0: 81%|████████ | 294/363 [00:13<00:01, 35.67it/s] Loading 0: 83%|████████▎ | 301/363 [00:13<00:01, 43.26it/s] Loading 0: 84%|████████▍ | 306/363 [00:13<00:01, 43.30it/s] Loading 0: 86%|████████▌ | 311/363 [00:13<00:01, 44.03it/s] Loading 0: 87%|████████▋ | 316/363 [00:13<00:01, 45.44it/s] Loading 0: 88%|████████▊ | 321/363 [00:14<00:01, 35.92it/s] Loading 0: 90%|█████████ | 328/363 [00:14<00:00, 42.14it/s] Loading 0: 92%|█████████▏| 333/363 [00:14<00:00, 41.80it/s] Loading 0: 93%|█████████▎| 338/363 [00:14<00:00, 40.45it/s] Loading 0: 94%|█████████▍| 343/363 [00:14<00:00, 42.25it/s] Loading 0: 96%|█████████▌| 348/363 [00:14<00:00, 36.19it/s] Loading 0: 98%|█████████▊| 355/363 [00:15<00:00, 31.17it/s] Loading 0: 99%|█████████▉| 359/363 [00:15<00:00, 31.81it/s]
Job sao10k-mn-12b-lyra-v4a1-v1-mkmlizer completed after 117.93s with status: succeeded
Stopping job with name sao10k-mn-12b-lyra-v4a1-v1-mkmlizer
Pipeline stage MKMLizer completed in 119.00s
run pipeline stage %s
Running pipeline stage MKMLTemplater
Pipeline stage MKMLTemplater completed in 0.09s
run pipeline stage %s
Running pipeline stage MKMLDeployer
Creating inference service sao10k-mn-12b-lyra-v4a1-v1
Waiting for inference service sao10k-mn-12b-lyra-v4a1-v1 to be ready
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Inference service sao10k-mn-12b-lyra-v4a1-v1 ready after 141.37545347213745s
Pipeline stage MKMLDeployer completed in 141.72s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 2.5358383655548096s
Received healthy response to inference request in 2.2759463787078857s
Received healthy response to inference request in 1.9903409481048584s
Received healthy response to inference request in 2.136533498764038s
Received healthy response to inference request in 2.1181774139404297s
5 requests
0 failed requests
5th percentile: 2.0159082412719727
10th percentile: 2.041475534439087
20th percentile: 2.0926101207733154
30th percentile: 2.121848630905151
40th percentile: 2.1291910648345946
50th percentile: 2.136533498764038
60th percentile: 2.192298650741577
70th percentile: 2.248063802719116
80th percentile: 2.3279247760772703
90th percentile: 2.43188157081604
95th percentile: 2.483859968185425
99th percentile: 2.5254426860809325
mean time: 2.2113673210144045
Pipeline stage StressChecker completed in 12.54s
run pipeline stage %s
Running pipeline stage TriggerMKMLProfilingPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
Pipeline stage TriggerMKMLProfilingPipeline completed in 7.40s
Shutdown handler de-registered
sao10k-mn-12b-lyra-v4a1_v1 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeleter
Skipping teardown as no inference service was successfully deployed
Pipeline stage MKMLProfilerDeleter completed in 0.16s
run pipeline stage %s
Running pipeline stage MKMLProfilerTemplater
Pipeline stage MKMLProfilerTemplater completed in 0.12s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeployer
Creating inference service sao10k-mn-12b-lyra-v4a1-v1-profiler
Waiting for inference service sao10k-mn-12b-lyra-v4a1-v1-profiler to be ready
Inference service sao10k-mn-12b-lyra-v4a1-v1-profiler ready after 150.37050604820251s
Pipeline stage MKMLProfilerDeployer completed in 150.75s
run pipeline stage %s
Running pipeline stage MKMLProfilerRunner
kubectl cp /code/guanaco/guanaco_inference_services/src/inference_scripts tenant-chaiml-guanaco/sao10k-mn-12b-lyra-v4a1-v1-profiler-predictor-00001-deploy8hmsq:/code/chaiverse_profiler_1725551630 --namespace tenant-chaiml-guanaco
kubectl exec -it sao10k-mn-12b-lyra-v4a1-v1-profiler-predictor-00001-deploy8hmsq --namespace tenant-chaiml-guanaco -- sh -c 'cd /code/chaiverse_profiler_1725551630 && python profiles.py profile --best_of_n 8 --auto_batch 5 --batches 1,5,10,15,20,25,30,35,40,45,50,55,60,65,70,75,80,85,90,95,100,105,110,115,120,125,130,135,140,145,150,155,160,165,170,175,180,185,190,195 --samples 200 --input_tokens 512 --output_tokens 64 --summary /code/chaiverse_profiler_1725551630/summary.json'
kubectl exec -it sao10k-mn-12b-lyra-v4a1-v1-profiler-predictor-00001-deploy8hmsq --namespace tenant-chaiml-guanaco -- bash -c 'cat /code/chaiverse_profiler_1725551630/summary.json'
Pipeline stage MKMLProfilerRunner completed in 951.47s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeleter
Checking if service sao10k-mn-12b-lyra-v4a1-v1-profiler is running
Tearing down inference service sao10k-mn-12b-lyra-v4a1-v1-profiler
Service sao10k-mn-12b-lyra-v4a1-v1-profiler has been torndown
Pipeline stage MKMLProfilerDeleter completed in 1.59s
Shutdown handler de-registered
sao10k-mn-12b-lyra-v4a1_v1 status is now inactive due to auto deactivation removed underperforming models

Usage Metrics

Latency Metrics