submission_id: oliveiratime-chaiverse_v1
developer_uid: Oliveiratime
alignment_samples: 11542
alignment_score: 0.5941574171197259
best_of: 1
celo_rating: 1137.67
display_name: oliveiratime-chaiverse_v1
formatter: {'memory_template': "{bot_name}'s Persona: {memory}\n####\n", 'prompt_template': '{prompt}\n<START>\n', 'bot_template': '{bot_name}: {message}\n', 'user_template': '{user_name}: {message}\n', 'response_template': '{bot_name}:', 'truncate_by_message': False}
generation_params: {'temperature': 1.0, 'top_p': 1.0, 'min_p': 0.9, 'top_k': 40, 'presence_penalty': 0.2, 'frequency_penalty': 0.1, 'stopping_words': ['\n'], 'max_input_tokens': 512, 'best_of': 1, 'max_output_tokens': 64}
gpu_counts: {'NVIDIA RTX A5000': 1}
is_internal_developer: False
language_model: Oliveiratime/chaiverse
latencies: [{'batch_size': 1, 'throughput': 1.0697728619300808, 'latency_mean': 0.9346838080883026, 'latency_p50': 0.9398816823959351, 'latency_p90': 1.0441334962844848}, {'batch_size': 5, 'throughput': 3.4821288126079923, 'latency_mean': 1.4275546729564668, 'latency_p50': 1.4220815896987915, 'latency_p90': 1.5930265426635741}, {'batch_size': 10, 'throughput': 5.189879972617139, 'latency_mean': 1.9064902651309967, 'latency_p50': 1.9012733697891235, 'latency_p90': 2.144068717956543}, {'batch_size': 15, 'throughput': 6.028763367461653, 'latency_mean': 2.4449210321903228, 'latency_p50': 2.4536566734313965, 'latency_p90': 2.7969090700149537}, {'batch_size': 20, 'throughput': 6.606935467276574, 'latency_mean': 2.967182936668396, 'latency_p50': 2.9672292470932007, 'latency_p90': 3.4128493070602417}, {'batch_size': 25, 'throughput': 7.01914548214838, 'latency_mean': 3.48530069231987, 'latency_p50': 3.460726737976074, 'latency_p90': 4.101570177078247}, {'batch_size': 30, 'throughput': 7.276605766000643, 'latency_mean': 4.0363370084762575, 'latency_p50': 3.988510012626648, 'latency_p90': 4.639559459686279}, {'batch_size': 35, 'throughput': 7.32271902356018, 'latency_mean': 4.65101724267006, 'latency_p50': 4.5930163860321045, 'latency_p90': 5.38740780353546}, {'batch_size': 40, 'throughput': 7.346595411607753, 'latency_mean': 5.253224321603775, 'latency_p50': 5.151739120483398, 'latency_p90': 6.442975378036499}]
max_input_tokens: 512
max_output_tokens: 64
model_architecture: LlamaForCausalLM
model_group: Oliveiratime/chaiverse
model_name: oliveiratime-chaiverse_v1
model_num_parameters: 8030261248.0
model_repo: Oliveiratime/chaiverse
model_size: 8B
num_battles: 11541
num_wins: 4435
propriety_score: 0.7621009268795057
propriety_total_count: 971.0
ranking_group: single
status: inactive
submission_type: basic
throughput_3p7s: 7.28
timestamp: 2024-09-10T05:43:28+00:00
us_pacific_date: 2024-09-09
win_ratio: 0.3842821245992548
Download Preference Data
Resubmit model
Shutdown handler not registered because Python interpreter is not running in the main thread
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLizer
Starting job with name oliveiratime-chaiverse-v1-mkmlizer
Waiting for job on oliveiratime-chaiverse-v1-mkmlizer to finish
oliveiratime-chaiverse-v1-mkmlizer: ╔═════════════════════════════════════════════════════════════════════╗
oliveiratime-chaiverse-v1-mkmlizer: ║ _____ __ __ ║
oliveiratime-chaiverse-v1-mkmlizer: ║ / _/ /_ ___ __/ / ___ ___ / / ║
oliveiratime-chaiverse-v1-mkmlizer: ║ / _/ / // / |/|/ / _ \/ -_) -_) / ║
oliveiratime-chaiverse-v1-mkmlizer: ║ /_//_/\_, /|__,__/_//_/\__/\__/_/ ║
oliveiratime-chaiverse-v1-mkmlizer: ║ /___/ ║
oliveiratime-chaiverse-v1-mkmlizer: ║ ║
oliveiratime-chaiverse-v1-mkmlizer: ║ Version: 0.10.1 ║
oliveiratime-chaiverse-v1-mkmlizer: ║ Copyright 2023 MK ONE TECHNOLOGIES Inc. ║
oliveiratime-chaiverse-v1-mkmlizer: ║ https://mk1.ai ║
oliveiratime-chaiverse-v1-mkmlizer: ║ ║
oliveiratime-chaiverse-v1-mkmlizer: ║ The license key for the current software has been verified as ║
oliveiratime-chaiverse-v1-mkmlizer: ║ belonging to: ║
oliveiratime-chaiverse-v1-mkmlizer: ║ ║
oliveiratime-chaiverse-v1-mkmlizer: ║ Chai Research Corp. ║
oliveiratime-chaiverse-v1-mkmlizer: ║ Account ID: 7997a29f-0ceb-4cc7-9adf-840c57b4ae6f ║
oliveiratime-chaiverse-v1-mkmlizer: ║ Expiration: 2024-10-15 23:59:59 ║
oliveiratime-chaiverse-v1-mkmlizer: ║ ║
oliveiratime-chaiverse-v1-mkmlizer: ╚═════════════════════════════════════════════════════════════════════╝
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
oliveiratime-chaiverse-v1-mkmlizer: Downloaded to shared memory in 37.231s
oliveiratime-chaiverse-v1-mkmlizer: quantizing model to /dev/shm/model_cache, profile:s0, folder:/tmp/tmp92gcqz97, device:0
oliveiratime-chaiverse-v1-mkmlizer: Saving flywheel model at /dev/shm/model_cache
oliveiratime-chaiverse-v1-mkmlizer: quantized model in 26.858s
oliveiratime-chaiverse-v1-mkmlizer: Processed model Oliveiratime/chaiverse in 64.089s
oliveiratime-chaiverse-v1-mkmlizer: creating bucket guanaco-mkml-models
oliveiratime-chaiverse-v1-mkmlizer: Bucket 's3://guanaco-mkml-models/' created
oliveiratime-chaiverse-v1-mkmlizer: uploading /dev/shm/model_cache to s3://guanaco-mkml-models/oliveiratime-chaiverse-v1
oliveiratime-chaiverse-v1-mkmlizer: cp /dev/shm/model_cache/config.json s3://guanaco-mkml-models/oliveiratime-chaiverse-v1/config.json
oliveiratime-chaiverse-v1-mkmlizer: cp /dev/shm/model_cache/special_tokens_map.json s3://guanaco-mkml-models/oliveiratime-chaiverse-v1/special_tokens_map.json
oliveiratime-chaiverse-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer_config.json s3://guanaco-mkml-models/oliveiratime-chaiverse-v1/tokenizer_config.json
oliveiratime-chaiverse-v1-mkmlizer: cp /dev/shm/model_cache/tokenizer.json s3://guanaco-mkml-models/oliveiratime-chaiverse-v1/tokenizer.json
oliveiratime-chaiverse-v1-mkmlizer: cp /dev/shm/model_cache/flywheel_model.0.safetensors s3://guanaco-mkml-models/oliveiratime-chaiverse-v1/flywheel_model.0.safetensors
oliveiratime-chaiverse-v1-mkmlizer: Loading 0: 0%| | 0/291 [00:00<?, ?it/s] Loading 0: 2%|▏ | 7/291 [00:00<00:05, 49.52it/s] Loading 0: 5%|▌ | 16/291 [00:00<00:04, 65.86it/s] Loading 0: 9%|▊ | 25/291 [00:00<00:03, 67.31it/s] Loading 0: 14%|█▎ | 40/291 [00:00<00:03, 82.18it/s] Loading 0: 17%|█▋ | 49/291 [00:00<00:02, 81.56it/s] Loading 0: 20%|█▉ | 58/291 [00:00<00:02, 79.75it/s] Loading 0: 23%|██▎ | 67/291 [00:00<00:02, 81.16it/s] Loading 0: 26%|██▌ | 76/291 [00:00<00:02, 80.71it/s] Loading 0: 29%|██▉ | 85/291 [00:02<00:10, 20.36it/s] Loading 0: 32%|███▏ | 94/291 [00:02<00:07, 25.71it/s] Loading 0: 35%|███▌ | 103/291 [00:02<00:05, 32.25it/s] Loading 0: 38%|███▊ | 112/291 [00:02<00:04, 39.30it/s] Loading 0: 42%|████▏ | 121/291 [00:02<00:03, 45.84it/s] Loading 0: 45%|████▍ | 130/291 [00:02<00:03, 49.20it/s] Loading 0: 48%|████▊ | 139/291 [00:02<00:02, 56.72it/s] Loading 0: 52%|█████▏ | 150/291 [00:03<00:02, 67.63it/s] Loading 0: 55%|█████▍ | 159/291 [00:03<00:01, 72.32it/s] Loading 0: 58%|█████▊ | 168/291 [00:03<00:01, 75.54it/s] Loading 0: 61%|██████ | 177/291 [00:03<00:01, 73.83it/s] Loading 0: 64%|██████▍ | 186/291 [00:03<00:01, 73.25it/s] Loading 0: 67%|██████▋ | 194/291 [00:04<00:04, 20.45it/s] Loading 0: 69%|██████▉ | 202/291 [00:04<00:03, 24.81it/s] Loading 0: 73%|███████▎ | 211/291 [00:04<00:02, 31.29it/s] Loading 0: 76%|███████▌ | 220/291 [00:05<00:01, 39.02it/s] Loading 0: 79%|███████▊ | 229/291 [00:05<00:01, 46.70it/s] Loading 0: 82%|████████▏ | 238/291 [00:05<00:00, 53.56it/s] Loading 0: 85%|████████▍ | 247/291 [00:05<00:00, 55.55it/s] Loading 0: 88%|████████▊ | 256/291 [00:05<00:00, 57.20it/s] Loading 0: 91%|█████████ | 265/291 [00:05<00:00, 61.60it/s] Loading 0: 94%|█████████▍| 274/291 [00:05<00:00, 66.27it/s] Loading 0: 97%|█████████▋| 283/291 [00:05<00:00, 67.81it/s] Loading 0: 100%|██████████| 291/291 [00:11<00:00, 5.03it/s]
Job oliveiratime-chaiverse-v1-mkmlizer completed after 84.51s with status: succeeded
Stopping job with name oliveiratime-chaiverse-v1-mkmlizer
Pipeline stage MKMLizer completed in 85.36s
run pipeline stage %s
Running pipeline stage MKMLTemplater
Pipeline stage MKMLTemplater completed in 0.11s
run pipeline stage %s
Running pipeline stage MKMLDeployer
Creating inference service oliveiratime-chaiverse-v1
Waiting for inference service oliveiratime-chaiverse-v1 to be ready
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Connection pool is full, discarding connection: %s. Connection pool size: %s
Inference service oliveiratime-chaiverse-v1 ready after 160.99913382530212s
Pipeline stage MKMLDeployer completed in 161.50s
run pipeline stage %s
Running pipeline stage StressChecker
Received healthy response to inference request in 1.7941255569458008s
Received healthy response to inference request in 2.3902928829193115s
Received healthy response to inference request in 1.663109540939331s
Received healthy response to inference request in 1.9971284866333008s
Received healthy response to inference request in 1.8920433521270752s
5 requests
0 failed requests
5th percentile: 1.689312744140625
10th percentile: 1.715515947341919
20th percentile: 1.7679223537445068
30th percentile: 1.8137091159820558
40th percentile: 1.8528762340545655
50th percentile: 1.8920433521270752
60th percentile: 1.9340774059295653
70th percentile: 1.9761114597320557
80th percentile: 2.075761365890503
90th percentile: 2.233027124404907
95th percentile: 2.311660003662109
99th percentile: 2.374566307067871
mean time: 1.9473399639129638
Pipeline stage StressChecker completed in 10.82s
run pipeline stage %s
Running pipeline stage TriggerMKMLProfilingPipeline
run_pipeline:run_in_cloud %s
starting trigger_guanaco_pipeline args=%s
Pipeline stage TriggerMKMLProfilingPipeline completed in 8.23s
Shutdown handler de-registered
oliveiratime-chaiverse_v1 status is now deployed due to DeploymentManager action
Shutdown handler registered
run pipeline %s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeleter
Skipping teardown as no inference service was successfully deployed
Pipeline stage MKMLProfilerDeleter completed in 0.11s
run pipeline stage %s
Running pipeline stage MKMLProfilerTemplater
Pipeline stage MKMLProfilerTemplater completed in 0.11s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeployer
Creating inference service oliveiratime-chaiverse-v1-profiler
Waiting for inference service oliveiratime-chaiverse-v1-profiler to be ready
Inference service oliveiratime-chaiverse-v1-profiler ready after 160.3702335357666s
Pipeline stage MKMLProfilerDeployer completed in 160.76s
run pipeline stage %s
Running pipeline stage MKMLProfilerRunner
kubectl cp /code/guanaco/guanaco_inference_services/src/inference_scripts tenant-chaiml-guanaco/oliveiratime-chaiverse-v1-profiler-predictor-00001-deploymkbkj6:/code/chaiverse_profiler_1725947484 --namespace tenant-chaiml-guanaco
kubectl exec -it oliveiratime-chaiverse-v1-profiler-predictor-00001-deploymkbkj6 --namespace tenant-chaiml-guanaco -- sh -c 'cd /code/chaiverse_profiler_1725947484 && python profiles.py profile --best_of_n 1 --auto_batch 5 --batches 1,5,10,15,20,25,30,35,40,45,50,55,60,65,70,75,80,85,90,95,100,105,110,115,120,125,130,135,140,145,150,155,160,165,170,175,180,185,190,195 --samples 200 --input_tokens 512 --output_tokens 64 --summary /code/chaiverse_profiler_1725947484/summary.json'
kubectl exec -it oliveiratime-chaiverse-v1-profiler-predictor-00001-deploymkbkj6 --namespace tenant-chaiml-guanaco -- bash -c 'cat /code/chaiverse_profiler_1725947484/summary.json'
Pipeline stage MKMLProfilerRunner completed in 461.73s
run pipeline stage %s
Running pipeline stage MKMLProfilerDeleter
Checking if service oliveiratime-chaiverse-v1-profiler is running
Tearing down inference service oliveiratime-chaiverse-v1-profiler
Service oliveiratime-chaiverse-v1-profiler has been torndown
Pipeline stage MKMLProfilerDeleter completed in 1.69s
Shutdown handler de-registered
oliveiratime-chaiverse_v1 status is now inactive due to auto deactivation removed underperforming models