1
- [2024-12-28 21:43:12,494 main.py:229 INFO] Detected system ID: KnownSystem.RTX4090x2
2
- [2024-12-28 21:43:13,048 generate_conf_files.py:107 INFO] Generated measurements/ entries for RTX4090x2_TRT/bert-99.9/Offline
3
- [2024-12-28 21:43:13,048 __init__.py:46 INFO] Running command: ./build/bin/harness_bert --logfile_outdir="/cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy" --logfile_prefix="mlperf_log_" --performance_sample_count=10833 --test_mode="AccuracyOnly" --gpu_batch_size=256 --mlperf_conf_path="/home/cmuser/CM/repos/local/cache/11daf5a55e5449f4 /inference/mlperf.conf" --tensor_path="build/preprocessed_data/squad_tokenized/input_ids.npy,build/preprocessed_data/squad_tokenized/segment_ids.npy,build/preprocessed_data/squad_tokenized/input_mask.npy" --use_graphs=false --user_conf_path="/home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/c6913ddf30654865abebcb8dbfb0abd9 .conf" --gpu_inference_streams=2 --gpu_copy_streams=2 --gpu_engines="./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan" --scenario Offline --model bert
4
- [2024-12-28 21:43:13,048 __init__.py:53 INFO] Overriding Environment
1
+ [2024-12-31 21:45:34,506 main.py:229 INFO] Detected system ID: KnownSystem.RTX4090x2
2
+ [2024-12-31 21:45:35,062 generate_conf_files.py:107 INFO] Generated measurements/ entries for RTX4090x2_TRT/bert-99.9/Offline
3
+ [2024-12-31 21:45:35,063 __init__.py:46 INFO] Running command: ./build/bin/harness_bert --logfile_outdir="/cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy" --logfile_prefix="mlperf_log_" --performance_sample_count=10833 --test_mode="AccuracyOnly" --gpu_batch_size=256 --mlperf_conf_path="/home/cmuser/CM/repos/local/cache/fe9ab9b6109f4c4b /inference/mlperf.conf" --tensor_path="build/preprocessed_data/squad_tokenized/input_ids.npy,build/preprocessed_data/squad_tokenized/segment_ids.npy,build/preprocessed_data/squad_tokenized/input_mask.npy" --use_graphs=false --user_conf_path="/home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/aa5ddd5ef2924d75a64fe1514566f612 .conf" --gpu_inference_streams=2 --gpu_copy_streams=2 --gpu_engines="./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan" --scenario Offline --model bert
4
+ [2024-12-31 21:45:35,063 __init__.py:53 INFO] Overriding Environment
5
5
benchmark : Benchmark.BERT
6
6
buffer_manager_thread_count : 0
7
7
coalesced_tensor : True
@@ -11,8 +11,8 @@ gpu_copy_streams : 2
11
11
gpu_inference_streams : 2
12
12
input_dtype : int32
13
13
input_format : linear
14
- log_dir : /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/logs/2024.12.28 -21.43.11
15
- mlperf_conf_path : /home/cmuser/CM/repos/local/cache/11daf5a55e5449f4 /inference/mlperf.conf
14
+ log_dir : /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/logs/2024.12.31 -21.45.33
15
+ mlperf_conf_path : /home/cmuser/CM/repos/local/cache/fe9ab9b6109f4c4b /inference/mlperf.conf
16
16
offline_expected_qps : 0.0
17
17
precision : fp16
18
18
preprocessed_data_dir : /home/cmuser/CM/repos/local/cache/4db00c74da1e44c8/preprocessed_data
@@ -21,7 +21,7 @@ system : SystemConfiguration(host_cpu_conf=CPUConfiguration(layout={CPU(name='In
21
21
tensor_path : build/preprocessed_data/squad_tokenized/input_ids.npy,build/preprocessed_data/squad_tokenized/segment_ids.npy,build/preprocessed_data/squad_tokenized/input_mask.npy
22
22
test_mode : AccuracyOnly
23
23
use_graphs : False
24
- user_conf_path : /home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/c6913ddf30654865abebcb8dbfb0abd9 .conf
24
+ user_conf_path : /home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/aa5ddd5ef2924d75a64fe1514566f612 .conf
25
25
system_id : RTX4090x2
26
26
config_name : RTX4090x2_bert_Offline
27
27
workload_setting : WorkloadSetting(HarnessType.Custom, AccuracyTarget.k_99_9, PowerSetting.MaxP)
@@ -34,8 +34,8 @@ skip_file_checks : True
34
34
power_limit : None
35
35
cpu_freq : None
36
36
&&&& RUNNING BERT_HARNESS # ./build/bin/harness_bert
37
- I1228 21:43:13.096694 20264 main_bert.cc:163] Found 2 GPUs
38
- I1228 21:43:13.218209 20264 bert_server.cc:147] Engine Path: ./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan
37
+ I1231 21:45:35.110177 20264 main_bert.cc:163] Found 2 GPUs
38
+ I1231 21:45:35.231178 20264 bert_server.cc:147] Engine Path: ./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan
39
39
[I] [TRT] Loaded engine size: 700 MiB
40
40
[I] [TRT] Loaded engine size: 700 MiB
41
41
[W] [TRT] Using an engine plan file across different models of devices is not recommended and is likely to affect performance or even cause errors.
@@ -45,53 +45,53 @@ I1228 21:43:13.218209 20264 bert_server.cc:147] Engine Path: ./build/engines/RTX
45
45
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +6, GPU +10, now: CPU 1018, GPU 1255 (MiB)
46
46
[I] [TRT] [MemUsageChange] Init cuDNN: CPU +1, GPU +10, now: CPU 1019, GPU 1265 (MiB)
47
47
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +1, GPU +576, now: CPU 1, GPU 1152 (MiB)
48
- I1228 21:43:13.877281 20264 bert_server.cc:208] Engines Creation Completed
49
- I1228 21:43:13.902434 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
50
- I1228 21:43:13.902443 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
51
- I1228 21:43:13.902449 20264 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
48
+ I1231 21:45:35.893620 20264 bert_server.cc:208] Engines Creation Completed
49
+ I1231 21:45:35.927969 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
50
+ I1231 21:45:35.927980 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
51
+ I1231 21:45:35.927986 20264 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
52
52
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 319, GPU 2859 (MiB)
53
53
[I] [TRT] [MemUsageChange] Init cuDNN: CPU +1, GPU +8, now: CPU 320, GPU 2867 (MiB)
54
- I1228 21:43:13.967170 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 0
55
- I1228 21:43:13.967195 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
54
+ I1231 21:45:35.992913 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 0
55
+ I1231 21:45:35.992939 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
56
56
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +0, now: CPU 1, GPU 1152 (MiB)
57
- I1228 21:43:13.967973 20264 bert_core_vs.cc:476] Setup complete
58
- I1228 21:43:13.968132 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
59
- I1228 21:43:13.968134 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
60
- I1228 21:43:13.968137 20264 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
57
+ I1231 21:45:35.993705 20264 bert_core_vs.cc:476] Setup complete
58
+ I1231 21:45:35.993885 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
59
+ I1231 21:45:35.993891 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
60
+ I1231 21:45:35.993894 20264 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
61
61
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 443, GPU 2603 (MiB)
62
62
[I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +8, now: CPU 443, GPU 2611 (MiB)
63
- I1228 21:43:14.032351 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 0
64
- I1228 21:43:14.032366 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
63
+ I1231 21:45:36.061340 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 0
64
+ I1231 21:45:36.061358 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
65
65
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +1, GPU +0, now: CPU 2, GPU 1152 (MiB)
66
- I1228 21:43:14.033135 20264 bert_core_vs.cc:476] Setup complete
67
- I1228 21:43:14.033304 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
68
- I1228 21:43:14.033309 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
69
- I1228 21:43:14.033313 20264 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
66
+ I1231 21:45:36.062158 20264 bert_core_vs.cc:476] Setup complete
67
+ I1231 21:45:36.062332 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
68
+ I1231 21:45:36.062336 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
69
+ I1231 21:45:36.062340 20264 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
70
70
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 566, GPU 4345 (MiB)
71
- [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0 , GPU +10, now: CPU 566 , GPU 4355 (MiB)
72
- I1228 21:43:14.097095 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 1
71
+ [I] [TRT] [MemUsageChange] Init cuDNN: CPU +1 , GPU +10, now: CPU 567 , GPU 4355 (MiB)
72
+ I1231 21:45:36.127660 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 1
73
73
[I] [TRT] Could not set default profile 0 for execution context. Profile index must be set explicitly.
74
74
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +1, GPU +0, now: CPU 3, GPU 1152 (MiB)
75
- I1228 21:43:14.097440 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
76
- I1228 21:43:14.098248 20264 bert_core_vs.cc:476] Setup complete
77
- I1228 21:43:14.098415 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
78
- I1228 21:43:14.098419 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
79
- I1228 21:43:14.098423 20264 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
75
+ I1231 21:45:36.128007 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
76
+ I1231 21:45:36.128804 20264 bert_core_vs.cc:476] Setup complete
77
+ I1231 21:45:36.128968 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
78
+ I1231 21:45:36.128973 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
79
+ I1231 21:45:36.128975 20264 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
80
80
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 690, GPU 4089 (MiB)
81
81
[I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +10, now: CPU 690, GPU 4099 (MiB)
82
- I1228 21:43:14.162550 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 1
82
+ I1231 21:45:36.193878 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 1
83
83
[I] [TRT] Could not set default profile 0 for execution context. Profile index must be set explicitly.
84
84
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +0, now: CPU 3, GPU 1152 (MiB)
85
- I1228 21:43:14.162904 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
86
- I1228 21:43:14.163702 20264 bert_core_vs.cc:476] Setup complete
87
- I1228 21:43:15.365103 20264 main_bert.cc:184] Starting running actual test.
88
- I1228 21:43:18.679199 20264 main_bert.cc:190] Finished running actual test.
85
+ I1231 21:45:36.194209 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
86
+ I1231 21:45:36.195005 20264 bert_core_vs.cc:476] Setup complete
87
+ I1231 21:45:37.395401 20264 main_bert.cc:184] Starting running actual test.
88
+ I1231 21:45:40.743121 20264 main_bert.cc:190] Finished running actual test.
89
89
90
90
No warnings encountered during test.
91
91
92
92
No errors encountered during test.
93
- [2024-12-28 21:43:18,912 run_harness.py:166 INFO] Result: Accuracy run detected.
94
- [2024-12-28 21:43:18,912 __init__.py:46 INFO] Running command: PYTHONPATH=code/bert/tensorrt/helpers python3 /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/inference/language/bert/accuracy-squad.py --log_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/mlperf_log_accuracy.json --vocab_file build/models/bert/vocab.txt --val_data /home/cmuser/CM/repos/local/cache/4db00c74da1e44c8/data/squad/dev-v1.1.json --out_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/predictions.json --output_dtype float16
93
+ [2024-12-31 21:45:40,987 run_harness.py:166 INFO] Result: Accuracy run detected.
94
+ [2024-12-31 21:45:40,987 __init__.py:46 INFO] Running command: PYTHONPATH=code/bert/tensorrt/helpers python3 /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/inference/language/bert/accuracy-squad.py --log_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/mlperf_log_accuracy.json --vocab_file build/models/bert/vocab.txt --val_data /home/cmuser/CM/repos/local/cache/4db00c74da1e44c8/data/squad/dev-v1.1.json --out_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/predictions.json --output_dtype float16
95
95
{"exact_match": 83.67076631977294, "f1": 90.8832407068292}
96
96
Reading examples...
97
97
Loading cached features from 'eval_features.pickle'...
0 commit comments