1
- [2024-12-27 21:28:18,054 main.py:229 INFO] Detected system ID: KnownSystem.RTX4090x2
2
- [2024-12-27 21:28:18,579 generate_conf_files.py:107 INFO] Generated measurements/ entries for RTX4090x2_TRT/bert-99.9/Offline
3
- [2024-12-27 21:28:18,579 __init__.py:46 INFO] Running command: ./build/bin/harness_bert --logfile_outdir="/cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy" --logfile_prefix="mlperf_log_" --performance_sample_count=10833 --test_mode="AccuracyOnly" --gpu_batch_size=256 --mlperf_conf_path="/home/cmuser/CM/repos/local/cache/13da9a9a9e4e460f /inference/mlperf.conf" --tensor_path="build/preprocessed_data/squad_tokenized/input_ids.npy,build/preprocessed_data/squad_tokenized/segment_ids.npy,build/preprocessed_data/squad_tokenized/input_mask.npy" --use_graphs=false --user_conf_path="/home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/1e993d10f7d444a99c5cc2482bfdc85c .conf" --gpu_inference_streams=2 --gpu_copy_streams=2 --gpu_engines="./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan" --scenario Offline --model bert
4
- [2024-12-27 21:28:18,579 __init__.py:53 INFO] Overriding Environment
1
+ [2024-12-28 21:43:12,494 main.py:229 INFO] Detected system ID: KnownSystem.RTX4090x2
2
+ [2024-12-28 21:43:13,048 generate_conf_files.py:107 INFO] Generated measurements/ entries for RTX4090x2_TRT/bert-99.9/Offline
3
+ [2024-12-28 21:43:13,048 __init__.py:46 INFO] Running command: ./build/bin/harness_bert --logfile_outdir="/cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy" --logfile_prefix="mlperf_log_" --performance_sample_count=10833 --test_mode="AccuracyOnly" --gpu_batch_size=256 --mlperf_conf_path="/home/cmuser/CM/repos/local/cache/11daf5a55e5449f4 /inference/mlperf.conf" --tensor_path="build/preprocessed_data/squad_tokenized/input_ids.npy,build/preprocessed_data/squad_tokenized/segment_ids.npy,build/preprocessed_data/squad_tokenized/input_mask.npy" --use_graphs=false --user_conf_path="/home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/c6913ddf30654865abebcb8dbfb0abd9 .conf" --gpu_inference_streams=2 --gpu_copy_streams=2 --gpu_engines="./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan" --scenario Offline --model bert
4
+ [2024-12-28 21:43:13,048 __init__.py:53 INFO] Overriding Environment
5
5
benchmark : Benchmark.BERT
6
6
buffer_manager_thread_count : 0
7
7
coalesced_tensor : True
@@ -11,8 +11,8 @@ gpu_copy_streams : 2
11
11
gpu_inference_streams : 2
12
12
input_dtype : int32
13
13
input_format : linear
14
- log_dir : /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/logs/2024.12.27 -21.28.16
15
- mlperf_conf_path : /home/cmuser/CM/repos/local/cache/13da9a9a9e4e460f /inference/mlperf.conf
14
+ log_dir : /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/logs/2024.12.28 -21.43.11
15
+ mlperf_conf_path : /home/cmuser/CM/repos/local/cache/11daf5a55e5449f4 /inference/mlperf.conf
16
16
offline_expected_qps : 0.0
17
17
precision : fp16
18
18
preprocessed_data_dir : /home/cmuser/CM/repos/local/cache/4db00c74da1e44c8/preprocessed_data
@@ -21,7 +21,7 @@ system : SystemConfiguration(host_cpu_conf=CPUConfiguration(layout={CPU(name='In
21
21
tensor_path : build/preprocessed_data/squad_tokenized/input_ids.npy,build/preprocessed_data/squad_tokenized/segment_ids.npy,build/preprocessed_data/squad_tokenized/input_mask.npy
22
22
test_mode : AccuracyOnly
23
23
use_graphs : False
24
- user_conf_path : /home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/1e993d10f7d444a99c5cc2482bfdc85c .conf
24
+ user_conf_path : /home/cmuser/CM/repos/mlcommons@mlperf-automations/script/generate-mlperf-inference-user-conf/tmp/c6913ddf30654865abebcb8dbfb0abd9 .conf
25
25
system_id : RTX4090x2
26
26
config_name : RTX4090x2_bert_Offline
27
27
workload_setting : WorkloadSetting(HarnessType.Custom, AccuracyTarget.k_99_9, PowerSetting.MaxP)
@@ -34,64 +34,64 @@ skip_file_checks : True
34
34
power_limit : None
35
35
cpu_freq : None
36
36
&&&& RUNNING BERT_HARNESS # ./build/bin/harness_bert
37
- I1227 21:28:18.626047 20263 main_bert.cc:163] Found 2 GPUs
38
- I1227 21:28:18.748579 20263 bert_server.cc:147] Engine Path: ./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan
39
- [I] [TRT] Loaded engine size: 699 MiB
40
- [I] [TRT] Loaded engine size: 699 MiB
37
+ I1228 21:43:13.096694 20264 main_bert.cc:163] Found 2 GPUs
38
+ I1228 21:43:13.218209 20264 bert_server.cc:147] Engine Path: ./build/engines/RTX4090x2/bert/Offline/bert-Offline-gpu-fp16_S_384_B_256_P_2_vs.custom_k_99_9_MaxP.plan
39
+ [I] [TRT] Loaded engine size: 700 MiB
40
+ [I] [TRT] Loaded engine size: 700 MiB
41
41
[W] [TRT] Using an engine plan file across different models of devices is not recommended and is likely to affect performance or even cause errors.
42
42
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +7, GPU +10, now: CPU 1008, GPU 1511 (MiB)
43
- [I] [TRT] [MemUsageChange] Init cuDNN: CPU +1 , GPU +10, now: CPU 1009 , GPU 1521 (MiB)
43
+ [I] [TRT] [MemUsageChange] Init cuDNN: CPU +2 , GPU +10, now: CPU 1010 , GPU 1521 (MiB)
44
44
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +1152, now: CPU 0, GPU 1152 (MiB)
45
- [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +6, GPU +10, now: CPU 1017 , GPU 1256 (MiB)
46
- [I] [TRT] [MemUsageChange] Init cuDNN: CPU +1, GPU +10, now: CPU 1018 , GPU 1266 (MiB)
45
+ [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +6, GPU +10, now: CPU 1018 , GPU 1255 (MiB)
46
+ [I] [TRT] [MemUsageChange] Init cuDNN: CPU +1, GPU +10, now: CPU 1019 , GPU 1265 (MiB)
47
47
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +1, GPU +576, now: CPU 1, GPU 1152 (MiB)
48
- I1227 21:28:19.399394 20263 bert_server.cc:208] Engines Creation Completed
49
- I1227 21:28:19.432389 20263 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
50
- I1227 21:28:19.432394 20263 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
51
- I1227 21:28:19.432399 20263 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
48
+ I1228 21:43:13.877281 20264 bert_server.cc:208] Engines Creation Completed
49
+ I1228 21:43:13.902434 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
50
+ I1228 21:43:13.902443 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
51
+ I1228 21:43:13.902449 20264 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
52
52
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 319, GPU 2859 (MiB)
53
- [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0 , GPU +8, now: CPU 319 , GPU 2867 (MiB)
54
- I1227 21:28:19.498804 20263 bert_core_vs.cc:426] Setting Opt.Prof. to 0
55
- I1227 21:28:19.498829 20263 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
53
+ [I] [TRT] [MemUsageChange] Init cuDNN: CPU +1 , GPU +8, now: CPU 320 , GPU 2867 (MiB)
54
+ I1228 21:43:13.967170 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 0
55
+ I1228 21:43:13.967195 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
56
56
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +0, now: CPU 1, GPU 1152 (MiB)
57
- I1227 21:28:19.499617 20263 bert_core_vs.cc:476] Setup complete
58
- I1227 21:28:19.499763 20263 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
59
- I1227 21:28:19.499768 20263 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
60
- I1227 21:28:19.499770 20263 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
61
- [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 442 , GPU 2604 (MiB)
62
- [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +8, now: CPU 442 , GPU 2612 (MiB)
63
- I1227 21:28:19.565945 20263 bert_core_vs.cc:426] Setting Opt.Prof. to 0
64
- I1227 21:28:19.565958 20263 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
57
+ I1228 21:43:13.967973 20264 bert_core_vs.cc:476] Setup complete
58
+ I1228 21:43:13.968132 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
59
+ I1228 21:43:13.968134 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
60
+ I1228 21:43:13.968137 20264 bert_core_vs.cc:415] Engine - Profile 0 maxDims 98304 Bmax=256 Smax=384
61
+ [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 443 , GPU 2603 (MiB)
62
+ [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +8, now: CPU 443 , GPU 2611 (MiB)
63
+ I1228 21:43:14.032351 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 0
64
+ I1228 21:43:14.032366 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
65
65
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +1, GPU +0, now: CPU 2, GPU 1152 (MiB)
66
- I1227 21:28:19.566730 20263 bert_core_vs.cc:476] Setup complete
67
- I1227 21:28:19.566908 20263 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
68
- I1227 21:28:19.566911 20263 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
69
- I1227 21:28:19.566915 20263 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
66
+ I1228 21:43:14.033135 20264 bert_core_vs.cc:476] Setup complete
67
+ I1228 21:43:14.033304 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
68
+ I1228 21:43:14.033309 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
69
+ I1228 21:43:14.033313 20264 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
70
70
[I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 566, GPU 4345 (MiB)
71
71
[I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +10, now: CPU 566, GPU 4355 (MiB)
72
- I1227 21:28:19.632443 20263 bert_core_vs.cc:426] Setting Opt.Prof. to 1
72
+ I1228 21:43:14.097095 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 1
73
73
[I] [TRT] Could not set default profile 0 for execution context. Profile index must be set explicitly.
74
74
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +1, GPU +0, now: CPU 3, GPU 1152 (MiB)
75
- I1227 21:28:19.632763 20263 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
76
- I1227 21:28:19.633572 20263 bert_core_vs.cc:476] Setup complete
77
- I1227 21:28:19.633745 20263 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
78
- I1227 21:28:19.633749 20263 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
79
- I1227 21:28:19.633752 20263 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
80
- [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 689 , GPU 4090 (MiB)
81
- [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +10, now: CPU 689 , GPU 4100 (MiB)
82
- I1227 21:28:19.699026 20263 bert_core_vs.cc:426] Setting Opt.Prof. to 1
75
+ I1228 21:43:14.097440 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
76
+ I1228 21:43:14.098248 20264 bert_core_vs.cc:476] Setup complete
77
+ I1228 21:43:14.098415 20264 bert_core_vs.cc:385] Engine - Device Memory requirements: 1409287680
78
+ I1228 21:43:14.098419 20264 bert_core_vs.cc:393] Engine - Number of Optimization Profiles: 2
79
+ I1228 21:43:14.098423 20264 bert_core_vs.cc:415] Engine - Profile 1 maxDims 98304 Bmax=256 Smax=384
80
+ [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +8, now: CPU 690 , GPU 4089 (MiB)
81
+ [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +10, now: CPU 690 , GPU 4099 (MiB)
82
+ I1228 21:43:14.162550 20264 bert_core_vs.cc:426] Setting Opt.Prof. to 1
83
83
[I] [TRT] Could not set default profile 0 for execution context. Profile index must be set explicitly.
84
84
[I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +0, now: CPU 3, GPU 1152 (MiB)
85
- I1227 21:28:19.699366 20263 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
86
- I1227 21:28:19.700189 20263 bert_core_vs.cc:476] Setup complete
87
- I1227 21:28:20.895964 20263 main_bert.cc:184] Starting running actual test.
88
- I1227 21:28:24.215916 20263 main_bert.cc:190] Finished running actual test.
85
+ I1228 21:43:14.162904 20264 bert_core_vs.cc:444] Context creation complete. Max supported batchSize: 256
86
+ I1228 21:43:14.163702 20264 bert_core_vs.cc:476] Setup complete
87
+ I1228 21:43:15.365103 20264 main_bert.cc:184] Starting running actual test.
88
+ I1228 21:43:18.679199 20264 main_bert.cc:190] Finished running actual test.
89
89
90
90
No warnings encountered during test.
91
91
92
92
No errors encountered during test.
93
- [2024-12-27 21:28:24,446 run_harness.py:166 INFO] Result: Accuracy run detected.
94
- [2024-12-27 21:28:24,446 __init__.py:46 INFO] Running command: PYTHONPATH=code/bert/tensorrt/helpers python3 /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/inference/language/bert/accuracy-squad.py --log_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/mlperf_log_accuracy.json --vocab_file build/models/bert/vocab.txt --val_data /home/cmuser/CM/repos/local/cache/4db00c74da1e44c8/data/squad/dev-v1.1.json --out_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/predictions.json --output_dtype float16
93
+ [2024-12-28 21:43:18,912 run_harness.py:166 INFO] Result: Accuracy run detected.
94
+ [2024-12-28 21:43:18,912 __init__.py:46 INFO] Running command: PYTHONPATH=code/bert/tensorrt/helpers python3 /home/cmuser/CM/repos/local/cache/94a57f78972843c6/repo/closed/NVIDIA/build/inference/language/bert/accuracy-squad.py --log_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/mlperf_log_accuracy.json --vocab_file build/models/bert/vocab.txt --val_data /home/cmuser/CM/repos/local/cache/4db00c74da1e44c8/data/squad/dev-v1.1.json --out_file /cm-mount/home/arjun/gh_action_results/valid_results/RTX4090x2-nvidia_original-gpu-tensorrt-vdefault-default_config/bert-99.9/offline/accuracy/predictions.json --output_dtype float16
95
95
{"exact_match": 83.67076631977294, "f1": 90.8832407068292}
96
96
Reading examples...
97
97
Loading cached features from 'eval_features.pickle'...
0 commit comments