Skip to content

Commit

Permalink
Auto-merge updates from auto-update branch
Browse files Browse the repository at this point in the history
  • Loading branch information
mlcommons-bot committed Feb 9, 2025
2 parents 62ef883 + 02680c2 commit 10611c4
Show file tree
Hide file tree
Showing 32 changed files with 848 additions and 854 deletions.
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
| Model | Scenario | Accuracy | Throughput | Latency (in ms) |
|---------|------------|------------|--------------|-------------------|
| bert-99 | offline | 80 | 6.273 | - |
| bert-99 | offline | 80 | 6.175 | - |
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ pip install -U mlcflow

mlc rm cache -f

mlc pull repo anandhu-eng@mlperf-automations --checkout=89d56a9917bae940aa71a9eef3f297e64480f8a1
mlc pull repo anandhu-eng@mlperf-automations --checkout=ba4eb1ee4ca39562f3ee55c8e9972f185b2663bc


```
Expand All @@ -40,4 +40,4 @@ Model Precision: fp32
`F1`: `80.0`, Required accuracy for closed division `>= 89.96526`

### Performance Results
`Samples per second`: `6.27262`
`Samples per second`: `6.175`
Original file line number Diff line number Diff line change
@@ -1,13 +1,13 @@
DeepSparse, Copyright 2021-present / Neuralmagic, Inc. version: 1.8.0 COMMUNITY | (e3778e93) (release) (optimized) (system=avx2, binary=avx2)
[7f269d4006c0 >WARN< operator() ./src/include/wand/utility/warnings.hpp:14] Generating emulated code for quantized (INT8) operations since no VNNI instructions were detected. Set NM_FAST_VNNI_EMULATION=1 to increase performance at the expense of accuracy.
2025-02-07 10:55:25 deepsparse.utils.onnx INFO Generating input 'input_ids', type = int64, shape = [1, 384]
2025-02-07 10:55:25 deepsparse.utils.onnx INFO Generating input 'attention_mask', type = int64, shape = [1, 384]
2025-02-07 10:55:25 deepsparse.utils.onnx INFO Generating input 'token_type_ids', type = int64, shape = [1, 384]
[7fd79b4006c0 >WARN< operator() ./src/include/wand/utility/warnings.hpp:14] Generating emulated code for quantized (INT8) operations since no VNNI instructions were detected. Set NM_FAST_VNNI_EMULATION=1 to increase performance at the expense of accuracy.
2025-02-09 01:19:22 deepsparse.utils.onnx INFO Generating input 'input_ids', type = int64, shape = [1, 384]
2025-02-09 01:19:22 deepsparse.utils.onnx INFO Generating input 'attention_mask', type = int64, shape = [1, 384]
2025-02-09 01:19:22 deepsparse.utils.onnx INFO Generating input 'token_type_ids', type = int64, shape = [1, 384]

No warnings encountered during test.

2 ERRORS encountered. See detailed log.
Loading ONNX model... /home/runner/MLC/repos/local/cache/extract-file_b231fa43/oBERT-Large_95sparse_block4_qat.onnx
Loading ONNX model... /home/runner/MLC/repos/local/cache/extract-file_21c67dc9/oBERT-Large_95sparse_block4_qat.onnx
Constructing SUT...
Finished constructing SUT.
Constructing QSL...
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,42 +18,42 @@ graph TD
download-and-extract,c67e81a4ce2649f5_(_wget,_url.https://zenodo.org/record/3733868/files/vocab.txt_) --> download,file,_wget,_url.https://zenodo.org/record/3733868/files/vocab.txt
get-dataset-squad-vocab,e38874fff5094577 --> download-and-extract,_wget,_url.https://zenodo.org/record/3733868/files/vocab.txt
app-mlperf-inference,d775cac873ee4231_(_reference,_bert-99,_deepsparse,_cpu,_test,_r5.0-dev_default,_int8,_offline_) --> get,dataset-aux,squad-vocab
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> detect,os
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> detect,cpu
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> detect,cpu
get-sys-utils-cm,bc90993277e84b8e --> detect,os
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,python
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,python
get-generic-python-lib,94b62a682bc44791_(_torch_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_torch_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_torch_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_torch_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_torch
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_torch
get-generic-python-lib,94b62a682bc44791_(_transformers_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_transformers_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_transformers_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_transformers_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_transformers
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_transformers
download-file,9cdc8dc41aae437e_(_cmutil,_url.https://github.com/mlcommons/inference_results_v2.1/raw/master/open/NeuralMagic/code/bert/deepsparse/models/oBERT-Large_95sparse_block4_qat.onnx.tar.xz_) --> detect,os
download-and-extract,c67e81a4ce2649f5_(_url.https://github.com/mlcommons/inference_results_v2.1/raw/master/open/NeuralMagic/code/bert/deepsparse/models/oBERT-Large_95sparse_block4_qat.onnx.tar.xz_) --> download,file,_cmutil,_url.https://github.com/mlcommons/inference_results_v2.1/raw/master/open/NeuralMagic/code/bert/deepsparse/models/oBERT-Large_95sparse_block4_qat.onnx.tar.xz
extract-file,3f0b76219d004817_(_path./home/runner/MLC/repos/local/cache/download-file_2d87160d/oBERT-Large_95sparse_block4_qat.onnx.tar.xz_) --> detect,os
download-and-extract,c67e81a4ce2649f5_(_url.https://github.com/mlcommons/inference_results_v2.1/raw/master/open/NeuralMagic/code/bert/deepsparse/models/oBERT-Large_95sparse_block4_qat.onnx.tar.xz_) --> extract,file,_path./home/runner/MLC/repos/local/cache/download-file_2d87160d/oBERT-Large_95sparse_block4_qat.onnx.tar.xz
get-ml-model-bert-large-squad,5e865dbdc65949d2_(_int8,_deepsparse_) --> download-and-extract,_url.https://github.com/mlcommons/inference_results_v2.1/raw/master/open/NeuralMagic/code/bert/deepsparse/models/oBERT-Large_95sparse_block4_qat.onnx.tar.xz
get-ml-model-bert-large-squad,5e865dbdc65949d2_(_int8,_deepsparse_) --> get,dataset-aux,squad-vocab
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,ml-model,language-processing,bert-large,raw,_int8,_deepsparse
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,dataset,squad,original
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,dataset-aux,squad-vocab
extract-file,3f0b76219d004817_(_path./home/runner/MLC/repos/local/cache/download-file_bd5fef6c/oBERT-Large_95sparse_block4_qat.onnx.tar.xz_) --> detect,os
download-and-extract,c67e81a4ce2649f5_(_url.https://github.com/mlcommons/inference_results_v2.1/raw/master/open/NeuralMagic/code/bert/deepsparse/models/oBERT-Large_95sparse_block4_qat.onnx.tar.xz_) --> extract,file,_path./home/runner/MLC/repos/local/cache/download-file_bd5fef6c/oBERT-Large_95sparse_block4_qat.onnx.tar.xz
get-ml-model-bert-large-squad,5e865dbdc65949d2_(_deepsparse,_int8_) --> download-and-extract,_url.https://github.com/mlcommons/inference_results_v2.1/raw/master/open/NeuralMagic/code/bert/deepsparse/models/oBERT-Large_95sparse_block4_qat.onnx.tar.xz
get-ml-model-bert-large-squad,5e865dbdc65949d2_(_deepsparse,_int8_) --> get,dataset-aux,squad-vocab
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,ml-model,language-processing,bert-large,raw,_deepsparse,_int8
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,dataset,squad,original
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,dataset-aux,squad-vocab
generate-mlperf-inference-user-conf,3af4475745964b93 --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
generate-mlperf-inference-user-conf,3af4475745964b93 --> detect,cpu
generate-mlperf-inference-user-conf,3af4475745964b93 --> get,python
generate-mlperf-inference-user-conf,3af4475745964b93 --> get,mlcommons,inference,src,_deeplearningexamples
get-mlperf-inference-sut-configs,c2fbf72009e2445b --> get,cache,dir,_name.mlperf-inference-sut-configs
generate-mlperf-inference-user-conf,3af4475745964b93 --> get,sut,configs
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> generate,user-conf,mlperf,inference
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> generate,user-conf,mlperf,inference
get-mlperf-inference-loadgen,64c3d98d0ba04950 --> detect,os
get-mlperf-inference-loadgen,64c3d98d0ba04950 --> get,python3
get-mlperf-inference-loadgen,64c3d98d0ba04950 --> get,mlcommons,inference,src
Expand Down Expand Up @@ -84,56 +84,56 @@ graph TD
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_package.setuptools_) --> get,generic-python-lib,_pip
get-mlperf-inference-loadgen,64c3d98d0ba04950 --> get,generic-python-lib,_package.setuptools
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,loadgen
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,mlcommons,inference,src,_deeplearningexamples
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,loadgen
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,mlcommons,inference,src,_deeplearningexamples
get-mlperf-inference-src,4b57186581024797 --> detect,os
get-mlperf-inference-src,4b57186581024797 --> get,python3
get-git-repo,ed603e7292974f10_(_branch.deepsparse,_repo.https://github.com/neuralmagic/inference_) --> detect,os
get-mlperf-inference-src,4b57186581024797 --> get,git,repo,_branch.deepsparse,_repo.https://github.com/neuralmagic/inference
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,mlcommons,inference,src
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,mlcommons,inference,src
get-generic-python-lib,94b62a682bc44791_(_package.psutil_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_package.psutil_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_package.psutil_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_package.psutil_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_package.psutil
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_package.psutil
get-generic-python-lib,94b62a682bc44791_(_deepsparse_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_deepsparse_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_deepsparse_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_deepsparse_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_deepsparse
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_deepsparse
get-generic-python-lib,94b62a682bc44791_(_package.pydantic_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_package.pydantic_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_package.pydantic_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_package.pydantic_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_package.pydantic
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_package.pydantic
get-generic-python-lib,94b62a682bc44791_(_tokenization_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_tokenization_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_tokenization_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_tokenization_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_tokenization
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_tokenization
get-generic-python-lib,94b62a682bc44791_(_six_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_six_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_six_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_six_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_six
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_six
get-generic-python-lib,94b62a682bc44791_(_package.absl-py_) --> detect,os
detect-cpu,586c8a43320142f7 --> detect,os
get-generic-python-lib,94b62a682bc44791_(_package.absl-py_) --> detect,cpu
get-generic-python-lib,94b62a682bc44791_(_package.absl-py_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_pip_) --> get,python3
get-generic-python-lib,94b62a682bc44791_(_package.absl-py_) --> get,generic-python-lib,_pip
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> get,generic-python-lib,_package.absl-py
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> get,generic-python-lib,_package.absl-py
detect-cpu,586c8a43320142f7 --> detect,os
benchmark-program,19f369ef47084895 --> detect,cpu
benchmark-program-mlperf,cfff0132a8aa4018 --> benchmark-program,program
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_cpu,_bert-99,_deepsparse,_offline_) --> benchmark-mlperf
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_int8,_bert-99,_deepsparse,_cpu,_offline_) --> benchmark-mlperf
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading

0 comments on commit 10611c4

Please sign in to comment.