You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to run the sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 model in ONNX Runtime Java using Intel's DNNL execution provider. I have exported the model to ONNX and also generated the build with the --use_dnnl flag.
However, when I attempt to infer with the model, I am encountering the following error:
[W:onnxruntime:, session_state.cc:1166 VerifyEachNodeIsAssignedToAnEp] Some nodes were not assigned to the preferred execution providers which may or may not have an negative impact on performance. e.g. ORT explicitly assigns shape related ops to CPU to improve perf.
[W:onnxruntime:, session_state.cc:1168 VerifyEachNodeIsAssignedToAnEp] Rerunning with verbose output on a non-minimal build will show node assignments.
[E:onnxruntime:, sequential_executor.cc:516 ExecuteKernel] Non-zero status code returned while running DNNL_17874086050993188016_0 node. Name:'DnnlExecutionProvider_DNNL_17874086050993188016_0_0' Status Message: /onnxruntime/onnxruntime/core/providers/dnnl/subgraph/dnnl_dequantizelinear.cc:191 void onnxruntime::ort_dnnl::DnnlDequantizeLinear::ValidateDims(onnxruntime::ort_dnnl::DnnlSubgraphPrimitive&, onnxruntime::ort_dnnl::DnnlNode&) x_scale and x_zero_point dimensions does not match
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi,
I am trying to run the sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 model in ONNX Runtime Java using Intel's DNNL execution provider. I have exported the model to ONNX and also generated the build with the
--use_dnnl
flag.However, when I attempt to infer with the model, I am encountering the following error:
Here is the snippet for loading the model:
The same model works fine in ONNX CPU executive provider.
Am I missing anything here? Please assist.
Beta Was this translation helpful? Give feedback.
All reactions