Releases · microsoft/onnxruntime · GitHub

August 12, 2025 at 12:00 AMai_discoveryinfo

Product

C3.ai

aimachine-learningautomation

Update Details

Comprehensive information about this update

Full Content

Release Notes
What's new? This release adds an optimized CPU/MLAS implementation of DequantizeLinear (8 bit) and introduces the build option client_package_build, which enables default options that are more appropriate for client/on-device workloads (e.g., disable thread spinning by default). Build System & Packages Add –client_package_build option (#25351) - @jywu-msft Remove the python installation steps from win-qnn-arm64-ci-pipeline.yml (#25552) - @snnn CPU EP Add multithreaded/vectorized implementation of DequantizeLinear for int8 and uint8 inputs (SSE2, NEON) (#24818) - @adrianlizarraga QNN EP Add support for the Upsample, Einsum, LSTM, and CumSum operators (#24265, #24616, #24646, #24820) - @quic-zhaoxul, @1duo, @chenweng-quic, @Akupadhye Fuse scale into Softmax (#24809) - @qti-yuduo Enable DSP queue polling when performance is set to “burst” mode (#25361) - @quic-calvnguy Update QNN SDK to version 2.36.1 (#25388) - @qti-jkilpatrick Include the license file from QNN SDK in the Microsoft.ML.OnnxRunitme.QNN NuGet package (#25158) - @HectorSVC

Published At

Tuesday, August 12, 2025

12:00:00 AM

Discovered At

Monday, August 25, 2025

10:25:34 PM

Confidence

1