solrbot opened a new pull request, #3657: URL: https://github.com/apache/solr/pull/3657
This PR contains the following updates: | Package | Type | Update | Change | |---|---|---|---| | [com.microsoft.onnxruntime:onnxruntime](https://microsoft.github.io/onnxruntime/) ([source](https://redirect.github.com/microsoft/onnxruntime)) | dependencies | minor | `1.20.0` -> `1.22.0` | --- ### Release Notes <details> <summary>microsoft/onnxruntime (com.microsoft.onnxruntime:onnxruntime)</summary> ### [`v1.22.0`](https://redirect.github.com/microsoft/onnxruntime/releases/tag/v1.22.0): ONNX Runtime v1.22 #### Announcements - This release introduces new API's for Model Editor, Auto EP infrastructure, and AOT Compile - OnnxRuntime GPU packages require CUDA 12.x , packages built for CUDA 11.x are no longer published. - The min supported Windows version is now 10.0.19041. #### GenAI & Advanced Model Features - **Constrained Decoding:** Introduced new capabilities for constrained decoding, offering more control over generative AI model outputs. #### Execution & Core Optimizations ##### Core - **Auto EP Selection Infrastructure:** Added foundational infrastructure to enable automatic selection of Execution Providers via selection policies, aiming to simplify configuration and optimize performance. (Pull Request [#​24430](https://redirect.github.com/microsoft/onnxruntime/issues/24430)) - **Compile API:** Introduced new APIs to support explicit compilation of ONNX models. - See: [OrtCompileApi Struct Reference](https://onnxruntime.ai/docs/api/c/struct_ort_compile_api.html) (Assuming a similar link structure for future documentation) - See: [EP Context Design](https://onnxruntime.ai/docs/api/c/struct_ort_ep_context.html) (Assuming a similar link structure for future documentation) - **Model Editor API** api's for creating or editing ONNX models - See: [OrtModelEditorApi](https://onnxruntime.ai/docs/api/c/struct_ort_model_editor_api.html#details) ##### Execution Provider (EP) Updates ##### CPU EP/MLAS - **KleidiAI Integration:** Integrated KleidiAI into ONNX Runtime/MLAS for enhanced performance on Arm architectures. - **MatMulNBits Support:** Added support for `MatMulNBits`, enabling matrix multiplication with weights quantized to 8 bits. - **GroupQueryAttention optimizations and enhancements** ##### OpenVINO EP - Added support up to OpenVINO 2025.1 - Introduced Intel compiler level optimizations for QDQ models. - Added support to select Intel devices based on LUID - Load\_config feature improvement to support AUTO, HETERO and MULTI plugin. - misc bugfixes/optimizations - For detailed updates, refer to Pull Request [#​24394](https://redirect.github.com/microsoft/onnxruntime/issues/24394): [ONNXRuntime OpenVINO - Release 1.22](https://redirect.github.com/microsoft/onnxruntime/pull/24394) ##### QNN EP - **SDK Update:** Added support for QNN SDK 2.33.2. - operator updates/support to Sum, Softmax, Upsample, Expand, ScatterND, Einsum - QNN EP can be built as shared or static library. - enable QnnGpu backend - For detailed updates refer to [recent QNN tagged PR's](https://redirect.github.com/microsoft/onnxruntime/pulls?q=is%3Apr+qnn+ep+is%3Aclosed+label%3Aep%3AQNN) ##### TensorRT EP - **TensorRT Version:** Added support for TensorRT 10.9. - **Note for onnx-tensorrt open-source parser users:** Please check [here](https://onnxruntime.ai/docs/build/eps.html#note-to-ort-1210-open-sourced-parser-users) for specific requirements (Referencing 1.21 link as a placeholder, this should be updated for 1.22). - **New Features:** - EP option to enable TRT Preview Feature - Support to load TensorRT V3 plugin - **Bug Fixes:** - Resolved an issue related to multithreading scenarios. - Fixed incorrect GPU usage that affected both TensorRT EP and CUDA EP. ##### NV TensorRT RTX EP - **New Execution Provider:** Introduced a new Execution Provider specifically for Nvidia RTX GPUs, leveraging TensorRT for optimized performance. ##### CUDA EP - **MatMulNBits Enhancement:** Added support for 8-bit weight-only quantization in `MatMulNBits`. - **Bug Fixes:** - Fixed incorrect GPU usage (also mentioned under TensorRT EP). ##### VitisAI EP - Miscellaneous bug fixes and improvements. #### Infrastructure & Build Improvements ##### Build System & Packages - **QNN Nuget Package:** The QNN Nuget package is now built as ARM64x. ##### Dependencies / Version Updates - **CUDA Version Update:** This release includes an update to the CUDA version. Users should consult the documentation for specific version requirements. CUDA 11 based GPU packages no longer published. #### Web - **WebGPU Expansion:** - Added WebGPU support to the node.js package (Windows and macOS). - Enabled WebGPU when building from source for macOS, Linux, and Windows. #### Mobile - No major updates of note this release. ### Contributions Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members: Yulong Wang, Jian Chen, Changming Sun, Satya Kumar Jandhyala, Hector Li, Prathik Rao, Adrian Lizarraga, Jiajia Qin, Scott McKay, Jie Chen, Tianlei Wu, Edward Chen, Wanming Lin, xhcao, vraspar, Dmitri Smirnov, Jing Fang, Yifan Li, Caroline Zhu, Jianhui Dai, Chi Lo, Guenther Schmuelling, Ryan Hill, Sushanth Rajasankar, Yi-Hong Lyu, Ankit Maheshkar, Artur Wojcik, Baiju Meswani, David Fan, Enrico Galli, Hans, Jambay Kinley, John Paul, Peishen Yan, Yateng Hong, amarin16, chuteng-quic, kunal-vaishnavi, quic-hungjuiw, Alessio Soldano, Andreas Hussing, Ashish Garg, Ashwath Shankarnarayan, Chengdong Liang, Clément Péron, Erick Muñoz, Fanchen Kong, George Wu, Haik Silm, Jagadish Krishnamoorthy, Justin Chu, Karim Vadsariya, Kevin Chen, Mark Schofield, Masaya, Kato, Michael Tyler, Nenad Banfic, Ningxin Hu, Praveen G, Preetha Veeramalai, Ranjit Ranjan, Seungtaek Kim, Ti-Tai Wang, Xiaofei Han, Yueqing Zhang, co63oc, derdeljan-msft, genmingz\@​AMD, jiangzhaoming, jing-bao, kuanyul-quic, liqun Fu, minfhong-quic, mingyue, quic-tirupath, quic-zhaoxul, saurabh, selenayang888, sfatimar, sheetalarkadam, virajwad, zz002, Ștefan Talpalaru ### [`v1.21.1`](https://redirect.github.com/microsoft/onnxruntime/releases/tag/v1.21.1): ONNX Runtime v1.21.1 ### What's new? - Extend CMAKE\_CUDA\_FLAGS with all Blackwell compute capacity [#​23928](https://redirect.github.com/microsoft/onnxruntime/issues/23928) - [@​yf711](https://redirect.github.com/yf711) - \[ARM CPU] Fix fp16 const initialization on no-fp16 platform [#​23978](https://redirect.github.com/microsoft/onnxruntime/issues/23978) - [@​fajin-corp](https://redirect.github.com/fajin-corp) - \[TensorRT EP] Call cudaSetDevice at compute function for handling multithreading scenario [#​24010](https://redirect.github.com/microsoft/onnxruntime/issues/24010) - [@​chilo-ms](https://redirect.github.com/chilo-ms) - Fix attention bias broadcast [#​24017](https://redirect.github.com/microsoft/onnxruntime/issues/24017) - [@​tianleiwu](https://redirect.github.com/tianleiwu) - Deleted the constant SKIP\_CUDA\_TEST\_WITH\_DML [#​24113](https://redirect.github.com/microsoft/onnxruntime/issues/24113) - [@​CodingSeaotter](https://redirect.github.com/CodingSeaotter) - \[QNN EP] ARM64EC python package remove --vcpkg in build [#​24174](https://redirect.github.com/microsoft/onnxruntime/issues/24174) - [@​jywu-msft](https://redirect.github.com/jywu-msft) - \[wasm] remove --vcpkg in wasm build [#​24179](https://redirect.github.com/microsoft/onnxruntime/issues/24179) - [@​fs-eire](https://redirect.github.com/fs-eire) ### [`v1.21.0`](https://redirect.github.com/microsoft/onnxruntime/releases/tag/v1.21.0): ONNX Runtime v1.21.0 #### Announcements - No large announcements of note this release! We've made a lot of small refinements to streamline your ONNX Runtime experience. #### GenAI & Advanced Model Features ##### Enhanced Decoding & Pipeline Support - Added "chat mode" support for CPU, GPU, and WebGPU. - Provided support for decoder model pipelines. - Added support for Java API for MultiLoRA. ##### API & Compatibility Updates - Chat mode introduced breaking changes in the API (see [migration guide](https://onnxruntime.ai/docs/genai/howto/migrate.html)). ##### Bug Fixes for Model Output - Fixed Phi series garbage output issues with long prompts. - Resolved gibberish issues with `top_k` on CPU. #### Execution & Core Optimizations ##### Core Refinements - Reduced default logger usage for improved efficiency([#​23030](https://redirect.github.com/microsoft/onnxruntime/issues/23030)). - Fixed a visibility issue in theadpool ([#​23098](https://redirect.github.com/microsoft/onnxruntime/issues/23098)). ##### Execution Provider (EP) Updates ##### General - Removed TVM EP from the source tree([#​22827](https://redirect.github.com/microsoft/onnxruntime/issues/22827)). - Marked NNAPI EP for deprecation (following Google's deprecation of NNAPI). - Fixed a DLL delay loading issue that impacts WebGPU EP and DirectML EP's usability on Windows ([#​23111](https://redirect.github.com/microsoft/onnxruntime/issues/23111), [#​23227](https://redirect.github.com/microsoft/onnxruntime/issues/23227)) ##### TensorRT EP Improvements - Added support for TensorRT 10.8. - [onnx-tensorrt](https://redirect.github.com/onnx/onnx-tensorrt) open-source parser user: please check [here](https://onnxruntime.ai/docs/build/eps.html#note-to-ort-1210-open-sourced-parser-users) for requirement. - Assigned DDS ops (`NMS`, `RoiAlign`, `NonZero`) to TensorRT by default. - Introduced option `trt_op_types_to_exclude` to exclude specific ops from TensorRT assignment. ##### CUDA EP Improvements - Added a python API [preload\_dlls](https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#preload-dlls) to [coexist with PyTorch](https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#compatibility-with-pytorch). - Miscellaneous enhancements for Flux model inference. ##### QNN EP Improvements - Introduced QNN shared memory support. - Improved performance for AI Hub models. - Added support for QAIRT/QNN SDK 2.31. - Added Python 3.13 package. - Miscellaneous bug fixes and enhancements. - QNN EP is now built as a shared library/DLL by default. To retain previous build behavior, use build option `--use_qnn static_lib`. ##### DirectML EP Support & Upgrades - Updated DirectML version from 1.15.2 to 1.15.4([#​22635](https://redirect.github.com/microsoft/onnxruntime/issues/22635)). ##### OpenVINO EP Improvements - Introduced OpenVINO EP Weights Sharing feature. - Added support for various contrib Ops in OVEP: - `SkipLayerNormalization`, `MatMulNBits`, `FusedGemm`, `FusedConv`, `EmbedLayerNormalization`, `BiasGelu`, `Attention`, `DynamicQuantizeMatMul`, `FusedMatMul`, `QuickGelu`, `SkipSimplifiedLayerNormalization` - Miscellaneous bug fixes and improvements. ##### VitisAI EP Improvements - Miscellaneous bug fixes and improvements. #### Mobile Platform Enhancements ##### CoreML Updates - Added support for caching generated CoreML models. #### Extensions & Tokenizer Improvements ##### Expanded Tokenizer Support - Now supports more tokenizer models, including `ChatGLM`, `Baichuan2`, `Phi-4`, etc. - Added full `Phi-4` pre/post-processing support for text, vision, and audio. - Introduced RegEx pattern loading from `tokenizer.json`. ##### Image Codec Enhancements - `ImageCodec` now links to native APIs if available; otherwise, falls back to built-in libraries. ##### Unified Tokenizer API - Introduced a new tokenizer op schema to unify the tokenizer codebase. - Added support for loading tokenizer data from a memory blob in the C API. #### Infrastructure & Build Improvements ##### Runtime Requirements All the prebuilt Windows packages now require VC++ Runtime version >= 14.40(instead of 14.38). If your VC++ runtime version is lower than that, you may see a crash when ONNX Runtime was initializing. See <https://github.com/microsoft/STL/wiki/Changelog#vs-2022-1710> for more details. Updated minimum iOS and Android SDK requirements to align with React Native 0.76: - iOS >= [15.1](https://support.apple.com/en-gb/108051#151) - Android API >= [24](https://developer.android.com/tools/releases/platforms#7.0) (Android 7) All macOS packages now require macOS version >= 13.3. ##### CMake File Changes CMake Version: Increased the minimum required CMake version from 3.26 to 3.28. Added support for CMake 4.0. Python Version: Increased the minimum required Python version from 3.8 to 3.10 for building ONNX Runtime from source. Improved VCPKG support Added the following cmake options for WebGPU EP - onnxruntime\_USE\_EXTERNAL\_DAWN - onnxruntime\_CUSTOM\_DAWN\_SRC\_PATH - onnxruntime\_BUILD\_DAWN\_MONOLITHIC\_LIBRARY - onnxruntime\_ENABLE\_PIX\_FOR\_WEBGPU\_EP - onnxruntime\_ENABLE\_DAWN\_BACKEND\_VULKAN - onnxruntime\_ENABLE\_DAWN\_BACKEND\_D3D12 Added cmake option onnxruntime\_BUILD\_QNN\_EP\_STATIC\_LIB for building with QNN EP as a static library. Removed cmake option onnxruntime\_USE\_PREINSTALLED\_EIGEN. Fixed a build issue with Visual Studio 2022 17.3 ([#​23911](https://redirect.github.com/microsoft/onnxruntime/issues/23911)) ##### Modernized Build Tools - Now using VCPKG for most package builds. - Upgraded Gradle from 7.x to 8.x. - Updated JDK from 11 to 17. - Enabled `onnxruntime_USE_CUDA_NHWC_OPS` by default for CUDA builds. - Added support for WASM64 (build from source; no package published). ##### Dependency Cleanup - Removed Google’s `nsync` from dependencies. ##### Others Updated Node.js installation script to support network proxy usage ([#​23231](https://redirect.github.com/microsoft/onnxruntime/issues/23231)) #### Web - No updates of note. ### Contributors Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members: Changming Sun, Yulong Wang, Tianlei Wu, Jian Chen, Wanming Lin, Adrian Lizarraga, Hector Li, Jiajia Qin, Yifan Li, Edward Chen, Prathik Rao, Jing Fang, shiyi, Vincent Wang, Yi Zhang, Dmitri Smirnov, Satya Kumar Jandhyala, Caroline Zhu, Chi Lo, Justin Chu, Scott McKay, Enrico Galli, Kyle, Ted Themistokleous, dtang317, wejoncy, Bin Miao, Jambay Kinley, Sushanth Rajasankar, Yueqing Zhang, amancini-N, ivberg, kunal-vaishnavi, liqun Fu, Corentin Maravat, Peishen Yan, Preetha Veeramalai, Ranjit Ranjan, Xavier Dupré, amarin16, jzm-intel, kailums, xhcao, A-Satti, Aleksei Nikiforov, Ankit Maheshkar, Javier Martinez, Jianhui Dai, Jie Chen, Jon Campbell, Karim Vadsariya, Michael Tyler, PARK DongHa, Patrice Vignola, Pranav Sharma, Sam Webster, Sophie Schoenmeyer, Ti-Tai Wang, Xu Xing, Yi-Hong Lyu, genmingz\@​AMD, junchao-zhao, sheetalarkadam, sushraja-msft, Akshay Sonawane, Alexis Tsogias, Ashrit Shetty, Bilyana Indzheva, Chen Feiyue, Christian Larson, David Fan, David Hotham, Dmitry D eshevoy, Frank Dong, Gavin Kinsey, George Wu, Grégoire, Guenther Schmuelling, Indy Zhu, Jean-Michaël Celerier, Jeff Daily, Joshua Lochner, Kee, Malik Shahzad Muzaffar, Matthieu Darbois, Michael Cho, Michael Sharp, Misha Chornyi, Po-Wei (Vincent), Sevag H, Takeshi Watanabe, Wu, Junze, Xiang Zhang, Xiaoyu, Xinpeng Dou, Xinya Zhang, Yang Gu, Yateng Hong, mindest, mingyue, raoanag, saurabh, shaoboyan091, sstamenk, tianf-fff, wonchung-microsoft, xieofxie, zz002 </details> --- ### Configuration 📅 **Schedule**: Branch creation - Every minute ( * * * * * ) (UTC), Automerge - At any time (no schedule defined). 🚦 **Automerge**: Disabled by config. Please merge this manually once you are satisfied. ♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 **Ignore**: Close this PR and you won't be reminded about this update again. --- - [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check this box --- This PR has been generated by [Renovate Bot](https://redirect.github.com/solrbot/renovate-github-action) <!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiI0MS44Mi4xMCIsInVwZGF0ZWRJblZlciI6IjQxLjgyLjEwIiwidGFyZ2V0QnJhbmNoIjoibWFpbiIsImxhYmVscyI6WyJleGVtcHQtc3RhbGUiXX0=--> -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
