solrbot opened a new pull request, #3657:
URL: https://github.com/apache/solr/pull/3657

   This PR contains the following updates:
   
   | Package | Type | Update | Change |
   |---|---|---|---|
   | 
[com.microsoft.onnxruntime:onnxruntime](https://microsoft.github.io/onnxruntime/)
 ([source](https://redirect.github.com/microsoft/onnxruntime)) | dependencies | 
minor | `1.20.0` -> `1.22.0` |
   
   ---
   
   ### Release Notes
   
   <details>
   <summary>microsoft/onnxruntime 
(com.microsoft.onnxruntime:onnxruntime)</summary>
   
   ### 
[`v1.22.0`](https://redirect.github.com/microsoft/onnxruntime/releases/tag/v1.22.0):
 ONNX Runtime v1.22
   
   #### Announcements
   
   - This release introduces new API's for Model Editor, Auto EP 
infrastructure, and AOT Compile
   - OnnxRuntime GPU packages require CUDA 12.x  ,  packages built for CUDA 
11.x are no longer published.
   - The min supported Windows version is now 10.0.19041.
   
   #### GenAI & Advanced Model Features
   
   - **Constrained Decoding:** Introduced new capabilities for constrained 
decoding, offering more control over generative AI model outputs.
   
   #### Execution & Core Optimizations
   
   ##### Core
   
   - **Auto EP Selection Infrastructure:** Added foundational infrastructure to 
enable automatic selection of Execution Providers via selection policies, 
aiming to simplify configuration and optimize performance. (Pull Request 
[#&#8203;24430](https://redirect.github.com/microsoft/onnxruntime/issues/24430))
   - **Compile API:** Introduced new APIs to support explicit compilation of 
ONNX models.
     - See: [OrtCompileApi Struct 
Reference](https://onnxruntime.ai/docs/api/c/struct_ort_compile_api.html) 
(Assuming a similar link structure for future documentation)
     - See: [EP Context 
Design](https://onnxruntime.ai/docs/api/c/struct_ort_ep_context.html) (Assuming 
a similar link structure for future documentation)
   - **Model Editor API** api's for creating or editing ONNX models
     - See: 
[OrtModelEditorApi](https://onnxruntime.ai/docs/api/c/struct_ort_model_editor_api.html#details)
   
   ##### Execution Provider (EP) Updates
   
   ##### CPU EP/MLAS
   
   - **KleidiAI Integration:** Integrated KleidiAI into ONNX Runtime/MLAS for 
enhanced performance on Arm architectures.
   - **MatMulNBits Support:** Added support for `MatMulNBits`, enabling matrix 
multiplication with weights quantized to 8 bits.
   - **GroupQueryAttention optimizations and enhancements**
   
   ##### OpenVINO EP
   
   - Added support up to OpenVINO 2025.1
   - Introduced Intel compiler level optimizations for QDQ models.
   - Added support to select Intel devices based on LUID
   - Load\_config feature improvement to support AUTO, HETERO and MULTI plugin.
   - misc bugfixes/optimizations
   - For detailed updates, refer to Pull Request 
[#&#8203;24394](https://redirect.github.com/microsoft/onnxruntime/issues/24394):
 [ONNXRuntime OpenVINO - Release 
1.22](https://redirect.github.com/microsoft/onnxruntime/pull/24394)
   
   ##### QNN EP
   
   - **SDK Update:** Added support for QNN SDK 2.33.2.
   - operator updates/support to Sum, Softmax, Upsample, Expand, ScatterND, 
Einsum
   - QNN EP can be built as shared or static library.
   - enable QnnGpu backend
   - For detailed updates refer to [recent QNN tagged 
PR's](https://redirect.github.com/microsoft/onnxruntime/pulls?q=is%3Apr+qnn+ep+is%3Aclosed+label%3Aep%3AQNN)
   
   ##### TensorRT EP
   
   - **TensorRT Version:** Added support for TensorRT 10.9.
     - **Note for onnx-tensorrt open-source parser users:** Please check 
[here](https://onnxruntime.ai/docs/build/eps.html#note-to-ort-1210-open-sourced-parser-users)
 for specific requirements (Referencing 1.21 link as a placeholder, this should 
be updated for 1.22).
   - **New Features:**
     - EP option to enable TRT Preview Feature
     - Support to load TensorRT V3 plugin
   - **Bug Fixes:**
     - Resolved an issue related to multithreading scenarios.
     - Fixed incorrect GPU usage that affected both TensorRT EP and CUDA EP.
   
   ##### NV TensorRT RTX EP
   
   - **New Execution Provider:** Introduced a new Execution Provider 
specifically for Nvidia RTX GPUs, leveraging TensorRT for optimized performance.
   
   ##### CUDA EP
   
   - **MatMulNBits Enhancement:** Added support for 8-bit weight-only 
quantization in `MatMulNBits`.
   - **Bug Fixes:**
     - Fixed incorrect GPU usage (also mentioned under TensorRT EP).
   
   ##### VitisAI EP
   
   - Miscellaneous bug fixes and improvements.
   
   #### Infrastructure & Build Improvements
   
   ##### Build System & Packages
   
   - **QNN Nuget Package:** The QNN Nuget package is now built as ARM64x.
   
   ##### Dependencies / Version Updates
   
   - **CUDA Version Update:** This release includes an update to the CUDA 
version. Users should consult the documentation for specific version 
requirements. CUDA 11 based GPU packages no longer published.
   
   #### Web
   
   - **WebGPU Expansion:**
     - Added WebGPU support to the node.js package (Windows and macOS).
     - Enabled WebGPU when building from source for macOS, Linux, and Windows.
   
   #### Mobile
   
   - No major updates of note this release.
   
   ### Contributions
   
   Contributors to ONNX Runtime include members across teams at Microsoft, 
along with our community members:
   
   Yulong Wang, Jian Chen, Changming Sun, Satya Kumar Jandhyala, Hector Li, 
Prathik Rao, Adrian Lizarraga, Jiajia Qin, Scott McKay, Jie Chen, Tianlei Wu, 
Edward Chen, Wanming Lin, xhcao, vraspar, Dmitri Smirnov, Jing Fang, Yifan Li, 
Caroline Zhu, Jianhui Dai, Chi Lo, Guenther Schmuelling, Ryan Hill, Sushanth 
Rajasankar, Yi-Hong Lyu, Ankit Maheshkar, Artur Wojcik, Baiju Meswani, David 
Fan, Enrico Galli, Hans, Jambay Kinley, John Paul, Peishen Yan, Yateng Hong, 
amarin16, chuteng-quic, kunal-vaishnavi, quic-hungjuiw, Alessio Soldano, 
Andreas Hussing, Ashish Garg, Ashwath Shankarnarayan, Chengdong Liang, Clément 
Péron, Erick Muñoz, Fanchen Kong, George Wu, Haik Silm, Jagadish 
Krishnamoorthy, Justin Chu, Karim Vadsariya, Kevin Chen, Mark Schofield, 
Masaya, Kato, Michael Tyler, Nenad Banfic, Ningxin Hu, Praveen G, Preetha 
Veeramalai, Ranjit Ranjan, Seungtaek Kim, Ti-Tai Wang, Xiaofei Han, Yueqing 
Zhang, co63oc, derdeljan-msft, genmingz\@&#8203;AMD, jiangzhaoming, jing-bao, 
kuanyul-quic,
  liqun Fu, minfhong-quic, mingyue, quic-tirupath, quic-zhaoxul, saurabh, 
selenayang888, sfatimar, sheetalarkadam, virajwad, zz002, Ștefan Talpalaru
   
   ### 
[`v1.21.1`](https://redirect.github.com/microsoft/onnxruntime/releases/tag/v1.21.1):
 ONNX Runtime v1.21.1
   
   ### What's new?
   
   - Extend CMAKE\_CUDA\_FLAGS with all Blackwell compute capacity  
[#&#8203;23928](https://redirect.github.com/microsoft/onnxruntime/issues/23928) 
- [@&#8203;yf711](https://redirect.github.com/yf711)
   - \[ARM CPU] Fix fp16 const initialization on no-fp16 platform 
[#&#8203;23978](https://redirect.github.com/microsoft/onnxruntime/issues/23978) 
- [@&#8203;fajin-corp](https://redirect.github.com/fajin-corp)
   - \[TensorRT EP] Call cudaSetDevice at compute function for handling 
multithreading scenario 
[#&#8203;24010](https://redirect.github.com/microsoft/onnxruntime/issues/24010) 
- [@&#8203;chilo-ms](https://redirect.github.com/chilo-ms)
   - Fix attention bias broadcast 
[#&#8203;24017](https://redirect.github.com/microsoft/onnxruntime/issues/24017) 
- [@&#8203;tianleiwu](https://redirect.github.com/tianleiwu)
   - Deleted the constant SKIP\_CUDA\_TEST\_WITH\_DML 
[#&#8203;24113](https://redirect.github.com/microsoft/onnxruntime/issues/24113) 
- [@&#8203;CodingSeaotter](https://redirect.github.com/CodingSeaotter)
   - \[QNN EP] ARM64EC python package remove --vcpkg in build 
[#&#8203;24174](https://redirect.github.com/microsoft/onnxruntime/issues/24174) 
- [@&#8203;jywu-msft](https://redirect.github.com/jywu-msft)
   - \[wasm] remove --vcpkg in wasm build 
[#&#8203;24179](https://redirect.github.com/microsoft/onnxruntime/issues/24179) 
- [@&#8203;fs-eire](https://redirect.github.com/fs-eire)
   
   ### 
[`v1.21.0`](https://redirect.github.com/microsoft/onnxruntime/releases/tag/v1.21.0):
 ONNX Runtime v1.21.0
   
   #### Announcements
   
   - No large announcements of note this release! We've made a lot of small 
refinements to streamline your ONNX Runtime experience.
   
   #### GenAI & Advanced Model Features
   
   ##### Enhanced Decoding & Pipeline Support
   
   - Added "chat mode" support for CPU, GPU, and WebGPU.
   - Provided support for decoder model pipelines.
   - Added support for Java API for MultiLoRA.
   
   ##### API & Compatibility Updates
   
   - Chat mode introduced breaking changes in the API (see [migration 
guide](https://onnxruntime.ai/docs/genai/howto/migrate.html)).
   
   ##### Bug Fixes for Model Output
   
   - Fixed Phi series garbage output issues with long prompts.
   - Resolved gibberish issues with `top_k` on CPU.
   
   #### Execution & Core Optimizations
   
   ##### Core Refinements
   
   - Reduced default logger usage for improved 
efficiency([#&#8203;23030](https://redirect.github.com/microsoft/onnxruntime/issues/23030)).
   - Fixed a visibility issue in theadpool 
([#&#8203;23098](https://redirect.github.com/microsoft/onnxruntime/issues/23098)).
   
   ##### Execution Provider (EP) Updates
   
   ##### General
   
   - Removed TVM EP from the source 
tree([#&#8203;22827](https://redirect.github.com/microsoft/onnxruntime/issues/22827)).
   - Marked NNAPI EP for deprecation (following Google's deprecation of NNAPI).
   - Fixed a DLL delay loading issue that impacts WebGPU EP and DirectML EP's 
usability on Windows 
([#&#8203;23111](https://redirect.github.com/microsoft/onnxruntime/issues/23111),
 
[#&#8203;23227](https://redirect.github.com/microsoft/onnxruntime/issues/23227))
   
   ##### TensorRT EP Improvements
   
   - Added support for TensorRT 10.8.
     - [onnx-tensorrt](https://redirect.github.com/onnx/onnx-tensorrt) 
open-source parser user: please check 
[here](https://onnxruntime.ai/docs/build/eps.html#note-to-ort-1210-open-sourced-parser-users)
 for requirement.
   - Assigned DDS ops (`NMS`, `RoiAlign`, `NonZero`) to TensorRT by default.
   - Introduced option `trt_op_types_to_exclude` to exclude specific ops from 
TensorRT assignment.
   
   ##### CUDA EP Improvements
   
   - Added a python API 
[preload\_dlls](https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#preload-dlls)
 to [coexist with 
PyTorch](https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#compatibility-with-pytorch).
   - Miscellaneous enhancements for Flux model inference.
   
   ##### QNN EP Improvements
   
   - Introduced QNN shared memory support.
   - Improved performance for AI Hub models.
   - Added support for QAIRT/QNN SDK 2.31.
   - Added Python 3.13 package.
   - Miscellaneous bug fixes and enhancements.
   - QNN EP is now built as a shared library/DLL by default. To retain previous 
build behavior, use build option `--use_qnn static_lib`.
   
   ##### DirectML EP Support & Upgrades
   
   - Updated DirectML version from 1.15.2 to 
1.15.4([#&#8203;22635](https://redirect.github.com/microsoft/onnxruntime/issues/22635)).
   
   ##### OpenVINO EP Improvements
   
   - Introduced OpenVINO EP Weights Sharing feature.
   - Added support for various contrib Ops in OVEP:
     - `SkipLayerNormalization`, `MatMulNBits`, `FusedGemm`, `FusedConv`, 
`EmbedLayerNormalization`, `BiasGelu`, `Attention`, `DynamicQuantizeMatMul`, 
`FusedMatMul`, `QuickGelu`, `SkipSimplifiedLayerNormalization`
   - Miscellaneous bug fixes and improvements.
   
   ##### VitisAI EP Improvements
   
   - Miscellaneous bug fixes and improvements.
   
   #### Mobile Platform Enhancements
   
   ##### CoreML Updates
   
   - Added support for caching generated CoreML models.
   
   #### Extensions & Tokenizer Improvements
   
   ##### Expanded Tokenizer Support
   
   - Now supports more tokenizer models, including `ChatGLM`, `Baichuan2`, 
`Phi-4`, etc.
   - Added full `Phi-4` pre/post-processing support for text, vision, and audio.
   - Introduced RegEx pattern loading from `tokenizer.json`.
   
   ##### Image Codec Enhancements
   
   - `ImageCodec` now links to native APIs if available; otherwise, falls back 
to built-in libraries.
   
   ##### Unified Tokenizer API
   
   - Introduced a new tokenizer op schema to unify the tokenizer codebase.
   - Added support for loading tokenizer data from a memory blob in the C API.
   
   #### Infrastructure & Build Improvements
   
   ##### Runtime Requirements
   
   All the prebuilt Windows packages now require VC++ Runtime version >= 
14.40(instead of 14.38).  If your VC++ runtime version is lower than that, you 
may see a crash when ONNX Runtime was initializing. See 
<https://github.com/microsoft/STL/wiki/Changelog#vs-2022-1710> for more details.
   
   Updated minimum iOS and Android SDK requirements to align with React Native 
0.76:
   
   - iOS  >=  [15.1](https://support.apple.com/en-gb/108051#151)
   - Android API >= 
[24](https://developer.android.com/tools/releases/platforms#7.0) (Android 7)
   
   All macOS packages now require macOS version >= 13.3.
   
   ##### CMake File Changes
   
   CMake Version: Increased the minimum required CMake version from 3.26 to 
3.28. Added support for CMake 4.0.
   Python Version: Increased the minimum required Python version from 3.8 to 
3.10 for building ONNX Runtime from source.
   Improved VCPKG support
   
   Added the following cmake options for WebGPU EP
   
   - onnxruntime\_USE\_EXTERNAL\_DAWN
   - onnxruntime\_CUSTOM\_DAWN\_SRC\_PATH
   - onnxruntime\_BUILD\_DAWN\_MONOLITHIC\_LIBRARY
   - onnxruntime\_ENABLE\_PIX\_FOR\_WEBGPU\_EP
   - onnxruntime\_ENABLE\_DAWN\_BACKEND\_VULKAN
   - onnxruntime\_ENABLE\_DAWN\_BACKEND\_D3D12
   
   Added cmake option onnxruntime\_BUILD\_QNN\_EP\_STATIC\_LIB for building 
with QNN EP as a static library.
   Removed cmake option onnxruntime\_USE\_PREINSTALLED\_EIGEN.
   
   Fixed a build issue with Visual Studio 2022 17.3 
([#&#8203;23911](https://redirect.github.com/microsoft/onnxruntime/issues/23911))
   
   ##### Modernized Build Tools
   
   - Now using VCPKG for most package builds.
   - Upgraded Gradle from 7.x to 8.x.
   - Updated JDK from 11 to 17.
   - Enabled `onnxruntime_USE_CUDA_NHWC_OPS` by default for CUDA builds.
   - Added support for WASM64 (build from source; no package published).
   
   ##### Dependency Cleanup
   
   - Removed Google’s `nsync` from dependencies.
   
   ##### Others
   
   Updated Node.js installation script to support network proxy usage 
([#&#8203;23231](https://redirect.github.com/microsoft/onnxruntime/issues/23231))
   
   #### Web
   
   - No updates of note.
   
   ### Contributors
   
   Contributors to ONNX Runtime include members across teams at Microsoft, 
along with our community members:
   
   Changming Sun, Yulong Wang, Tianlei Wu, Jian Chen, Wanming Lin, Adrian 
Lizarraga, Hector Li, Jiajia Qin, Yifan Li, Edward Chen, Prathik Rao, Jing 
Fang, shiyi, Vincent Wang, Yi Zhang, Dmitri Smirnov, Satya Kumar Jandhyala, 
Caroline Zhu, Chi Lo, Justin Chu, Scott McKay, Enrico Galli, Kyle, Ted 
Themistokleous, dtang317, wejoncy, Bin Miao, Jambay Kinley, Sushanth 
Rajasankar, Yueqing Zhang, amancini-N, ivberg, kunal-vaishnavi, liqun Fu, 
Corentin Maravat, Peishen Yan, Preetha Veeramalai, Ranjit Ranjan, Xavier Dupré, 
amarin16, jzm-intel, kailums, xhcao, A-Satti, Aleksei Nikiforov, Ankit 
Maheshkar, Javier Martinez, Jianhui Dai, Jie Chen, Jon Campbell, Karim 
Vadsariya, Michael Tyler, PARK DongHa, Patrice Vignola, Pranav Sharma, Sam 
Webster, Sophie Schoenmeyer, Ti-Tai Wang, Xu Xing, Yi-Hong Lyu, 
genmingz\@&#8203;AMD, junchao-zhao, sheetalarkadam, sushraja-msft, Akshay 
Sonawane, Alexis Tsogias, Ashrit Shetty, Bilyana Indzheva, Chen Feiyue, 
Christian Larson, David Fan, David Hotham, Dmitry D
 eshevoy, Frank Dong, Gavin Kinsey, George Wu, Grégoire, Guenther Schmuelling, 
Indy Zhu, Jean-Michaël Celerier, Jeff Daily, Joshua Lochner, Kee, Malik Shahzad 
Muzaffar, Matthieu Darbois, Michael Cho, Michael Sharp, Misha Chornyi, Po-Wei 
(Vincent), Sevag H, Takeshi Watanabe, Wu, Junze, Xiang Zhang, Xiaoyu, Xinpeng 
Dou, Xinya Zhang, Yang Gu, Yateng Hong, mindest, mingyue, raoanag, saurabh, 
shaoboyan091, sstamenk, tianf-fff, wonchung-microsoft, xieofxie, zz002
   
   </details>
   
   ---
   
   ### Configuration
   
   📅 **Schedule**: Branch creation - Every minute ( * * * * * ) (UTC), 
Automerge - At any time (no schedule defined).
   
   🚦 **Automerge**: Disabled by config. Please merge this manually once you are 
satisfied.
   
   ♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry 
checkbox.
   
   🔕 **Ignore**: Close this PR and you won't be reminded about this update 
again.
   
   ---
   
    - [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check this 
box
   
   ---
   
   This PR has been generated by [Renovate 
Bot](https://redirect.github.com/solrbot/renovate-github-action)
   
<!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiI0MS44Mi4xMCIsInVwZGF0ZWRJblZlciI6IjQxLjgyLjEwIiwidGFyZ2V0QnJhbmNoIjoibWFpbiIsImxhYmVscyI6WyJleGVtcHQtc3RhbGUiXX0=-->
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to