Releases: huggingface/optimum
v1.21.2: Patch release
- Remove inplace op in mistral patcher by @IlyasMoutawwakil in #1938
- Fix ORTModelForFeatureExtraction modeling by @moria97 in #1941
Full Changelog: v1.21.1...v1.21.2
v1.21.1: Patch release
- Fix sentence transformers model patching by @echarlaix in #1936
- Update Intel extra by @echarlaix in #1935
- Update Habana extra by @regisss in #1937
Full Changelog: v1.21.0...v1.21.1
v1.21.0: many bugfixes, transformers 4.42 compatibility
What's Changed
- ORTOptimizer for the model type Segformer by @zachmayer in #1820
- fix: device consistence by @Daya-Jin in #1891
- Allow optimum to discover and load subpackages by @dacorvo in #1894
- feat(ci): add trufflehog secrets detector by @McPatate in #1899
- fix(ci): remove unnecessary permissions by @McPatate in #1904
- Remove read token by @fxmarty in #1903
- Remove dataset with restrictive license by @echarlaix in #1910
- Fix Windows and onnx dtype compatibility by @IlyasMoutawwakil in #1886
- Deprecated
use_auth_token
by @IlyasMoutawwakil in #1837 - Add redirection for optimum intel doc by @echarlaix in #1918
- Read use_external_data_format from ORTConfig file by @idruker-cerence in #1917
- Pin numpy v1 for onnxruntime by @IlyasMoutawwakil in #1921
- Fix GPTQ CI by @IlyasMoutawwakil in #1878
- Fix code quality by @IlyasMoutawwakil in #1928
- Fix incorrect names for usage blenderbot for causallm by @eaidova in #1887
- Fixed bug key error "last_hidden_state" by @satishsilveri in #1674
- Support transformers 4.42 by @fxmarty in #1929
New Contributors
- @zachmayer made their first contribution in #1820
- @Daya-Jin made their first contribution in #1891
- @dacorvo made their first contribution in #1894
- @McPatate made their first contribution in #1899
- @idruker-cerence made their first contribution in #1917
- @satishsilveri made their first contribution in #1674
Full Changelog: v1.20.0...v1.21.0
v1.20.0: VITS, Phi-3 ONNX export
Extended ONNX export
- VITS ONNX export by @echarlaix in #1607
- Phi-3 ONNX export by @JingyaHuang in #1870
- Add Phi-3 normalized config by @kunal-vaishnavi in #1841
- Add Phi-3 small normalized config by @JingyaHuang in #1864
Other changes and bugfixes
-
Bump transformers version by @echarlaix in #1824
-
Remove call to
apt update
beforeapt purge
in the main doc build workflow by @regisss in #1830 -
Update github workflows by @echarlaix in #1829
-
Remove bad PPA in main doc build workflow by @regisss in #1831
-
Fix sentence transformers models infer library by @echarlaix in #1832
-
Fix random initialization of bias when using GPTQ quantization with models without bias by @B-201 in #1827
-
Update the Transformers dependency in the Habana extra by @regisss in #1851
-
Make stable diffusion unet and vae number of channels static by @eaidova in #1840
-
Fix compatibility with transformers v4.41.0 for ONNX by @echarlaix in #1860
-
Fix FX CI by @IlyasMoutawwakil in #1866
-
Fix Utils CI by @IlyasMoutawwakil in #1867
-
Fix BT CI by @IlyasMoutawwakil in #1872
-
Fix ORTConfig loading by @mr-sarthakgupta in #1879
-
Update ORT doc for ROCM 6.0 by @mht-sharma in #1862
-
Fix ort config instantiation (from_pretrained) and saving (save_pretrained) by @IlyasMoutawwakil in #1865
-
Fix ORT CI by @IlyasMoutawwakil in #1875
-
Update optimum intel extra by @echarlaix in #1882
-
Bump transformers version for neuron extras by @JingyaHuang in #1881
New Contributors
- @B-201 made their first contribution in #1827
- @mr-sarthakgupta made their first contribution in #1879
Full Changelog: v1.19.0...v1.20.0
v1.19.2: Patch release
Full Changelog: v1.19.1...v1.19.2
v1.19.1: Patch release
- Bump transformers version by @echarlaix in #1824
- Remove call to
apt update
beforeapt purge
in the main doc build workflow by @regisss in #1830
Full Changelog: v1.19.0...v1.19.1
v1.19.0: Musicgen, MarkupLM ONNX export
Extended ONNX export
Musicgen and MarkupLM models from Transformers can now be exported to ONNX through optimum-cli export onnx
. Musicgen ONNX export is used to run the model locally in a browser through transformers.js.
- Musicgen ONNX export (text-conditional only) by @fxmarty in #1779
- Add support for markuplm ONNX export by @pogzyb in #1784
Other changes and bugfixes
- Fix IR version for merged ONNX decoders by @fxmarty in #1780
- Update test model id by @echarlaix in #1785
- Add Nvidia and Neuron to README by @JingyaHuang in #1791
- adds debug options to dump onnx graphs by @prathikr in #1789
- Improve PR template by @fxmarty in #1799
- Add Google TPU to the mix by @mfuntowicz in #1797
- Add redirection for Optimum TPU by @regisss in #1801
- Add Nvidia and Neuron to the installation doc by @JingyaHuang in #1803
- Update installation instructions by @echarlaix in #1806
- Fix offline compatibility by @fxmarty in #1805
- Remove unnecessary constants for > 2GB ONNX models by @fxmarty in #1808
- Add onnx export function for pix2struct model by @naormatania in #1815
New Contributors
- @pogzyb made their first contribution in #1784
- @naormatania made their first contribution in #1815
Full Changelog: v1.18.0...v1.19.0
v1.18.1: Patch release
Fix the installation for Optimum Neuron v0.0.21 release
- Improve the installation of optimum-neuron through optimum extras #1778
Fix the task inference of stable diffusion
- Fix infer task for stable diffusion #1793
Full Changelog: v1.18.0...v1.18.1
v1.18.0: Gemma, OWLv2, MPNet Qwen2 ONNX support
New architectures ONNX export :
- OWLv2 by @xenova in #1689
- Gemma by @fxmarty in #1714
- MPNet by @nathan-az in #1471
- Qwen2 by @uniartisan in #1746
Other changes and bugfixes
v1.17.1: Patch release
Update Transformers dependency for the release of Optimum Habana v1.10.2
- Update Transformers dependency in Habana extra #1700
Full Changelog: v1.17.0...v1.17.1