-
Notifications
You must be signed in to change notification settings - Fork 576
Comparing changes
Open a pull request
base repository: huggingface/optimum
base: v1.26.1
head repository: huggingface/optimum
compare: v1.27.0
- 20 commits
- 54 files changed
- 10 contributors
Commits on Jun 17, 2025
-
Bump protobuf for onnxruntime training example (#2292)
Bump protobuf in /examples/onnxruntime/training/language-modeling Bumps [protobuf](https://github.com/protocolbuffers/protobuf) from 3.20.2 to 4.25.8. - [Release notes](https://github.com/protocolbuffers/protobuf/releases) - [Changelog](https://github.com/protocolbuffers/protobuf/blob/main/protobuf_release.bzl) - [Commits](protocolbuffers/protobuf@v3.20.2...v4.25.8) --- updated-dependencies: - dependency-name: protobuf dependency-version: 4.25.8 dependency-type: direct:production ... Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 0c6dbd4 - Browse repository at this point
Copy the full SHA 0c6dbd4View commit details -
Add task onnx register (#2291)
* Move task register in onnx exporters * fix * remove unused * fix * fix
Configuration menu - View commit details
-
Copy full SHA for de10347 - Browse repository at this point
Copy the full SHA de10347View commit details -
No more forcing separators (#2279)
* use model_types as god intended for them * use the names transformers has given them * fix perceiver * fix perceiver text vision distinction * fix other models * add reference to official model types in transformers * use automated script * data2vec * remove duplicate * fix * fix * runscript with latest transformers * add fix chinese clip * apply script * fix --------- Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com> Co-authored-by: Ella Charlaix <ella@huggingface.co>
Configuration menu - View commit details
-
Copy full SHA for 5246fc8 - Browse repository at this point
Copy the full SHA 5246fc8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 05aef41 - Browse repository at this point
Copy the full SHA 05aef41View commit details
Commits on Jun 18, 2025
-
ExporterConfig refactorization (#2157)
* add type checking * tflite config exporters * add exporters config * fix style * fix * fix * remove constant * fix style * move * rm comments * fix masked-im output names * style * rm comment * fix doc generation * task attribute already set * model patcher * add model patcher * add processor * move patching specs to onnx config * rename * add comment * fix * style * rename * style * remove todo comments --------- Co-authored-by: IlyasMoutawwakil <moutawwakil.ilyas.tsi@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for be76dde - Browse repository at this point
Copy the full SHA be76ddeView commit details
Commits on Jun 19, 2025
-
Remove flatten_inputs from taskmanager (#2296)
* remove unused method * remove unused
Configuration menu - View commit details
-
Copy full SHA for 824e368 - Browse repository at this point
Copy the full SHA 824e368View commit details -
Fix 'Block pattern could not be match. Pass
block_name_to_quantize
……argument in `quantize_model`' while loading Qwen VL GPTQ model (#2295) Updated BLOCK_PATTERNS
Configuration menu - View commit details
-
Copy full SHA for ed70d27 - Browse repository at this point
Copy the full SHA ed70d27View commit details
Commits on Jun 20, 2025
-
Remove timm from exporters extra (#2299)
remove timm from required dependencies
Configuration menu - View commit details
-
Copy full SHA for cf6d4cc - Browse repository at this point
Copy the full SHA cf6d4ccView commit details
Commits on Jun 23, 2025
-
Configuration menu - View commit details
-
Copy full SHA for ffd91d8 - Browse repository at this point
Copy the full SHA ffd91d8View commit details
Commits on Jun 24, 2025
-
Adding DummyInputGenerator for VisualBert (#2303)
Co-authored-by: Abdennacer Badaoui <abdennacerbadaoui@MacBook-Air-de-Abdennacer.local>
Configuration menu - View commit details
-
Copy full SHA for a4152c0 - Browse repository at this point
Copy the full SHA a4152c0View commit details
Commits on Jun 25, 2025
-
Configuration menu - View commit details
-
Copy full SHA for eb187fa - Browse repository at this point
Copy the full SHA eb187faView commit details -
Configuration menu - View commit details
-
Copy full SHA for f7d6e88 - Browse repository at this point
Copy the full SHA f7d6e88View commit details
Commits on Jun 27, 2025
-
add dummy input generator for VisualBert (#2307)
* add dummy input generator * fix style * add default shapes * Apply suggestions from code review --------- Co-authored-by: Ilyas Moutawwakil <57442720+IlyasMoutawwakil@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for fb6016d - Browse repository at this point
Copy the full SHA fb6016dView commit details -
Configuration menu - View commit details
-
Copy full SHA for 70aaf48 - Browse repository at this point
Copy the full SHA 70aaf48View commit details -
Configuration menu - View commit details
-
Copy full SHA for ec756fd - Browse repository at this point
Copy the full SHA ec756fdView commit details
Commits on Jul 10, 2025
-
Deprecate tflite, ort-training and bettertransformer (#2316)
* deprecate tflite * bt and ort training
Configuration menu - View commit details
-
Copy full SHA for 9564248 - Browse repository at this point
Copy the full SHA 9564248View commit details
Commits on Jul 23, 2025
-
Transformers 4.53 support, SmolLM3 and Fix old Transformers support (#…
…2319) * test * fix seq2seq patched sdpa * patch qwen3_moe, out_attentions, and eager_mask * fix * use optimum model * editable subpackages * Apply suggestions from code review * smollm3 support * deprecate tensorflow onnx export and add smollm3 to export tests * write a more general sdpa_mask without vmap that's also vectorized and broadcastable * better and more generic sdpa_mask_without_vmap implementation * style and fix * fix * patch find_packed_sequence_indices as it's untraceable * fix * fix * revert tests removal until refactor * fix temporary hub repo import * fix * fix external data tests on windows * update phi and phi3 min version * condition modernbert optimization test * get back old (pre 4.44) bloom modeling support and remove the need for normalized config in ort modeling * fix test was using hardcoded architecture * unparallelize test that uses remote code * support older versions of mpt and phi (4.36) * remove parallelism from slow tests * fix vision to text pipelines test * more specific version handling for find_packed_sequence_indices * fix
Configuration menu - View commit details
-
Copy full SHA for 53f39a6 - Browse repository at this point
Copy the full SHA 53f39a6View commit details
Commits on Jul 29, 2025
-
Propagate library_name parameter in from_pretrained to export (#2328)
* Propagate library_name parameter in from_pretrained to export Required to avoid automatic inferring of the library_name * Use class attribute for ORTModel instead Under modeling_diffusion it looks like ORTModel isn't used * Add test case * Update optimum/onnxruntime/modeling_ort.py --------- Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 689c0b5 - Browse repository at this point
Copy the full SHA 689c0b5View commit details
Commits on Jul 30, 2025
-
Fix batched inference/generation, position_ids creation, falcon alibi…
…, gpt_bigcode multi-query,.. (#2326) * test left-padded batched inference * demonstrate batched tex generation failure * fix remote code * fix * fix position_ids generation inside ORTModelForCausalLM class * it works until transformers 4.52 -_- * now run with latest transformers * bolean 4D mask is actually not supported by torch onnx exporter * only test generation with batched inputs, for logits are a bit off because of transformers using boolean mask * boolean mask safe softmax batched inference * style * use old typing * don't do unnecessary patching * try to avoid spamming the hub for an image * update min transformers version * better and direct torch patching * more batched generation special cases * style * initialize the il image instead of downloading it * use random pil image * test different versions of transformers in fast tests * fix * revert diffusers changes for now * mask padding kv cache as well * fix masking for old bloom * use constant image to image loading errors * style * test diffusers in series to avoid runner dying * fix * cleanup and some comments * fix and test falcon alibi * style * fix, support and test multi_query=False as well * only apply masked testing for transformers version previous to 4.39 * Update optimum/onnxruntime/modeling_decoder.py * use text decoder position ids onnx config but test its sync with list * fix opt * style * fix sdpa without overriting torch onnx exporter * use inplace op ;-; * fix st test * patch directly in onnx because patch needs to happen after softmax
Configuration menu - View commit details
-
Copy full SHA for 31d4ea9 - Browse repository at this point
Copy the full SHA 31d4ea9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5118493 - Browse repository at this point
Copy the full SHA 5118493View commit details
This comparison is taking too long to generate.
Unfortunately it looks like we can’t render this comparison for you right now. It might be too big, or there might be something weird with your repository.
You can try running this command locally to see the comparison on your machine:
git diff v1.26.1...v1.27.0