Skip to content
Permalink

Comparing changes

Choose two branches to see what’s changed or to start a new pull request. If you need to, you can also or learn more about diff comparisons.

Open a pull request

Create a new pull request by comparing changes across two branches. If you need to, you can also . Learn more about diff comparisons here.
base repository: huggingface/optimum
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: v1.26.1
Choose a base ref
...
head repository: huggingface/optimum
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: v1.27.0
Choose a head ref
  • 20 commits
  • 54 files changed
  • 10 contributors

Commits on Jun 17, 2025

  1. Bump protobuf for onnxruntime training example (#2292)

    Bump protobuf in /examples/onnxruntime/training/language-modeling
    
    Bumps [protobuf](https://github.com/protocolbuffers/protobuf) from 3.20.2 to 4.25.8.
    - [Release notes](https://github.com/protocolbuffers/protobuf/releases)
    - [Changelog](https://github.com/protocolbuffers/protobuf/blob/main/protobuf_release.bzl)
    - [Commits](protocolbuffers/protobuf@v3.20.2...v4.25.8)
    
    ---
    updated-dependencies:
    - dependency-name: protobuf
      dependency-version: 4.25.8
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <support@github.com>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored Jun 17, 2025
    Configuration menu
    Copy the full SHA
    0c6dbd4 View commit details
    Browse the repository at this point in the history
  2. Add task onnx register (#2291)

    * Move task register in onnx exporters
    
    * fix
    
    * remove unused
    
    * fix
    
    * fix
    echarlaix authored Jun 17, 2025
    Configuration menu
    Copy the full SHA
    de10347 View commit details
    Browse the repository at this point in the history
  3. No more forcing separators (#2279)

    * use model_types as god intended for them
    
    * use the names transformers has given them
    
    * fix perceiver
    
    * fix perceiver text vision distinction
    
    * fix other models
    
    * add reference to official model types in transformers
    
    * use automated script
    
    * data2vec
    
    * remove duplicate
    
    * fix
    
    * fix
    
    * runscript with latest transformers
    
    * add fix chinese clip
    
    * apply script
    
    * fix
    
    ---------
    
    Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
    Co-authored-by: Ella Charlaix <ella@huggingface.co>
    3 people authored Jun 17, 2025
    Configuration menu
    Copy the full SHA
    5246fc8 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    05aef41 View commit details
    Browse the repository at this point in the history

Commits on Jun 18, 2025

  1. ExporterConfig refactorization (#2157)

    * add type checking
    
    * tflite config exporters
    
    * add exporters config
    
    * fix style
    
    * fix
    
    * fix
    
    * remove constant
    
    * fix style
    
    * move
    
    * rm comments
    
    * fix masked-im output names
    
    * style
    
    * rm comment
    
    * fix doc generation
    
    * task attribute already set
    
    * model patcher
    
    * add model patcher
    
    * add processor
    
    * move patching specs to onnx config
    
    * rename
    
    * add comment
    
    * fix
    
    * style
    
    * rename
    
    * style
    
    * remove todo comments
    
    ---------
    
    Co-authored-by: IlyasMoutawwakil <moutawwakil.ilyas.tsi@gmail.com>
    echarlaix and IlyasMoutawwakil authored Jun 18, 2025
    Configuration menu
    Copy the full SHA
    be76dde View commit details
    Browse the repository at this point in the history

Commits on Jun 19, 2025

  1. Remove flatten_inputs from taskmanager (#2296)

    * remove unused method
    
    * remove unused
    echarlaix authored Jun 19, 2025
    Configuration menu
    Copy the full SHA
    824e368 View commit details
    Browse the repository at this point in the history
  2. Fix 'Block pattern could not be match. Pass block_name_to_quantize

    …argument in `quantize_model`' while loading Qwen VL GPTQ model (#2295)
    
    Updated BLOCK_PATTERNS
    arunmadhusud authored Jun 19, 2025
    Configuration menu
    Copy the full SHA
    ed70d27 View commit details
    Browse the repository at this point in the history

Commits on Jun 20, 2025

  1. Remove timm from exporters extra (#2299)

    remove timm from required dependencies
    echarlaix authored Jun 20, 2025
    Configuration menu
    Copy the full SHA
    cf6d4cc View commit details
    Browse the repository at this point in the history

Commits on Jun 23, 2025

  1. Configuration menu
    Copy the full SHA
    ffd91d8 View commit details
    Browse the repository at this point in the history

Commits on Jun 24, 2025

  1. Adding DummyInputGenerator for VisualBert (#2303)

    Co-authored-by: Abdennacer Badaoui <abdennacerbadaoui@MacBook-Air-de-Abdennacer.local>
    Abdennacer-Badaoui and Abdennacer Badaoui authored Jun 24, 2025
    Configuration menu
    Copy the full SHA
    a4152c0 View commit details
    Browse the repository at this point in the history

Commits on Jun 25, 2025

  1. Configuration menu
    Copy the full SHA
    eb187fa View commit details
    Browse the repository at this point in the history
  2. Revert "Adding DummyInputGenerator for VisualBert" (#2306)

    Revert "Adding DummyInputGenerator for VisualBert (#2303)"
    
    This reverts commit a4152c0.
    echarlaix authored Jun 25, 2025
    Configuration menu
    Copy the full SHA
    f7d6e88 View commit details
    Browse the repository at this point in the history

Commits on Jun 27, 2025

  1. add dummy input generator for VisualBert (#2307)

    * add dummy input generator
    
    * fix style
    
    * add default shapes
    
    * Apply suggestions from code review
    
    ---------
    
    Co-authored-by: Ilyas Moutawwakil <57442720+IlyasMoutawwakil@users.noreply.github.com>
    Abdennacer-Badaoui and IlyasMoutawwakil authored Jun 27, 2025
    Configuration menu
    Copy the full SHA
    fb6016d View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    70aaf48 View commit details
    Browse the repository at this point in the history
  3. Fix tflite cli (#2310)

    fix tflite
    IlyasMoutawwakil authored Jun 27, 2025
    Configuration menu
    Copy the full SHA
    ec756fd View commit details
    Browse the repository at this point in the history

Commits on Jul 10, 2025

  1. Deprecate tflite, ort-training and bettertransformer (#2316)

    * deprecate tflite
    
    * bt and ort training
    IlyasMoutawwakil authored Jul 10, 2025
    Configuration menu
    Copy the full SHA
    9564248 View commit details
    Browse the repository at this point in the history

Commits on Jul 23, 2025

  1. Transformers 4.53 support, SmolLM3 and Fix old Transformers support (#…

    …2319)
    
    * test
    
    * fix seq2seq patched sdpa
    
    * patch qwen3_moe, out_attentions, and eager_mask
    
    * fix
    
    * use optimum model
    
    * editable subpackages
    
    * Apply suggestions from code review
    
    * smollm3 support
    
    * deprecate tensorflow onnx export and add smollm3 to export tests
    
    * write a more general sdpa_mask without vmap that's also vectorized and broadcastable
    
    * better and more generic sdpa_mask_without_vmap implementation
    
    * style and fix
    
    * fix
    
    * patch find_packed_sequence_indices as it's untraceable
    
    * fix
    
    * fix
    
    * revert tests removal until refactor
    
    * fix temporary hub repo import
    
    * fix
    
    * fix external data tests on windows
    
    * update phi and phi3 min version
    
    * condition modernbert optimization test
    
    * get back old (pre 4.44) bloom modeling support and remove the need for normalized config in ort modeling
    
    * fix test was using hardcoded architecture
    
    * unparallelize test that uses remote code
    
    * support older versions of mpt and phi (4.36)
    
    * remove parallelism from slow tests
    
    * fix vision to text pipelines test
    
    * more specific version handling for find_packed_sequence_indices
    
    * fix
    IlyasMoutawwakil authored Jul 23, 2025
    Configuration menu
    Copy the full SHA
    53f39a6 View commit details
    Browse the repository at this point in the history

Commits on Jul 29, 2025

  1. Propagate library_name parameter in from_pretrained to export (#2328)

    * Propagate library_name parameter in from_pretrained to export
    
    Required to avoid automatic inferring of the library_name
    
    * Use class attribute for ORTModel instead
    
    Under modeling_diffusion it looks like ORTModel isn't used
    
    * Add test case
    
    * Update optimum/onnxruntime/modeling_ort.py
    
    ---------
    
    Co-authored-by: Ella Charlaix <80481427+echarlaix@users.noreply.github.com>
    tomaarsen and echarlaix authored Jul 29, 2025
    Configuration menu
    Copy the full SHA
    689c0b5 View commit details
    Browse the repository at this point in the history

Commits on Jul 30, 2025

  1. Fix batched inference/generation, position_ids creation, falcon alibi…

    …, gpt_bigcode multi-query,.. (#2326)
    
    * test left-padded batched inference
    
    * demonstrate batched tex generation failure
    
    * fix remote code
    
    * fix
    
    * fix position_ids generation inside ORTModelForCausalLM class
    
    * it works until transformers 4.52 -_-
    
    * now run with latest transformers
    
    * bolean 4D mask is actually not supported by torch onnx exporter
    
    * only test generation with batched inputs, for logits are a bit off because of transformers using boolean mask
    
    * boolean mask safe softmax batched inference
    
    * style
    
    * use old typing
    
    * don't do unnecessary patching
    
    * try to avoid spamming the hub for an image
    
    * update min transformers version
    
    * better and direct torch patching
    
    * more batched generation special cases
    
    * style
    
    * initialize the il image instead of downloading it
    
    * use random pil image
    
    * test different versions of transformers in fast tests
    
    * fix
    
    * revert diffusers changes for now
    
    * mask padding kv cache as well
    
    * fix masking for old bloom
    
    * use constant image to image loading errors
    
    * style
    
    * test diffusers in series to avoid runner dying
    
    * fix
    
    * cleanup and some comments
    
    * fix and test falcon alibi
    
    * style
    
    * fix, support and test multi_query=False as well
    
    * only apply masked testing for transformers version previous to 4.39
    
    * Update optimum/onnxruntime/modeling_decoder.py
    
    * use text decoder position ids onnx config but test its sync with list
    
    * fix opt
    
    * style
    
    * fix sdpa without overriting torch onnx exporter
    
    * use inplace op ;-;
    
    * fix st test
    
    * patch directly in onnx because patch needs to happen after softmax
    IlyasMoutawwakil authored Jul 30, 2025
    Configuration menu
    Copy the full SHA
    31d4ea9 View commit details
    Browse the repository at this point in the history
  2. Release: v1.27.0

    IlyasMoutawwakil committed Jul 30, 2025
    Configuration menu
    Copy the full SHA
    5118493 View commit details
    Browse the repository at this point in the history
Loading