Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loading GGUF support #2

Draft
wants to merge 253 commits into
base: main
Choose a base branch
from
Draft

Loading GGUF support #2

wants to merge 253 commits into from
This pull request is big! We’re only showing the most recent 250 commits.

Commits on Apr 19, 2024

  1. Fix config + attn_implementation in AutoModelForCausalLM.from_pretrai…

    …ned (huggingface#30299)
    
    * Update modeling_utils.py
    
    * Update test_modeling_utils.py
    
    * Update test_modeling_utils.py
    
    * Update test_modeling_utils.py
    hiyouga authored Apr 19, 2024
    Configuration menu
    Copy the full SHA
    21c912e View commit details
    Browse the repository at this point in the history
  2. Add TF swiftformer (huggingface#23342)

    * Duplicate swiftformer
    
    * Convert SwiftFormerPatchEmbedding
    
    * Convert SwiftFormerEmbeddings
    
    * Convert TFSwiftFormerMlp
    
    * Convert TFSwiftFormerConvEncoder
    
    * Convert TFSwiftFormerLocalRepresentation
    
    * convert TFSwiftFormerEncoderBlock
    
    * Convert SwiftFormerStage
    
    * Convert SwiftFormerEncoder
    
    * Add TFSWiftFormerPreTrainedModel
    
    * Convert SwiftFormerForImageClassification
    
    * Add kwargs and start drop path
    
    * Fix syntax
    
    * Change Model class name
    
    * Add TFSwiftFormer to __init__
    
    * Duplicate test_modeling_swiftformer
    
    * First test conversions
    
    * Change require_torch to require_tf
    
    * Add exports to swiftformer __init__
    
    * Add TFSwiftFormerModel wrapper
    
    * Fix __init__ and run black
    
    * Remove docstring from MainLayer, fix padding
    
    * Use keras.layers.Activation on keras.Sequential
    
    * Fix swiftformer exports
    
    * Fix activation layer from config
    
    * Remove post_inits
    
    * Use tf.keras.layers.ZeroPadding2D
    
    * Convert torch normalize
    
    * Change tf test input shape
    
    * Fix softmax and reduce_sum
    
    * Convert expand_dims and repeat
    
    * Add missing reshape and tranpose
    
    * Simplify TFSwiftFormerEncoderBlock.call
    
    * Fix mismatch in patch embeddings
    
    * Fix expected output shape to match channels last
    
    * Fix swiftformer typo
    
    * Disable test_onnx
    
    * Fix TFSwiftFormerForImageClassification call
    
    * Add unpack inputs
    
    * Convert flatten(2).mean(-1)
    
    * Change vision dummy inputs (to be reviewed)
    
    * Change test_forward_signature to use .call
    
    * Fix @unpack_inputs
    
    * Set return_tensors="tf" and rename class
    
    * Rename wrongly named patch_embeddings layer
    
    * Add serving_output and change dummy_input shape
    
    * Make dimensions BCHW and transpose inside embedding layer
    
    * Change SwiftFormerEncoderBlock
    
    * Fix ruff problems
    
    * Add image size to swiftformer config
    
    * Change tranpose to MainLayer and use -1 for reshape
    
    * Remove serving_outputs and dummy_inputs
    
    * Remove test_initialization test from tf model
    
    * Make Sequential component a separate layer
    
    * Fix layers' names
    
    * Tranpose encoder outputs
    
    * Fix tests and check if hidden states is not None
    
    * Fix TFSwiftFormerForImageClassification
    
    * Run make fixup
    
    * Run make fix-copies
    
    * Update modeling_tf_auto
    
    * Update docs
    
    * Fix modeling auto mapping
    
    * Update modelint_tf_swiftformer docs
    
    * Fill image_size doc and type
    
    * Add reduction=None to loss computation
    
    * Update docs
    
    * make style
    
    * Debug: Delete the tip to see if that changes anything
    
    * Re-add tip
    
    * Remove add_code_sample_docstrings
    
    * Remove unused import
    
    * Get the debug to actually tell us the problem it has with the docs
    
    * Try a substitution to match the PyTorch file?
    
    * Add swiftformer to ignore list
    
    * Add build() methods
    
    * Update copyright year
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Remove FIXME comment
    
    * Remove from_pt
    
    * Update copyright year
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Rename one-letter variables
    
    * Remove FIXMEs related to momentum
    
    * Remove old TODO comment
    
    * Remove outstanding FIXME comments
    
    * Get dropout rate from config
    
    * Add specific dropout config for MLP
    
    * Add convencoder dropout to config
    
    * Pass config to SwiftFormerDropPath layer
    
    * Fix drop_path variable name and add Adapted from comment
    
    * Run ruff
    
    * Removed copied from comment
    
    * Run fix copies
    
    * Change drop_path to identity to match pt
    
    * Cleanup build() methods and move to new keras imports
    
    * Update docs/source/en/model_doc/swiftformer.md
    
    Co-authored-by: Matt <[email protected]>
    
    * Raise error if drop_path_rate > 0.0
    
    * Apply suggestions from code review
    
    Replace (self.dim), with self.dim,
    
    Co-authored-by: Matt <[email protected]>
    
    * Remove drop_path function
    
    * Add training to TFSwiftFormerEncoder
    
    * Set self.built = True last
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Should have been added to previous commit
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Apply suggestions from code review
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Change default_feature_extractor to default_image_processor
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Import Keras from modeling_tf_utils
    
    * Remove relative import
    
    * Run ruff --fix
    
    * Move import keras to tf_available
    
    * Add copied from comment to test_forward_signature
    
    * Reduce batch size and num_labels
    
    * Extract loss logic to hf_compute_loss
    
    * Run ruff format
    
    ---------
    
    Co-authored-by: Matt <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    Co-authored-by: Matt <[email protected]>
    4 people authored Apr 19, 2024
    Configuration menu
    Copy the full SHA
    d2cec09 View commit details
    Browse the repository at this point in the history
  3. [Grounding DINO] Add resources (huggingface#30232)

    * Add resources
    
    * Address comments
    
    * Apply suggestions from code review
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    NielsRogge and amyeroberts authored Apr 19, 2024
    Configuration menu
    Copy the full SHA
    8c12690 View commit details
    Browse the repository at this point in the history

Commits on Apr 22, 2024

  1. Configuration menu
    Copy the full SHA
    81e4324 View commit details
    Browse the repository at this point in the history
  2. fix tests

    younesbelkada committed Apr 22, 2024
    Configuration menu
    Copy the full SHA
    8a0d5b8 View commit details
    Browse the repository at this point in the history
  3. Nits for model docs (huggingface#29795)

    * Update llava_next.md
    
    * Update seggpt.md
    merveenoyan authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    b20b017 View commit details
    Browse the repository at this point in the history
  4. Enable multi-device for more models (huggingface#30379)

    * feat: support for vitmatte
    
    * feat: support for vivit
    
    * feat: support for beit
    
    * feat: support for blip :D
    
    * feat: support for data2vec
    jla524 authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    8b02bb6 View commit details
    Browse the repository at this point in the history
  5. GenerationConfig: warn if pad token is negative (huggingface#30187)

    * warn if pad token is negative
    
    * Update src/transformers/generation/configuration_utils.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/generation/configuration_utils.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/configuration_utils.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    Co-authored-by: Joao Gante <[email protected]>
    3 people authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    9138935 View commit details
    Browse the repository at this point in the history
  6. Add FSDP config for CPU RAM efficient loading through accelerate (hug…

    …gingface#30002)
    
    * Add FSDP config for CPU RAM efficient loading
    
    * Style fix
    
    * Update src/transformers/training_args.py
    
    Co-authored-by: Zach Mueller <[email protected]>
    
    * Update src/transformers/training_args.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Add sync_module_states and cpu_ram_efficient_loading validation logic
    
    * Update src/transformers/training_args.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Style
    
    ---------
    
    Co-authored-by: Zach Mueller <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    3 people authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    f16caf4 View commit details
    Browse the repository at this point in the history
  7. Update doc

    LysandreJik committed Apr 22, 2024
    Configuration menu
    Copy the full SHA
    08534f3 View commit details
    Browse the repository at this point in the history
  8. Style

    LysandreJik committed Apr 22, 2024
    Configuration menu
    Copy the full SHA
    ebd9944 View commit details
    Browse the repository at this point in the history
  9. Llama family, fix use_cache=False generation (huggingface#30380)

    * nit to make sure cache positions are not sliced
    
    * fix other models
    
    * nit
    
    * style
    ArthurZucker authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    2d92db8 View commit details
    Browse the repository at this point in the history
  10. Update docstrings for text generation pipeline (huggingface#30343)

    * Update docstrings for text generation pipeline
    
    * Fix docstring arg
    
    * Update docstring to explain chat mode
    
    * Fix doctests
    
    * Fix doctests
    Rocketknight1 authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    0e9d44d View commit details
    Browse the repository at this point in the history
  11. Terminator strings for generate() (huggingface#28932)

    * stash commit (will discard all of this)
    
    * stash commit
    
    * First commit - needs a lot of testing!
    
    * Add a test
    
    * Fix imports and make the tests actually test something
    
    * Tests pass!
    
    * Rearrange test
    
    * Add comments (but it's still a bit confusing)
    
    * Stop storing the tokenizer
    
    * Comment fixup
    
    * Fix for input_ids with a single sequence
    
    * Update tests to test single sequences
    
    * make fixup
    
    * Fix incorrect use of isin()
    
    * Expand tests to catch more cases
    
    * Expand tests to catch more cases
    
    * make fixup
    
    * Fix length calculation and update tests
    
    * Handle Ġ as a space replacement too
    
    * Update src/transformers/generation/stopping_criteria.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Add optimizations from Joao's suggestion
    
    * Remove TODO
    
    * Update src/transformers/generation/stopping_criteria.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update tests/generation/test_stopping_criteria.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * make fixup
    
    * Rename some variables and remove some debugging clauses for clarity
    
    * Add tests for the sub-methods
    
    * Clarify one test slightly
    
    * Add stop_strings to GenerationConfig
    
    * generate() supports stop_string arg, asks for tokenizer if not provided
    
    * make fixup
    
    * Cleanup code and rename variables for clarity
    
    * Update tokenizer error
    
    * Update tokenizer passing, handle generation on GPU
    
    * Slightly more explanation cleanup
    
    * More comment cleanup
    
    * Factor out the token cleanup so it's more obvious what we're doing, and we can change it later
    
    * Careful with that cleanup!
    
    * Cleanup + optimizations to _get_matching_positions
    
    * More minor performance tweaks
    
    * Implement caching and eliminate some expensive ops (startup time: 200ms -> 9ms)
    
    * Remove the pin_memory call
    
    * Parallelize across all stop strings!
    
    * Quick fix for tensor devices
    
    * Update embeddings test for the new format
    
    * Fix test imports
    
    * Manual patching for BERT-like tokenizers
    
    * Return a bool vector instead of a single True/False
    
    * Better comment
    
    * Better comment
    
    * Add tests from @zucchini-nlp
    
    * Amy's list creation nit
    
    * tok_list -> token_list
    
    * Push a big expanded docstring (should we put it somewhere else?)
    
    * Expand docstrings
    
    * Docstring fixups
    
    * Rebase
    
    * make fixup
    
    * Make a properly general method for figuring out token strings
    
    * Fix naming throughout the functions
    
    * Move cache, refactor, fix tests
    
    * Add comment
    
    * Remove finished TODO
    
    * Remove finished TODO
    
    * make fixup
    
    * Update src/transformers/generation/stopping_criteria.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update and shorten docstring
    
    * Update tests to be shorter/clearer and test specific cases
    
    ---------
    
    Co-authored-by: Joao Gante <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    3 people authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    0d84901 View commit details
    Browse the repository at this point in the history
  12. Docs

    LysandreJik committed Apr 22, 2024
    Configuration menu
    Copy the full SHA
    5c913ec View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    f3b3533 View commit details
    Browse the repository at this point in the history
  14. Jamba: fix left-padding test (huggingface#30389)

    fix test
    gante authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    6c7335e View commit details
    Browse the repository at this point in the history
  15. Fix DETA save_pretrained (huggingface#30326)

    * Add class_embed to tied weights for DETA
    
    * Fix test_tied_weights_keys for DETA model
    
    * Replace error raise with assert statement
    qubvel authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    13b3b90 View commit details
    Browse the repository at this point in the history
  16. FIX / PEFT: Pass device correctly to peft (huggingface#30397)

    pass device correctly to peft
    younesbelkada authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    367a0db View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    8b81bfb View commit details
    Browse the repository at this point in the history
  18. fix CI

    younesbelkada committed Apr 22, 2024
    Configuration menu
    Copy the full SHA
    c49f1a8 View commit details
    Browse the repository at this point in the history
  19. Add sdpa and fa2 the Wav2vec2 family. (huggingface#30121)

    * add sdpa to wav2vec.
    Co-authored-by: kamilakesbi <[email protected]>
    Co-authored-by: jp1924 <[email protected]>
    
    * add fa2 to wav2vec2
    
    * add tests
    
    * fix attention_mask compatibility with fa2
    
    * minor dtype fix
    
    * replace fa2 slow test
    
    * fix fa2 slow test
    
    * apply code review + add fa2 batch test
    
    * add sdpa and fa2 to hubert
    
    * sdpa and fa2 to data2vec_audio
    
    * sdpa and fa2 to Sew
    
    * sdpa to unispeech + unispeech sat
    
    * small fix
    
    * attention mask in tests
    
    Co-authored-by: Sanchit Gandhi <[email protected]>
    
    * add_speedup_benchmark_to_doc
    
    ---------
    
    Co-authored-by: [email protected] <[email protected]>
    Co-authored-by: Sanchit Gandhi <[email protected]>
    3 people authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    569743f View commit details
    Browse the repository at this point in the history
  20. Configuration menu
    Copy the full SHA
    7fa538b View commit details
    Browse the repository at this point in the history
  21. Configuration menu
    Copy the full SHA
    5485327 View commit details
    Browse the repository at this point in the history
  22. [FEAT]: EETQ quantizer support (huggingface#30262)

    * [FEAT]: EETQ quantizer support
    
    * Update quantization.md
    
    * Update docs/source/en/main_classes/quantization.md
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update docs/source/en/quantization.md
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update docs/source/en/quantization.md
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/integrations/__init__.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/integrations/__init__.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/integrations/eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/integrations/eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/integrations/eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update tests/quantization/eetq_integration/test_eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/quantizers/auto.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/quantizers/auto.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/quantizers/auto.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/quantizers/quantizer_eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update tests/quantization/eetq_integration/test_eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update src/transformers/quantizers/quantizer_eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update tests/quantization/eetq_integration/test_eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * Update tests/quantization/eetq_integration/test_eetq.py
    
    Co-authored-by: Marc Sun <[email protected]>
    
    * [FEAT]: EETQ quantizer support
    
    * [FEAT]: EETQ quantizer support
    
    * remove whitespaces
    
    * update quantization.md
    
    * style
    
    * Update docs/source/en/quantization.md
    
    Co-authored-by: Younes Belkada <[email protected]>
    
    * add copyright
    
    * Update quantization.md
    
    * Update docs/source/en/quantization.md
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update docs/source/en/quantization.md
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Address the comments by amyeroberts
    
    * style
    
    ---------
    
    Co-authored-by: Marc Sun <[email protected]>
    Co-authored-by: Marc Sun <[email protected]>
    Co-authored-by: Younes Belkada <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    5 people authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    b4c18a8 View commit details
    Browse the repository at this point in the history
  23. [docs] LLM inference (huggingface#29791)

    * first draft
    
    * feedback
    
    * static cache snippet
    
    * feedback
    
    * feedback
    stevhliu authored Apr 22, 2024
    Configuration menu
    Copy the full SHA
    e74d793 View commit details
    Browse the repository at this point in the history

Commits on Apr 23, 2024

  1. Configuration menu
    Copy the full SHA
    6c1295a View commit details
    Browse the repository at this point in the history
  2. Add inputs embeds in generation (huggingface#30269)

    * Add inputs embeds in generation
    
    * always scale embeds
    
    * fix-copies
    
    * fix failing test
    
    * fix copies once more
    
    * remove embeds for models with scaling
    
    * second try to revert
    
    * codestyle
    zucchini-nlp authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    408453b View commit details
    Browse the repository at this point in the history
  3. [Grounding DINO] Add support for cross-attention in GroundingDinoMult…

    …iHeadAttention (huggingface#30364)
    
    * Added cross attention support
    
    * Fixed dtypes
    
    * Fixed assumption
    
    * Moved to decoder
    EduardoPach authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    c651ea9 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    179ab09 View commit details
    Browse the repository at this point in the history
  5. fix: link to HF repo/tree/revision when a file is missing (huggingfac…

    …e#30406)
    
    fix: link to HF repo tree when a file is missing
    mapmeld authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    04ac324 View commit details
    Browse the repository at this point in the history
  6. [tests] add require_torch_sdpa for test that needs sdpa support (hu…

    …ggingface#30408)
    
    * add cuda flag
    
    * check for sdpa
    
    * add bitsandbytes
    faaany authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    2d61823 View commit details
    Browse the repository at this point in the history
  7. Jax: scipy version pin (huggingface#30402)

    scipy pin for jax
    gante authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    31921d8 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    074f05e View commit details
    Browse the repository at this point in the history
  9. Compute merges

    LysandreJik committed Apr 23, 2024
    Configuration menu
    Copy the full SHA
    ca8363e View commit details
    Browse the repository at this point in the history
  10. Fix on "cache position" for assisted generation (huggingface#30068)

    * clean commit history I hope
    
    * get kv seq length correctly
    
    * PR suggestions
    
    * Update src/transformers/testing_utils.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * add comment
    
    * give gpt bigcode it's own overriden method
    
    * remove code
    
    ---------
    
    Co-authored-by: Joao Gante <[email protected]>
    zucchini-nlp and gante authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    77b59dc View commit details
    Browse the repository at this point in the history
  11. fix for itemsize => element_size() for torch backwards compat (huggin…

    …gface#30133)
    
    * fix for itemsize => element_size() for torch backwards compat
    
    * improve handling of element counting
    
    * Update src/transformers/modeling_utils.py
    
    * fixup
    
    * Update src/transformers/modeling_utils.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: Younes Belkada <[email protected]>
    Co-authored-by: Younes Belkada <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    4 people authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    57fc00f View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    4b63d01 View commit details
    Browse the repository at this point in the history
  13. FIX: re-add bnb on docker image (huggingface#30427)

    Update Dockerfile
    younesbelkada authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    d179b9d View commit details
    Browse the repository at this point in the history
  14. Fix LayoutLMv2 init issue and doctest (huggingface#30278)

    * fix
    
    * try suggestion
    
    * update
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    416fdba View commit details
    Browse the repository at this point in the history
  15. Remove old TF port docs (huggingface#30426)

    * Remove old TF port guide
    
    * repo-consistency
    
    * Remove some translations as well for consistency
    
    * Remove some translations as well for consistency
    Rocketknight1 authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    696eded View commit details
    Browse the repository at this point in the history
  16. Rename torch.run to torchrun (huggingface#30405)

    torch.run does not exist anywhere as far as I can tell.
    steven-safeai authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    b8b1e44 View commit details
    Browse the repository at this point in the history
  17. Fix use_cache for xla fsdp (huggingface#30353)

    * Fix use_cache for xla fsdp
    
    * Fix linters
    alanwaketan authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    12c39e5 View commit details
    Browse the repository at this point in the history
  18. [LlamaTokenizerFast] Refactor default llama (huggingface#28881)

    * push legacy to fast as well
    
    * super strange
    
    * Update src/transformers/convert_slow_tokenizer.py
    
    * make sure we are BC
    
    * fix Llama test
    
    * nit
    
    * revert
    
    * more test
    
    * style
    
    * update
    
    * small update w.r.t tokenizers
    
    * nit
    
    * don't split
    
    * lol
    
    * add a test for `add_prefix_space=False`
    
    * fix gemma tokenizer as well
    
    * update
    
    * fix gemma
    
    * nicer failures
    
    * fixup
    
    * update
    
    * fix the example for legacy = False
    
    * use `huggyllama/llama-7b` for the PR doctest
    
    * nit
    
    * use from_slow
    
    * fix llama
    ArthurZucker authored Apr 23, 2024
    Configuration menu
    Copy the full SHA
    e34da3e View commit details
    Browse the repository at this point in the history

Commits on Apr 24, 2024

  1. Configuration menu
    Copy the full SHA
    0eb8fbc View commit details
    Browse the repository at this point in the history
  2. Remove add-new-model in favor of add-new-model-like (huggingface#30424)

    * Remove add-new-model in favor of add-new-model-like
    
    * nits
    LysandreJik authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    d4e92f1 View commit details
    Browse the repository at this point in the history
  3. Remove mentions of models in the READMEs and link to the documentatio…

    …n page in which they are featured. (huggingface#30420)
    
    * REAMDEs
    
    * REAMDEs v2
    LysandreJik authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    c6bba94 View commit details
    Browse the repository at this point in the history
  4. New model PR needs green (slow tests) CI (huggingface#30341)

    * You should not pass
    
    Co-authored-by: Arthur <[email protected]>
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    Co-authored-by: Arthur <[email protected]>
    3 people authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    fc34f84 View commit details
    Browse the repository at this point in the history
  5. Add llama3 (huggingface#30334)

    * nuke
    
    * add co-author
    
    * add co-author
    
    * update card
    
    * fixup and fix copies to please our ci
    
    * nit fixup
    
    * super small nits
    
    * remove tokenizer_path from call to `write_model`
    
    * always safe serialize by default
    
    ---------
    
    Co-authored-by: pcuenca <[email protected]>
    Co-authored-by: xenova <[email protected]>
    3 people authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    89c510d View commit details
    Browse the repository at this point in the history
  6. Fix YOLOS image processor resizing (huggingface#30436)

    * Add test for square image that fails
    
    * Fix for square images
    
    * Extend test cases
    
    * Fix resizing in tests
    
    * Style fixup
    qubvel authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    767e351 View commit details
    Browse the repository at this point in the history
  7. [Llava] + CIs fix red cis and llava integration tests (huggingface#…

    …30440)
    
    * nit
    
    * nit and fmt skip
    
    * fixup
    
    * Update src/transformers/convert_slow_tokenizer.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * set to true
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    ArthurZucker and amyeroberts authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    9a4a119 View commit details
    Browse the repository at this point in the history
  8. [tests] make test device-agnostic (huggingface#30444)

    * make device-agnostic
    
    * clean code
    faaany authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    16c8e17 View commit details
    Browse the repository at this point in the history
  9. fix uncaught init of linear layer in clip's/siglip's for image classi…

    …fication models (huggingface#30435)
    
    * fix clip's/siglip's _init_weights to reflect linear layers in "for image classification"
    
    * trigger slow tests
    vasqu authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    5d64ae9 View commit details
    Browse the repository at this point in the history
  10. fix jamba slow foward for multi-gpu (huggingface#30418)

    * fix jamba slow foward for multi-gpu
    
    * remove comm
    
    * oups
    
    * style
    SunMarc authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    37fa1f6 View commit details
    Browse the repository at this point in the history
  11. [SegGPT] Fix loss calculation (huggingface#30421)

    * Fixed main train issues
    
    * Added loss test
    
    * Update src/transformers/models/seggpt/modeling_seggpt.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Added missing labels arg in SegGptModel forward
    
    * Fixed typo
    
    * Added slow test to test loss calculation
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    EduardoPach and amyeroberts authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    d26c141 View commit details
    Browse the repository at this point in the history
  12. Add paths filter to avoid the chance of being triggered (huggingfac…

    …e#30453)
    
    * trigger
    
    * remove the last job
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    42fed15 View commit details
    Browse the repository at this point in the history
  13. Phi-3 (huggingface#30423)

    * chore(root): Initial commit of Phi-3 files.
    
    * fix(root): Fixes Phi-3 missing on readme.
    
    * fix(root): Ensures files are consistent.
    
    * fix(phi3): Fixes unit tests.
    
    * fix(tests): Fixes style of phi-3 test file.
    
    * chore(tests): Adds integration tests for Phi-3.
    
    * fix(phi3): Removes additional flash-attention usage, .e.g, swiglu and rmsnorm.
    
    * fix(phi3): Fixes incorrect docstrings.
    
    * fix(phi3): Fixes docstring typos.
    
    * fix(phi3): Adds support for Su and Yarn embeddings.
    
    * fix(phi3): Improves according first batch of reviews.
    
    * fix(phi3): Uses up_states instead of y in Phi3MLP.
    
    * fix(phi3): Uses gemma rotary embedding to support torch.compile.
    
    * fix(phi3): Improves how rotary embedding classes are defined.
    
    * fix(phi3): Fixes inv_freq not being re-computed for extended RoPE.
    
    * fix(phi3): Adds last suggestions to modeling file.
    
    * fix(phi3): Splits inv_freq calculation in two lines.
    gugarosa authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    c9693db View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    d0d430f View commit details
    Browse the repository at this point in the history
  15. [research_project] Most of the security issues come from this requi…

    …rement.txt (huggingface#29977)
    
    update most of decision transformers research project
    ArthurZucker authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    661190b View commit details
    Browse the repository at this point in the history
  16. Neuron: When save_safetensor=False, no need to move model to CPU (hug…

    …gingface#29703)
    
    save_safetensor=True is default as of release 4.35.0, which then
    required TPU hotfix huggingface#27799
    (issue huggingface#27578).
    However, when the flag save_safetensor is set to False (compatibility mode),
    moving the model to CPU causes generation of too many graphs
    during checkpoint huggingface#28438.
    This PR disable moving of model to CPU when save_safetensor=False.
    jeffhataws authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    d1d94d7 View commit details
    Browse the repository at this point in the history
  17. Enable fp16 on CPU (huggingface#30459)

    * Check removing flag for torch
    
    * LLM oops
    
    * Getting there...
    
    * More discoveries
    
    * Change
    
    * Clean up and prettify
    
    * Logic check
    
    * Not
    muellerzr authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    5c57463 View commit details
    Browse the repository at this point in the history
  18. Non blocking support to torch DL's (huggingface#30465)

    * Non blocking support
    
    * Check for optimization
    
    * Doc
    muellerzr authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    6ad9c8f View commit details
    Browse the repository at this point in the history
  19. consistent job / pytest report / artifact name correspondence (huggin…

    …gface#30392)
    
    * better names
    
    * run better names
    
    * update
    
    * update
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Apr 24, 2024
    Configuration menu
    Copy the full SHA
    fbb41cd View commit details
    Browse the repository at this point in the history

Commits on Apr 25, 2024

  1. Configuration menu
    Copy the full SHA
    2a0c9b0 View commit details
    Browse the repository at this point in the history
  2. Workflow / ENH: Add SSH into our runners workflow (huggingface#30425)

    * add SSH into our runners workflow
    
    * fix
    
    * fix
    
    * fix
    
    * use our previous approaches
    
    * forward contrib credits from discussions
    
    ---------
    
    Co-authored-by: Yih-Dar <[email protected]>
    younesbelkada and ydshieh authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    cebb072 View commit details
    Browse the repository at this point in the history
  3. FIX / Workflow: Change tailscale trigger condition (huggingface#30471)

    Update push-important-models.yml
    younesbelkada authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    cd0cd12 View commit details
    Browse the repository at this point in the history
  4. FIX / Workflow: Fix SSH workflow bug (huggingface#30474)

    Update ssh-runner.yml
    younesbelkada authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    e9b1635 View commit details
    Browse the repository at this point in the history
  5. [fix codellama conversion] (huggingface#30472)

    * fix codellama conversion
    
    * nit
    ArthurZucker authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    c60749d View commit details
    Browse the repository at this point in the history
  6. Script for finding candidate models for deprecation (huggingface#29686)

    * Add utility for finding candidate models for deprecation
    
    * Better model filtering
    
    * Update
    
    * Add warning tip
    
    * Fix up
    
    * Review comments
    
    * Filter requests based on tags
    
    * Add copyright header
    amyeroberts authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    30ee508 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    fac7bb3 View commit details
    Browse the repository at this point in the history
  8. Fix SigLip classification doctest (huggingface#30475)

    * Fix SigLip classification doctest
    
    * Remove extra line
    
    * Update src/transformers/models/siglip/modeling_siglip.py
    amyeroberts authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    4fed29e View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    aca4a10 View commit details
    Browse the repository at this point in the history
  10. Prevent crash with WandbCallback with third parties (huggingface#30477

    )
    
    * Use EAFP principle to prevent crash with third parties
    
    * Remove leftover debugging code
    
    * Add info-level logger message
    tomaarsen authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    ce5ae5a View commit details
    Browse the repository at this point in the history
  11. 🚨 Add training compatibility for Musicgen-like models (huggingface#29802

    )
    
    * first modeling code
    
    * make repository
    
    * still WIP
    
    * update model
    
    * add tests
    
    * add latest change
    
    * clean docstrings and copied from
    
    * update docstrings md and readme
    
    * correct chroma function
    
    * correct copied from and remove unreleated test
    
    * add doc to toctree
    
    * correct imports
    
    * add convert script to notdoctested
    
    * Add suggestion from Sanchit
    
    Co-authored-by: Sanchit Gandhi <[email protected]>
    
    * correct get_uncoditional_inputs docstrings
    
    * modify README according to SANCHIT feedback
    
    * add chroma to audio utils
    
    * clean librosa and torchaudio hard dependencies
    
    * fix FE
    
    * refactor audio decoder -> audio encoder for consistency with previous musicgen
    
    * refactor conditional -> encoder
    
    * modify sampling rate logics
    
    * modify license at the beginning
    
    * refactor all_self_attns->all_attentions
    
    * remove ignore copy from causallm generate
    
    * add copied from for from_sub_models
    
    * fix make copies
    
    * add warning if audio is truncated
    
    * add copied from where relevant
    
    * remove artefact
    
    * fix convert script
    
    * fix torchaudio and FE
    
    * modify chroma method according to feedback-> better naming
    
    * refactor input_values->input_features
    
    * refactor input_values->input_features and fix import fe
    
    * add input_features to docstrigs
    
    * correct inputs_embeds logics
    
    * remove dtype conversion
    
    * refactor _prepare_conditional_hidden_states_kwargs_for_generation ->_prepare_encoder_hidden_states_kwargs_for_generation
    
    * change warning for chroma length
    
    * Update src/transformers/models/musicgen_melody/convert_musicgen_melody_transformers.py
    
    Co-authored-by: Sanchit Gandhi <[email protected]>
    
    * change way to save wav, using soundfile
    
    * correct docs and change to soundfile
    
    * fix import
    
    * fix init proj layers
    
    * add draft training
    
    * fix cross entropy
    
    * clean loss computation
    
    * fix labels
    
    * remove line breaks from md
    
    * fix issue with docstrings
    
    * add FE suggestions
    
    * improve is in logics and remove useless imports
    
    * remove custom from_pretrained
    
    * simplify docstring code
    
    * add suggestions for modeling tests
    
    * make style
    
    * update converting script with sanity check
    
    * remove encoder attention mask from conditional generation
    
    * replace musicgen melody checkpoints with official orga
    
    * rename ylacombe->facebook in checkpoints
    
    * fix copies
    
    * remove unecessary warning
    
    * add shape in code docstrings
    
    * add files to slow doc tests
    
    * fix md bug and add md to not_tested
    
    * make fix-copies
    
    * fix hidden states test and batching
    
    * update training code
    
    * add training tests for melody
    
    * add training for o.g musicgen
    
    * fix copied from
    
    * remove final todos
    
    * make style
    
    * fix style
    
    * add suggestions from review
    
    * add ref to the original loss computation code
    
    * rename method + fix labels in tests
    
    * make style
    
    ---------
    
    Co-authored-by: Sanchit Gandhi <[email protected]>
    ylacombe and sanchit-gandhi authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    90cb55b View commit details
    Browse the repository at this point in the history
  12. Add WSD scheduler (huggingface#30231)

    * Added WSD scheduler.
    
    * Added tests.
    
    * Fixed errors.
    
    * Fix formatting.
    
    * CI fixes.
    visheratin authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    7b1170b View commit details
    Browse the repository at this point in the history
  13. Fix Issue huggingface#29817 Video Classification Task Guide Using Un…

    …declared Variables (huggingface#30457)
    
    * Fix issue huggingface#29817
    
    Video Classification Task Guide Using Undeclared Variables
    
    * Update docs/source/en/tasks/video_classification.md
    
    updated with review comments
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Fix issue huggingface#29817
    
    Add line space following PR comments
    
    ---------
    
    Co-authored-by: manju-rangam <Manju1@Git>
    Co-authored-by: amyeroberts <[email protected]>
    3 people authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    9283313 View commit details
    Browse the repository at this point in the history
  14. Make accelerate install non-torch dependent (huggingface#30463)

    * Pin accelerate w/o eager
    
    * Eager
    
    * Update .circleci/create_circleci_config.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Expound
    
    * Expound squared
    
    * PyTorch -> dependency
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    muellerzr and amyeroberts authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    86f2569 View commit details
    Browse the repository at this point in the history
  15. Introduce Stateful Callbacks (huggingface#29666)

    * Introduce saveable callbacks
    
    * Add note
    
    * Test for non-present and flag
    
    * Support early stopping and refusing to train further
    
    * Update docstring
    
    * More saving
    
    * Import oopsie
    
    * Apply suggestions from code review
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Make it go through TrainerArguments
    
    * Document
    
    * Fix test
    
    * Apply suggestions from code review
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Rework to allow for duplicates
    
    * CLean
    
    * Fix failing tests
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    muellerzr and amyeroberts authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    ad697f1 View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    e60491a View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    bc274a2 View commit details
    Browse the repository at this point in the history
  18. Add sidebar tutorial for chat models (huggingface#30401)

    * Draft tutorial for talking to chat models
    
    * Reformat lists and text snippets
    
    * Cleanups and clarifications
    
    * Finish up remaining TODOs
    
    * Correct section link
    
    * Small fix
    
    * Add proper quantization examples
    
    * Add proper quantization examples
    
    * Add proper quantization examples
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/_toctree.yml
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Update docs/source/en/conversations.md
    
    Co-authored-by: Steven Liu <[email protected]>
    
    * Fix Text Generation Pipeline link and add a ref to the LLM inference guide
    
    * intelligent -> capable
    
    * Small intro cleanup
    
    * Small text cleanup
    
    * Small text cleanup
    
    * Clarification about system message
    
    * Clarification about system message
    
    ---------
    
    Co-authored-by: Steven Liu <[email protected]>
    Rocketknight1 and stevhliu authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    f396271 View commit details
    Browse the repository at this point in the history
  19. Quantization: HfQuantizer quant method update (huggingface#30484)

    ensure popular quant methods are supported
    younesbelkada authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    26ddc58 View commit details
    Browse the repository at this point in the history
  20. [docs] Spanish translation of pipeline_tutorial.md (huggingface#30252)

    * add pipeline_webserver to es/
    
    * add pipeline_webserver to es/, translate first section
    
    * add comment for checking link
    
    * translate pipeline_webserver
    
    * edit pipeline_webserver
    
    * fix typo
    aaronjimv authored Apr 25, 2024
    Configuration menu
    Copy the full SHA
    a98c417 View commit details
    Browse the repository at this point in the history

Commits on Apr 26, 2024

  1. FEAT: PEFT support for EETQ (huggingface#30449)

    Update quantizer_eetq.py
    younesbelkada authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    19cfdf0 View commit details
    Browse the repository at this point in the history
  2. Fix the bitsandbytes error formatting ("Some modules are dispatched…

    … on ...") (huggingface#30494)
    
    Fix the `bitsandbytes` error when some modules are not properly offloaded.
    kyo-takano authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    59e715f View commit details
    Browse the repository at this point in the history
  3. Update dtype_byte_size to handle torch.float8_e4m3fn/float8_e5m2 ty…

    …pes (huggingface#30488)
    
    * Update modeling_utils/dtype_byte_size to handle float8 types
    
    * Add a test for dtype_byte_size
    
    * Format
    
    * Fix bool
    mgoin authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    20081c7 View commit details
    Browse the repository at this point in the history
  4. Use the Keras set_random_seed in tests (huggingface#30504)

    Use the Keras set_random_seed to ensure reproducible weight initialization
    Rocketknight1 authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    2de5cb1 View commit details
    Browse the repository at this point in the history
  5. [BERT] Add support for sdpa (huggingface#28802)

    * Adding SDPA support for BERT
    
    * Using the proper input name for testing model input in inference()
    
    * Adding documentation for SDPA in BERT model page
    
    * Use the stable link for the documentation
    
    * Adding a gate to only call .contiguous() for torch < 2.2.0
    
    * Additions and fixes to the documentation
    
    * Minor updates to documentation
    
    * Adding extra requirements needed for the contiguous() bug
    
    * Adding "Adapted from" in plcae of the "Copied from"
    
    * Add benchmark speedup tables to the documentation
    
    * Minor fixes to the documentation
    
    * Use ClapText as a replacemenet for Bert in the Copied-From
    
    * Some more fixes for the fix-copies references
    
    * Overriding the test_eager_matches_sdpa_generate in bert tests to not load with low_cpu_mem_usage
    
    [test all]
    
    * Undo changes to separate test
    
    * Refactored SDPA self attention code for KV projections
    
    * Change use_sdpa to attn_implementation
    
    * Fix test_sdpa_can_dispatch_on_flash by preparing input (required for MultipleChoice models)
    hackyon authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    dfa7b58 View commit details
    Browse the repository at this point in the history
  6. Remove skipping logic now that set_epoch exists (huggingface#30501)

    * Remove skipping logic now that set_epoch exists
    
    * Working version, clean
    muellerzr authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    77ff304 View commit details
    Browse the repository at this point in the history
  7. [DETR] Remove timm hardcoded logic in modeling files (huggingface#2…

    …9038)
    
    * Enable instantiating model with pretrained backbone weights
    
    * Clarify pretrained import
    
    * Use load_backbone instead
    
    * Add backbone_kwargs to config
    
    * Fix up
    
    * Add tests
    
    * Tidy up
    
    * Enable instantiating model with pretrained backbone weights
    
    * Update tests so backbone checkpoint isn't passed in
    
    * Clarify pretrained import
    
    * Update configs - docs and validation check
    
    * Update src/transformers/utils/backbone_utils.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Clarify exception message
    
    * Update config init in tests
    
    * Add test for when use_timm_backbone=True
    
    * Use load_backbone instead
    
    * Add use_timm_backbone to the model configs
    
    * Add backbone_kwargs to config
    
    * Pass kwargs to constructors
    
    * Draft
    
    * Fix tests
    
    * Add back timm - weight naming
    
    * More tidying up
    
    * Whoops
    
    * Tidy up
    
    * Handle when kwargs are none
    
    * Update tests
    
    * Revert test changes
    
    * Deformable detr test - don't use default
    
    * Don't mutate; correct model attributes
    
    * Add some clarifying comments
    
    * nit - grammar is hard
    
    ---------
    
    Co-authored-by: Arthur <[email protected]>
    amyeroberts and ArthurZucker authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    aafa7ce View commit details
    Browse the repository at this point in the history
  8. [examples] update whisper fine-tuning (huggingface#29938)

    * [examples] update whisper fine-tuning
    
    * deprecate forced/suppress tokens
    
    * item assignment
    
    * update readme
    
    * final fix
    sanchit-gandhi authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    38b53da View commit details
    Browse the repository at this point in the history
  9. Fix GroundingDINO, DPR after BERT SDPA update (huggingface#30506)

    Fix GroundingDINO, DPR after BET SDPA update
    amyeroberts authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    e7d52a1 View commit details
    Browse the repository at this point in the history
  10. load_image - decode b64encode and encodebytes strings (huggingface#30192

    )
    
    * Decode b64encode and encodebytes strings
    
    * Remove conditional encode -- image is always a string
    amyeroberts authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    c793b26 View commit details
    Browse the repository at this point in the history
  11. [SegGPT] Fix seggpt image processor (huggingface#29550)

    * Fixed SegGptImageProcessor to handle 2D and 3D prompt mask inputs
    
    * Added new test to check prompt mask equivalence
    
    * New proposal
    
    * Better proposal
    
    * Removed unnecessary method
    
    * Updated seggpt docs
    
    * Introduced do_convert_rgb
    
    * nits
    EduardoPach authored Apr 26, 2024
    Configuration menu
    Copy the full SHA
    6d4cabd View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    73014b5 View commit details
    Browse the repository at this point in the history

Commits on Apr 29, 2024

  1. Allow boolean FSDP options in fsdp_config (huggingface#30439)

    * Allow boolean FSDP options in fsdp_config
    
    * Use lower() to be safe
    helloworld1 authored Apr 29, 2024
    Configuration menu
    Copy the full SHA
    80126f9 View commit details
    Browse the repository at this point in the history
  2. Pass attn_implementation when using AutoXXX.from_config (huggingface#…

    …30507)
    
    * Pass attn_implementation when using AutoXXX.from_config
    
    * Fix
    amyeroberts authored Apr 29, 2024
    Configuration menu
    Copy the full SHA
    e8acb70 View commit details
    Browse the repository at this point in the history
  3. Fix broken link to Transformers notebooks (huggingface#30512)

    Co-authored-by: Clint Adams <[email protected]>
    clinty and clinty authored Apr 29, 2024
    Configuration menu
    Copy the full SHA
    bdbe166 View commit details
    Browse the repository at this point in the history
  4. Update runner tag for PR slow CI (huggingface#30535)

    fix
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Apr 29, 2024
    Configuration menu
    Copy the full SHA
    c024218 View commit details
    Browse the repository at this point in the history
  5. Fix repo. fetch/checkout in PR slow CI job (huggingface#30537)

    fix
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Apr 29, 2024
    Configuration menu
    Copy the full SHA
    87be06c View commit details
    Browse the repository at this point in the history
  6. Reenable SDPA's FA2 During Training with torch.compile (huggingface#3…

    …0442)
    
    * Reenable SDPA's FA2 during training with torch.compile
    
    * fix Olmo's SDPA FA2 dispatching too
    
    * update formatting
    
    * improved SDPA comment
    
    * formatting and explanatory comment
    
    * is_causal if statement to one-liner
    warner-benjamin authored Apr 29, 2024
    Configuration menu
    Copy the full SHA
    9df8b30 View commit details
    Browse the repository at this point in the history
  7. Include safetensors as part of _load_best_model (huggingface#30553)

    * Include safetensors
    
    * Cleanup
    muellerzr authored Apr 29, 2024
    Configuration menu
    Copy the full SHA
    a3aabc7 View commit details
    Browse the repository at this point in the history

Commits on Apr 30, 2024

  1. Pass use_cache in kwargs for GPTNeoX (huggingface#30538)

    pass use_cache in kwargs
    zucchini-nlp authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    c712d05 View commit details
    Browse the repository at this point in the history
  2. Enable multi-device for more models (huggingface#30409)

    * feat: support for dinov2
    
    * feat: support for depth_anything
    
    * feat: support for efficientformer
    
    * feat: support for bert (is this right?)
    
    * update: embedding split
    
    * remove: empty string
    
    * feat: support for align
    
    * fix: copies
    
    * fix: QAQBertEmbeddings
    
    * fix: more consistency issues
    
    * revert: support for effientformer
    
    * feat: support for altclip
    
    * feat: support for blip_text
    
    * support for ChineseCLIP
    
    * feat: support for depth anything
    
    * feat: support for dpt
    
    * feat: support for dpt
    
    * feat: support for git
    
    * feat: support for groupvit
    
    * update: format
    
    * fix: support for clip
    
    * fix: consistency
    
    * feat: support for pvt
    
    * feat: support for vit_msn
    
    * fix: consistency
    
    * fix: other copies
    
    * remove: device transfer
    
    * revert: in-place add
    
    * update: support for align
    
    * update: support for bert
    
    * update: support for Chinese CLIP
    
    * revert: changes to efficientformer
    
    * update: support for dpt
    
    * update: support for efficientformer
    
    * revert: changes to git
    
    * revert: changes to groupvit
    
    * revert: changes to roc_bert
    
    * update: support for vit_msn
    
    * revert: changes to dpt
    
    * remove: extra space
    
    * style: extra space
    jla524 authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    0ae789e View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    45983db View commit details
    Browse the repository at this point in the history
  4. change logic

    younesbelkada committed Apr 30, 2024
    Configuration menu
    Copy the full SHA
    e6c6f6c View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    a6cd08c View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    6611877 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    455163b View commit details
    Browse the repository at this point in the history
  8. change logic

    younesbelkada committed Apr 30, 2024
    Configuration menu
    Copy the full SHA
    42d5815 View commit details
    Browse the repository at this point in the history
  9. Update src/transformers/modeling_utils.py

    Co-authored-by: amyeroberts <[email protected]>
    younesbelkada and amyeroberts authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    1d3acec View commit details
    Browse the repository at this point in the history
  10. change

    younesbelkada committed Apr 30, 2024
    Configuration menu
    Copy the full SHA
    af3c42c View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    a27db0c View commit details
    Browse the repository at this point in the history
  12. Apply suggestions from code review

    Co-authored-by: amyeroberts <[email protected]>
    younesbelkada and amyeroberts authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    14ad10c View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    ab621a7 View commit details
    Browse the repository at this point in the history
  14. put back comment

    younesbelkada committed Apr 30, 2024
    Configuration menu
    Copy the full SHA
    207820a View commit details
    Browse the repository at this point in the history
  15. Configuration menu
    Copy the full SHA
    1fef8ad View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    9ae7363 View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    75bbfd5 View commit details
    Browse the repository at this point in the history
  18. Configuration menu
    Copy the full SHA
    1bff6a0 View commit details
    Browse the repository at this point in the history
  19. Configuration menu
    Copy the full SHA
    78a57c5 View commit details
    Browse the repository at this point in the history
  20. Fix seq2seq collator padding (huggingface#30556)

    * fix seq2seq data collator to respect the given padding strategy
    
    further added tests for the seq2seq data collator in the style of the `data_collator_for_token_classification` (pt, tf, np)
    
    * formatting and change bool equals "==" to "is"
    
    * add missed return types in tests
    
    * update numpy test as it can handle unequal shapes, not like pt or tf
    vasqu authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    9112520 View commit details
    Browse the repository at this point in the history
  21. BlipModel: get_multimodal_features method (huggingface#30438)

    * add_blip_get_multimodal_feautres
    
    * Fix docstring error
    
    * reimplement get_multimodal_features
    
    * fix error
    
    * recheck code quality
    
    * add new necessary tests
    XavierSpycy authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    0cdb6b3 View commit details
    Browse the repository at this point in the history
  22. Add chat templating support for KeyDataset in text-generation pipeline (

    huggingface#30558)
    
    * added chat templating support for keydataset in generation pipeline
    
    * fixed and improved test
    
    * fix formatting test failures
    
    * Fix tests
    
    * Fix tests
    DarshanDeshpande authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    2ecefc3 View commit details
    Browse the repository at this point in the history
  23. Fix generation doctests (huggingface#30263)

    * fix doctest
    
    * fix torch doctest
    
    * make CI happy
    
    * raise error
    
    * make fixup
    zucchini-nlp authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    b8ac4d0 View commit details
    Browse the repository at this point in the history
  24. General PR slow CI (huggingface#30540)

    * More general PR slow CI
    
    * Update utils/pr_slow_ci_models.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    3 people authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    87927b2 View commit details
    Browse the repository at this point in the history
  25. Remove use_square_size after loading (huggingface#30567)

    * fix
    
    * add test
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Apr 30, 2024
    Configuration menu
    Copy the full SHA
    78fdd64 View commit details
    Browse the repository at this point in the history

Commits on May 1, 2024

  1. Use text config's vocab size in testing models (huggingface#30568)

    use text config's vocab size
    zucchini-nlp authored May 1, 2024
    Configuration menu
    Copy the full SHA
    9d31b32 View commit details
    Browse the repository at this point in the history
  2. Encoder-decoder models: move embedding scale to nn.Module (huggingfac…

    …e#30410)
    
    * move scaling to nn.Module
    
    * let the test be here for now (need to fix)
    
    * failing tests
    
    * last failing models
    
    * Revert commit 4c14817
    
    * clean-up
    
    * oops forgot
    
    * codestyle
    
    * raise NotImplemented when possible
    
    * Update tests/test_modeling_common.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * skip tests in respective modeling files
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    zucchini-nlp and amyeroberts authored May 1, 2024
    Configuration menu
    Copy the full SHA
    38a4bf7 View commit details
    Browse the repository at this point in the history
  3. Fix Marian model conversion (huggingface#30173)

    * fix marian model coversion
    
    * uncomment that line
    
    * remove unnecessary code
    
    * revert tie_weights, doesn't hurt
    zucchini-nlp authored May 1, 2024
    Configuration menu
    Copy the full SHA
    4bc9cb3 View commit details
    Browse the repository at this point in the history
  4. Refactor default chat template warnings (huggingface#30551)

    * Temporarily silence warnings in apply_chat_template until we can properly deprecate default chat templates
    
    * make fixup
    
    * Move the default chat template warning into apply_chat_template itself
    
    * make fixup
    Rocketknight1 authored May 1, 2024
    Configuration menu
    Copy the full SHA
    4b4da18 View commit details
    Browse the repository at this point in the history
  5. Fix QA example (huggingface#30580)

    * Handle cases when CLS token is absent
    
    * Use BOS token as a fallback
    Rocketknight1 authored May 1, 2024
    Configuration menu
    Copy the full SHA
    1e05671 View commit details
    Browse the repository at this point in the history
  6. remove jax example (huggingface#30498)

    remove example
    ArthurZucker authored May 1, 2024
    Configuration menu
    Copy the full SHA
    3c69d81 View commit details
    Browse the repository at this point in the history
  7. Fix canonical model --model_type in examples (huggingface#30480)

    Fix --model_type in examples
    amyeroberts authored May 1, 2024
    Configuration menu
    Copy the full SHA
    bbaa8ce View commit details
    Browse the repository at this point in the history
  8. Gemma: update activation warning (huggingface#29995)

    * Gemma: only display act. warning when necessary
    
    This is a nit PR, but I was confused. I got the warning even after I
    had changed `hidden_act` to `gelu_pytorch_tanh`, telling me that I
    was using the "legacy" `gelu_pytorch_tanh`.
    
    Another option is to keep the warning but change the message to say
    something like "`hidden_act` is ignored, please use `hidden_activation`
    instead. Setting Gemma's activation function to `gelu_pytorch_tanh`".
    
    * Change message, and set `config.hidden_activation`
    pcuenca authored May 1, 2024
    Configuration menu
    Copy the full SHA
    f4f18af View commit details
    Browse the repository at this point in the history
  9. Bump gitpython from 3.1.32 to 3.1.41 in /examples/research_projects/d…

    …ecision_transformer (huggingface#30587)
    
    Bump gitpython in /examples/research_projects/decision_transformer
    
    Bumps [gitpython](https://github.com/gitpython-developers/GitPython) from 3.1.32 to 3.1.41.
    - [Release notes](https://github.com/gitpython-developers/GitPython/releases)
    - [Changelog](https://github.com/gitpython-developers/GitPython/blob/main/CHANGES)
    - [Commits](gitpython-developers/GitPython@3.1.32...3.1.41)
    
    ---
    updated-dependencies:
    - dependency-name: gitpython
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    b71f512 View commit details
    Browse the repository at this point in the history
  10. Bump grpcio from 1.44.0 to 1.53.2 in /examples/research_projects/deci…

    …sion_transformer (huggingface#30585)
    
    Bump grpcio in /examples/research_projects/decision_transformer
    
    Bumps [grpcio](https://github.com/grpc/grpc) from 1.44.0 to 1.53.2.
    - [Release notes](https://github.com/grpc/grpc/releases)
    - [Changelog](https://github.com/grpc/grpc/blob/master/doc/grpc_release_schedule.md)
    - [Commits](grpc/grpc@v1.44.0...v1.53.2)
    
    ---
    updated-dependencies:
    - dependency-name: grpcio
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    ff8f624 View commit details
    Browse the repository at this point in the history
  11. Bump gitpython from 3.1.32 to 3.1.41 in /examples/research_projects/d…

    …istillation (huggingface#30586)
    
    Bump gitpython in /examples/research_projects/distillation
    
    Bumps [gitpython](https://github.com/gitpython-developers/GitPython) from 3.1.32 to 3.1.41.
    - [Release notes](https://github.com/gitpython-developers/GitPython/releases)
    - [Changelog](https://github.com/gitpython-developers/GitPython/blob/main/CHANGES)
    - [Commits](gitpython-developers/GitPython@3.1.32...3.1.41)
    
    ---
    updated-dependencies:
    - dependency-name: gitpython
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    7164171 View commit details
    Browse the repository at this point in the history
  12. Bump pyarrow from 7.0.0 to 15.0.0 in /examples/research_projects/deci…

    …sion_transformer (huggingface#30582)
    
    Bump pyarrow in /examples/research_projects/decision_transformer
    
    Bumps [pyarrow](https://github.com/apache/arrow) from 7.0.0 to 15.0.0.
    - [Commits](apache/arrow@go/v7.0.0...go/v15.0.0)
    
    ---
    updated-dependencies:
    - dependency-name: pyarrow
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    aefbdfe View commit details
    Browse the repository at this point in the history
  13. Bump pyarrow from 1.0.1 to 15.0.0 in /examples/research_projects/visu…

    …al_bert (huggingface#30583)
    
    Bump pyarrow in /examples/research_projects/visual_bert
    
    Bumps [pyarrow](https://github.com/apache/arrow) from 1.0.1 to 15.0.0.
    - [Commits](apache/arrow@apache-arrow-1.0.1...go/v15.0.0)
    
    ---
    updated-dependencies:
    - dependency-name: pyarrow
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    0ec3003 View commit details
    Browse the repository at this point in the history
  14. Bump pyarrow from 1.0.1 to 15.0.0 in /examples/research_projects/lxme…

    …rt (huggingface#30584)
    
    Bumps [pyarrow](https://github.com/apache/arrow) from 1.0.1 to 15.0.0.
    - [Commits](apache/arrow@apache-arrow-1.0.1...go/v15.0.0)
    
    ---
    updated-dependencies:
    - dependency-name: pyarrow
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    b33f01f View commit details
    Browse the repository at this point in the history
  15. Bump torch from 1.6.0 to 1.13.1 in /examples/research_projects/lxmert (

    …huggingface#21174)
    
    Bumps [torch](https://github.com/pytorch/pytorch) from 1.6.0 to 1.13.1.
    - [Release notes](https://github.com/pytorch/pytorch/releases)
    - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md)
    - [Commits](pytorch/pytorch@v1.6.0...v1.13.1)
    
    ---
    updated-dependencies:
    - dependency-name: torch
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    7a29c57 View commit details
    Browse the repository at this point in the history
  16. Bump torch from 1.11.0 to 1.13.1 in /examples/research_projects/codep…

    …arrot (huggingface#21170)
    
    Bump torch in /examples/research_projects/codeparrot
    
    Bumps [torch](https://github.com/pytorch/pytorch) from 1.11.0 to 1.13.1.
    - [Release notes](https://github.com/pytorch/pytorch/releases)
    - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md)
    - [Commits](pytorch/pytorch@v1.11.0...v1.13.1)
    
    ---
    updated-dependencies:
    - dependency-name: torch
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    ce66c0e View commit details
    Browse the repository at this point in the history
  17. Bump torch from 1.6.0 to 1.13.1 in /examples/research_projects/visual…

    …_bert (huggingface#21172)
    
    Bump torch in /examples/research_projects/visual_bert
    
    Bumps [torch](https://github.com/pytorch/pytorch) from 1.6.0 to 1.13.1.
    - [Release notes](https://github.com/pytorch/pytorch/releases)
    - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md)
    - [Commits](pytorch/pytorch@v1.6.0...v1.13.1)
    
    ---
    updated-dependencies:
    - dependency-name: torch
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    6e0cba3 View commit details
    Browse the repository at this point in the history
  18. Fix image segmentation example - don't reopen image (huggingface#30481)

    Fix image segmentation example - don't repoen image
    amyeroberts authored May 1, 2024
    Configuration menu
    Copy the full SHA
    d2feb54 View commit details
    Browse the repository at this point in the history
  19. Improve object detection task guideline (huggingface#29967)

    * Add improvements
    
    * Address comment
    NielsRogge authored May 1, 2024
    Configuration menu
    Copy the full SHA
    dc401d3 View commit details
    Browse the repository at this point in the history
  20. Configuration menu
    Copy the full SHA
    d57ffb4 View commit details
    Browse the repository at this point in the history
  21. Fix llava half precision and autocast issues (huggingface#29721)

    * Ensure input_embeds and image_features are the same dtype in autocast
    
    * Fix nans in half precision llava-next and fix autocasting behavior.
    
    * Fix styling issues.
    
    * fix randn newline instantiation
    
    * fix broken slow llava test
    
    * Fix llava next init.
    
    * fix styling issues
    
    * [run-slow]llava,llava_next
    
    * fix styling issues
    frasermince authored May 1, 2024
    Configuration menu
    Copy the full SHA
    5090ea3 View commit details
    Browse the repository at this point in the history
  22. Bump torch from 1.11.0 to 1.13.1 in /examples/research_projects/decis…

    …ion_transformer (huggingface#21171)
    
    Bump torch in /examples/research_projects/decision_transformer
    
    Bumps [torch](https://github.com/pytorch/pytorch) from 1.11.0 to 1.13.1.
    - [Release notes](https://github.com/pytorch/pytorch/releases)
    - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md)
    - [Commits](pytorch/pytorch@v1.11.0...v1.13.1)
    
    ---
    updated-dependencies:
    - dependency-name: torch
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    6f465d4 View commit details
    Browse the repository at this point in the history
  23. Bump torch from 1.9.0+cpu to 1.13.1 in /examples/research_projects/ja…

    …x-projects/hybrid_clip (huggingface#21167)
    
    Bump torch in /examples/research_projects/jax-projects/hybrid_clip
    
    Bumps [torch](https://github.com/pytorch/pytorch) from 1.9.0+cpu to 1.13.1.
    - [Release notes](https://github.com/pytorch/pytorch/releases)
    - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md)
    - [Commits](https://github.com/pytorch/pytorch/commits/v1.13.1)
    
    ---
    updated-dependencies:
    - dependency-name: torch
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    4f3c7af View commit details
    Browse the repository at this point in the history
  24. Bump pillow from 10.0.1 to 10.2.0 in /examples/research_projects/deci…

    …sion_transformer (huggingface#28655)
    
    Bump pillow in /examples/research_projects/decision_transformer
    
    Bumps [pillow](https://github.com/python-pillow/Pillow) from 10.0.1 to 10.2.0.
    - [Release notes](https://github.com/python-pillow/Pillow/releases)
    - [Changelog](https://github.com/python-pillow/Pillow/blob/main/CHANGES.rst)
    - [Commits](python-pillow/Pillow@10.0.1...10.2.0)
    
    ---
    updated-dependencies:
    - dependency-name: pillow
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 1, 2024
    Configuration menu
    Copy the full SHA
    3a36597 View commit details
    Browse the repository at this point in the history
  25. Bump torch from 1.9.0+cpu to 1.13.1 in /examples/flax/vision (hugging…

    …face#21168)
    
    Bumps [torch](https://github.com/pytorch/pytorch) from 1.9.0+cpu to 1.13.1.
    - [Release notes](https://github.com/pytorch/pytorch/releases)
    - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md)
    - [Commits](https://github.com/pytorch/pytorch/commits/v1.13.1)
    
    ---
    updated-dependencies:
    - dependency-name: torch
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    Co-authored-by: amyeroberts <[email protected]>
    dependabot[bot] and amyeroberts authored May 1, 2024
    Configuration menu
    Copy the full SHA
    c681b58 View commit details
    Browse the repository at this point in the history

Commits on May 2, 2024

  1. Fix: failing CI after huggingface#30568 (huggingface#30599)

    * failiing CI
    
    * no let's keep it intil full deprecation in  v4.42
    zucchini-nlp authored May 2, 2024
    Configuration menu
    Copy the full SHA
    5cf3e6b View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    fbabd67 View commit details
    Browse the repository at this point in the history
  3. Fix memory leak with CTC training script on Chinese languages (huggin…

    …gface#30358)
    
    * Fix memory leak with CTC training script on Chinese languages
    
    * Fix lint
    lucky-bai authored May 2, 2024
    Configuration menu
    Copy the full SHA
    12c5544 View commit details
    Browse the repository at this point in the history
  4. 🚨 Update image_processing_vitmatte.py (huggingface#30566)

    * Update image_processing_vitmatte.py
    
    * add test
    
    * [run-slow]vitmatte
    rb-synth authored May 2, 2024
    Configuration menu
    Copy the full SHA
    f953025 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    4ad5ada View commit details
    Browse the repository at this point in the history
  6. fix:missing output_router_logits in SwitchTransformers (huggingface…

    …#30573)
    
    * fix:missing `output_router_logits` in SwitchTransformers
    
    * fix whitespace in blank line
    lausannel authored May 2, 2024
    Configuration menu
    Copy the full SHA
    a65da83 View commit details
    Browse the repository at this point in the history
  7. Use contiguous() in clip checkpoint conversion script (huggingface#…

    …30613)
    
    * fix
    
    * fix
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored May 2, 2024
    Configuration menu
    Copy the full SHA
    f57f014 View commit details
    Browse the repository at this point in the history
  8. phi3 chat_template does not support system role (huggingface#30606)

    * phi3 chat_template does not support system role
    
    * fix doc test error
    amitportnoy authored May 2, 2024
    Configuration menu
    Copy the full SHA
    801894e View commit details
    Browse the repository at this point in the history
  9. Docs: fix generate-related rendering issues (huggingface#30600)

    * does this work?
    
    * like this?
    
    * fix the other generate links
    
    * missing these
    gante authored May 2, 2024
    Configuration menu
    Copy the full SHA
    aa55ff4 View commit details
    Browse the repository at this point in the history
  10. Docs: add missing StoppingCriteria autodocs (huggingface#30617)

    * add missing docstrings to docs
    
    * Update src/transformers/generation/stopping_criteria.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    gante and amyeroberts authored May 2, 2024
    Configuration menu
    Copy the full SHA
    66abe13 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    9719202 View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    39359e5 View commit details
    Browse the repository at this point in the history
  13. Output None as attention when layer is skipped (huggingface#30597)

    * Output `None` as attention when layer is skipped
    
    * Add test for output_attentions
    jonghwanhyeon authored May 2, 2024
    Configuration menu
    Copy the full SHA
    4c94093 View commit details
    Browse the repository at this point in the history
  14. Add HQQ quantization support (huggingface#29637)

    * update HQQ transformers integration
    
    * push import_utils.py
    
    * add force_hooks check in modeling_utils.py
    
    * fix | with Optional
    
    * force bias as param
    
    * check bias is Tensor
    
    * force forward for multi-gpu
    
    * review fixes pass
    
    * remove torch grad()
    
    * if any key in linear_tags fix
    
    * add cpu/disk check
    
    * isinstance return
    
    * add multigpu test + refactor tests
    
    * clean hqq_utils imports in hqq.py
    
    * clean hqq_utils imports in quantizer_hqq.py
    
    * delete hqq_utils.py
    
    * Delete src/transformers/utils/hqq_utils.py
    
    * ruff init
    
    * remove torch.float16 from __init__ in test
    
    * refactor test
    
    * isinstance -> type in quantizer_hqq.py
    
    * cpu/disk device_map check in quantizer_hqq.py
    
    * remove type(module) nn.linear check in quantizer_hqq.py
    
    * add BaseQuantizeConfig import inside HqqConfig init
    
    * remove hqq import in hqq.py
    
    * remove accelerate import from test_hqq.py
    
    * quant config.py doc update
    
    * add hqqconfig to main_classes doc
    
    * make style
    
    * __init__ fix
    
    * ruff __init__
    
    * skip_modules list
    
    * hqqconfig format fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * hqqconfig doc fix
    
    * test_hqq.py remove mistral comment
    
    * remove self.using_multi_gpu is False
    
    * torch_dtype default val set and logger.info
    
    * hqq.py isinstance fix
    
    * remove torch=None
    
    * torch_device test_hqq
    
    * rename test_hqq
    
    * MODEL_ID in test_hqq
    
    * quantizer_hqq setattr fix
    
    * quantizer_hqq typo fix
    
    * imports quantizer_hqq.py
    
    * isinstance quantizer_hqq
    
    * hqq_layer.bias reformat quantizer_hqq
    
    * Step 2 as comment in quantizer_hqq
    
    * prepare_for_hqq_linear() comment
    
    * keep_in_fp32_modules fix
    
    * HqqHfQuantizer reformat
    
    * quantization.md hqqconfig
    
    * quantization.md model example reformat
    
    * quantization.md # space
    
    * quantization.md space   })
    
    * quantization.md space   })
    
    * quantization_config fix doc
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * axis value check in quantization_config
    
    * format
    
    * dynamic config explanation
    
    * quant config method in quantization.md
    
    * remove shard-level progress
    
    * .cuda fix modeling_utils
    
    * test_hqq fixes
    
    * make fix-copies
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    mobicham and amyeroberts authored May 2, 2024
    Configuration menu
    Copy the full SHA
    5995299 View commit details
    Browse the repository at this point in the history
  15. Fix CI after huggingface#30410 (huggingface#30612)

    * Fix CI after huggingface#30410
    
    * [run-slow] blenderbot
    zucchini-nlp authored May 2, 2024
    Configuration menu
    Copy the full SHA
    a0e77a1 View commit details
    Browse the repository at this point in the history

Commits on May 3, 2024

  1. add mlp bias for llama models (huggingface#30031)

    * add bias
    
    * fix quality
    mayank31398 authored May 3, 2024
    Configuration menu
    Copy the full SHA
    425e1a0 View commit details
    Browse the repository at this point in the history
  2. Fix W&B run name (huggingface#30462)

    * Remove comparison to output_dir
    
    * Update docs for `run_name`
    
    * Add warning
    qubvel authored May 3, 2024
    Configuration menu
    Copy the full SHA
    66f675e View commit details
    Browse the repository at this point in the history
  3. HQQ: PEFT support for HQQ (huggingface#30632)

    Update quantizer_hqq.py
    younesbelkada authored May 3, 2024
    Configuration menu
    Copy the full SHA
    d0c72c1 View commit details
    Browse the repository at this point in the history
  4. Prevent TextGenerationPipeline._sanitize_parameters from overriding…

    … previously provided parameters (huggingface#30362)
    
    * Fixed TextGenerationPipeline._sanitize_parameters default params
    
    * removed empty spaces
    
    ---------
    
    Co-authored-by: Ng, Yen Ting <[email protected]>
    yting27 and yentingn authored May 3, 2024
    Configuration menu
    Copy the full SHA
    deb7605 View commit details
    Browse the repository at this point in the history
  5. Avoid duplication in PR slow CI model list (huggingface#30634)

    update
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored May 3, 2024
    Configuration menu
    Copy the full SHA
    91d155e View commit details
    Browse the repository at this point in the history

Commits on May 6, 2024

  1. [CI update] Try to use dockers and no cache (huggingface#29202)

    * change cis
    
    * nits
    
    * update
    
    * minor updates
    
    * [push-ci-image]
    
    * nit [push-ci-image]
    
    * nitsssss
    
    * [build-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * both
    
    * [push-ci-image]
    
    * this?
    
    * [push-ci-image]
    
    * pypi-kenlm needs g++
    
    * [push-ci-image]
    
    * nit
    
    * more nits [push-ci-image]
    
    * nits [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * add vision
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * add new dummy file but will need to update them [push-ci-image]
    
    * [push-ci-image]
    
    * show package size as well
    
    * [push-ci-image]
    
    * potentially ignore failures
    
    * workflow updates
    
    * nits [push-ci-image]
    
    * [push-ci-image]
    
    * fix consistency
    
    * clean nciida triton
    
    * also show big packages [push-ci-image]
    
    * nit
    
    * update
    
    * another one
    
    * line escape?
    
    * add accelerate [push-ci-image]
    
    * updates [push-ci-image]
    
    * nits to run tests, no push-ci
    
    * try to parse skip reason to make sure nothing is skipped that should no be skippped
    
    * nit?
    
    * always show skipped reasons
    
    * nits
    
    * better parsing of the test outputs
    
    * action="store_true",
    
    * failure on failed
    
    * show matched
    
    * debug
    
    * update short summary with skipped, failed and errors
    
    * nits
    
    * nits
    
    * coolu pdates
    
    * remove docbuilder
    
    * fix
    
    * always run checks
    
    * oups
    
    * nits
    
    * don't error out on library printing
    
    * non zero exi codes
    
    * no warning
    
    * nit
    
    * WAT?
    
    * format nit
    
    * [push-ci-image]
    
    * fail if fail is needed
    
    * [push-ci-image]
    
    * sound file for torch light?
    
    * [push-ci-image]
    
    * order is important [push-ci-image]
    
    * [push-ci-image] reduce even further
    
    * [push-ci-image]
    
    * use pytest rich !
    
    * yes [push-ci-image]
    
    * oupsy
    
    * bring back the full traceback, but pytest rich should help
    
    * nit
    
    * [push-ci-image]
    
    * re run
    
    * nit
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * empty push to trigger
    
    * [push-ci-image]
    
    * nit? [push-ci-image]
    
    * empty
    
    * try to install timm with no deps
    
    * [push-ci-image]
    
    * oups [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image] ?
    
    * [push-ci-image] open ssh client for git checkout fast
    
    * empty for torch light
    
    * updates [push-ci-image]
    
    * nit
    
    * @v4 for checkout
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * fix fetch tests with parallelism
    
    * [push-ci-image]
    
    * more parallelism
    
    * nit
    
    * more nits
    
    * empty to re-trigger
    
    * empty to re-trigger
    
    * split by timing
    
    * did not work with previous commit
    
    * junit.xml
    
    * no path?
    
    * mmm this?
    
    * junitxml format
    
    * split by timing
    
    * nit
    
    * fix junit family
    
    * now we can test if the xunit1 is compatible!
    
    * this?
    
    * fully list tests
    
    * update
    
    * update
    
    * oups
    
    * finally
    
    * use classname
    
    * remove working directory to make sure the path does not interfere
    
    * okay no juni should have the correct path
    
    * name split?
    
    * sort by classname is what make most sense
    
    * some testing
    
    * naem
    
    * oups
    
    * test something fun
    
    * autodetect
    
    * 18?
    
    * nit
    
    * file size?
    
    * uip
    
    * 4 is best
    
    * update to see versions
    
    * better print
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * please install the correct keras version
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * uv is fucking me up
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * nits
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * install issues an pins
    
    * tapas as well
    
    * nits
    
    * more paralellism
    
    * short tb
    
    * soundfile
    
    * soundfile
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * oups
    
    * [push-ci-image]
    
    * fix some things
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * use torch-light for hub
    
    * small git lfs for hub job
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * fix tf tapas
    
    * [push-ci-image]
    
    * nits
    
    * [push-ci-image]
    
    * don't update the test
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * no use them
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * update tf proba
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * woops
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * test with built dockers
    
    * [push-ci-image]
    
    * skip annoying tests
    
    * revert fix copy
    
    * update test values
    
    * update
    
    * last skip and fixup
    
    * nit
    
    * ALL GOOOD
    
    * quality
    
    * Update tests/models/layoutlmv2/test_image_processing_layoutlmv2.py
    
    * Update docker/quality.dockerfile
    
    Co-authored-by: Lysandre Debut <[email protected]>
    
    * Update src/transformers/models/tapas/modeling_tf_tapas.py
    
    Co-authored-by: Lysandre Debut <[email protected]>
    
    * Apply suggestions from code review
    
    Co-authored-by: Lysandre Debut <[email protected]>
    
    * use torch-speed
    
    * updates
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    * fuck ken-lm [push-ci-image]
    
    * [push-ci-image]
    
    * [push-ci-image]
    
    ---------
    
    Co-authored-by: Lysandre Debut <[email protected]>
    ArthurZucker and LysandreJik authored May 6, 2024
    Configuration menu
    Copy the full SHA
    307f632 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    09edd77 View commit details
    Browse the repository at this point in the history
  3. Hotfix-change-ci (huggingface#30669)

    * dmmy change
    
    * fiux
    
    * revert change
    ArthurZucker authored May 6, 2024
    Configuration menu
    Copy the full SHA
    a45c514 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    9c772ac View commit details
    Browse the repository at this point in the history
  5. Fix llava next tie_word_embeddings config (huggingface#30640)

    * fix llava next embedding
    
    * add docstring
    
    * Update src/transformers/models/llava_next/configuration_llava_next.py
    
    Co-authored-by: NielsRogge <[email protected]>
    
    ---------
    
    Co-authored-by: NielsRogge <[email protected]>
    SunMarc and NielsRogge authored May 6, 2024
    Configuration menu
    Copy the full SHA
    aa64f08 View commit details
    Browse the repository at this point in the history
  6. Trainer._load_from_checkpoint - support loading multiple Peft adapters (

    huggingface#30505)
    
    * Trainer: load checkpoint model with multiple adapters
    
    * Trainer._load_from_checkpoint support multiple active adapters
    
    * PeftModel.set_adapter does not support multiple adapters yet
    
    * Trainer._load_from_checkpoint test multiple adapters
    
    ---------
    
    Co-authored-by: Clara Luise Pohland <[email protected]>
    claralp and Clara Luise Pohland authored May 6, 2024
    Configuration menu
    Copy the full SHA
    e076953 View commit details
    Browse the repository at this point in the history
  7. Trainer - add cache clearing and the option for batched eval metrics …

    …computation (huggingface#28769)
    
    * Added cache clearing for GPU efficiency.
    
    * Added cache clearing for GPU efficiency.
    
    * Added batch_eval_metrics capability
    
    * Ran make fixup
    
    * Fixed bug
    
    * Fixed whitespace issue
    
    * Fixed outdated condition
    
    * Updated docstrings with instructions for batch_eval_metrics. Updated end of dataloader logic
    
    * Added first version of batch_eval_metrics Trainer test
    
    * Fixed batch_eval_metrics Trainer tests for both eval and predict
    
    * Fixed batch_eval_metrics behavior for new Trainer variables
    
    * Fixed batch_eval_metrics Trainer tests
    
    * Ran fixup
    FoamoftheSea authored May 6, 2024
    Configuration menu
    Copy the full SHA
    df475bf View commit details
    Browse the repository at this point in the history
  8. Fix typo: llama3.md (huggingface#30653)

    Update llama3.md
    
    fix typo
    mimbres authored May 6, 2024
    Configuration menu
    Copy the full SHA
    277db23 View commit details
    Browse the repository at this point in the history
  9. Respect resume_download deprecation (huggingface#30620)

    * Deprecate resume_download
    
    * remove default resume_download value
    
    ---------
    
    Co-authored-by: Lysandre Debut <[email protected]>
    Wauplin and LysandreJik authored May 6, 2024
    Configuration menu
    Copy the full SHA
    835de4c View commit details
    Browse the repository at this point in the history

Commits on May 7, 2024

  1. top-k instead of top-p in MixtralConfig docstring (huggingface#30687)

    top-k instead of top-p in docstring
    sorgfresser authored May 7, 2024
    Configuration menu
    Copy the full SHA
    4980d62 View commit details
    Browse the repository at this point in the history
  2. Bump jinja2 from 3.1.3 to 3.1.4 in /examples/research_projects/decisi…

    …on_transformer (huggingface#30680)
    
    Bump jinja2 in /examples/research_projects/decision_transformer
    
    Bumps [jinja2](https://github.com/pallets/jinja) from 3.1.3 to 3.1.4.
    - [Release notes](https://github.com/pallets/jinja/releases)
    - [Changelog](https://github.com/pallets/jinja/blob/main/CHANGES.rst)
    - [Commits](pallets/jinja@3.1.3...3.1.4)
    
    ---
    updated-dependencies:
    - dependency-name: jinja2
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 7, 2024
    Configuration menu
    Copy the full SHA
    a898fb9 View commit details
    Browse the repository at this point in the history
  3. Bump werkzeug from 3.0.1 to 3.0.3 in /examples/research_projects/deci…

    …sion_transformer (huggingface#30679)
    
    Bump werkzeug in /examples/research_projects/decision_transformer
    
    Bumps [werkzeug](https://github.com/pallets/werkzeug) from 3.0.1 to 3.0.3.
    - [Release notes](https://github.com/pallets/werkzeug/releases)
    - [Changelog](https://github.com/pallets/werkzeug/blob/main/CHANGES.rst)
    - [Commits](pallets/werkzeug@3.0.1...3.0.3)
    
    ---
    updated-dependencies:
    - dependency-name: werkzeug
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 7, 2024
    Configuration menu
    Copy the full SHA
    ce47582 View commit details
    Browse the repository at this point in the history
  4. Adding _tie_weights() to prediction heads to support low_cpu_mem_usag…

    …e=True (huggingface#29024)
    
    * Adding _tie_weights() to prediction heads to support low_cpu_mem_usage=True
    
    * Testing for the non-safe-tensors case, since the default is safe-tensors already
    
    * Running fixup/fix-copies
    
    * Adding accelerate annotations to tests
    hackyon authored May 7, 2024
    Configuration menu
    Copy the full SHA
    54a2361 View commit details
    Browse the repository at this point in the history
  5. Fix cache_position initialisation for generation with `use_cache=Fa…

    …lse` (huggingface#30485)
    
    * Fix cache_position init for generation
    
    * Update src/transformers/generation/utils.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Fix cache position update
    
    ---------
    
    Co-authored-by: Arthur <[email protected]>
    nurlanov-zh and ArthurZucker authored May 7, 2024
    Configuration menu
    Copy the full SHA
    4fda78c View commit details
    Browse the repository at this point in the history
  6. Word-level timestamps broken for short-form audio (huggingface#30325)

    * force chunk_length_s in AutomaticSpeechRecognitionPipeline
    
    * compute num_frames even when stride is None
    
    * add slow tests
    
    * fix test
    
    * Update src/transformers/pipelines/automatic_speech_recognition.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update tests/pipelines/test_pipelines_automatic_speech_recognition.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * add input validation
    
    * fixup
    
    * small fix
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    kamilakesbi and amyeroberts authored May 7, 2024
    Configuration menu
    Copy the full SHA
    9c8979e View commit details
    Browse the repository at this point in the history
  7. Updated docs of forward in Idefics2ForConditionalGeneration with …

    …correct `ignore_index` value (huggingface#30678)
    
    updated docs of `forward` in `Idefics2ForConditionalGeneration` with correct `ignore_index` value
    zafstojano authored May 7, 2024
    Configuration menu
    Copy the full SHA
    e5f71ec View commit details
    Browse the repository at this point in the history
  8. Bump tqdm from 4.63.0 to 4.66.3 in /examples/research_projects/decisi…

    …on_transformer (huggingface#30646)
    
    Bump tqdm in /examples/research_projects/decision_transformer
    
    Bumps [tqdm](https://github.com/tqdm/tqdm) from 4.63.0 to 4.66.3.
    - [Release notes](https://github.com/tqdm/tqdm/releases)
    - [Commits](tqdm/tqdm@v4.63.0...v4.66.3)
    
    ---
    updated-dependencies:
    - dependency-name: tqdm
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 7, 2024
    Configuration menu
    Copy the full SHA
    4051d36 View commit details
    Browse the repository at this point in the history
  9. Bump tqdm from 4.48.2 to 4.66.3 in /examples/research_projects/visual…

    …_bert (huggingface#30645)
    
    Bump tqdm in /examples/research_projects/visual_bert
    
    Bumps [tqdm](https://github.com/tqdm/tqdm) from 4.48.2 to 4.66.3.
    - [Release notes](https://github.com/tqdm/tqdm/releases)
    - [Commits](tqdm/tqdm@v4.48.2...v4.66.3)
    
    ---
    updated-dependencies:
    - dependency-name: tqdm
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 7, 2024
    Configuration menu
    Copy the full SHA
    3733391 View commit details
    Browse the repository at this point in the history
  10. Reboot Agents (huggingface#30387)

    * Create CodeAgent and ReactAgent
    
    * Fix formatting errors
    
    * Update documentation for agents
    
    * Add custom errors, improve logging
    
    * Support variable usage in ReactAgent
    
    * add messages
    
    * Add message passing format
    
    * Create React Code Agent
    
    * Update
    
    * Refactoring
    
    * Fix errors
    
    * Improve python interpreter
    
    * Only non-tensor inputs should be sent to device
    
    * Calculator tool slight refactor
    
    * Improve docstrings
    
    * Refactor
    
    * Fix tests
    
    * Fix more tests
    
    * Fix even more tests
    
    * Fix tests by replacing output and input types
    
    * Fix operand type issue
    
    * two small fixes
    
    * EM TTS
    
    * Fix agent running type errors
    
    * Change text to speech tests to allow changed outputs
    
    * Update doc with new agent types
    
    * Improve code interpreter
    
    * If max iterations reached, provide a real answer instead of an error
    
    * Add edge case in interpreter
    
    * Add safe imports to the interpreter
    
    * Interpreter tweaks: tuples and listcomp
    
    * Make style
    
    * Make quality
    
    * Add dictcomp to interpreter
    
    * Rename ReactJSONAgent to ReactJsonAgent
    
    * Misc changes
    
    * ToolCollection
    
    * Rename agent's logger to self.logger
    
    * Add while loops to interpreter
    
    * Update doc with new tools. still need to mention collections
    
    * Add collections to the doc
    
    * Small fixes on logs and interpretor
    
    * Fix toolbox return type
    
    * Docs + fixup
    
    * Skip doctests
    
    * Correct prompts with improved examples and formatting
    
    * Update prompt
    
    * Remove outdated docs
    
    * Change agent to accept Toolbox object for tools
    
    * Remove calculator tool
    
    * Propagate removal of calculator in doc
    
    * Fix 2 failing workflows
    
    * Simplify additional argument passing
    
    * AgentType audio
    
    * Minor changes: function name, types
    
    * Remove calculator tests
    
    * Fix test
    
    * Fix torch requirement
    
    * Fix final answer tests
    
    * Style fixes
    
    * Fix tests
    
    * Update docstrings with calculator removal
    
    * Small type hint fixes
    
    * Update tests/agents/test_translation.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update tests/agents/test_python_interpreter.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/agents/default_tools.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/agents/tools.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update tests/agents/test_agents.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/models/bert/configuration_bert.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/agents/tools.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/agents/speech_to_text.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update tests/agents/test_speech_to_text.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update tests/agents/test_tools_common.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * pygments
    
    * Answer comments
    
    * Cleaning up
    
    * Simplifying init for all agents
    
    * Improving prompts and making code nicer
    
    * Style fixes
    
    * Add multiple comparator test in interpreter
    
    * Style fixes
    
    * Improve BERT example in documentation
    
    * Add examples to doc
    
    * Fix python interpreter quality
    
    * Logging improvements
    
    * Change test flag to agents
    
    * Quality fix
    
    * Add example for HfEngine
    
    * Improve conversation example for HfEngine
    
    * typo fix
    
    * Verify doc
    
    * Update docs/source/en/agents.md
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/agents/agents.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/agents/prompts.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/agents/python_interpreter.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update docs/source/en/agents.md
    
    Co-authored-by: Arthur <[email protected]>
    
    * Fix style issues
    
    * local s2t tool
    
    ---------
    
    Co-authored-by: Cyril Kondratenko <[email protected]>
    Co-authored-by: Lysandre <[email protected]>
    Co-authored-by: Lysandre <[email protected]>
    Co-authored-by: Arthur <[email protected]>
    5 people authored May 7, 2024
    Configuration menu
    Copy the full SHA
    0ba15ce View commit details
    Browse the repository at this point in the history
  11. Bump tqdm from 4.48.2 to 4.66.3 in /examples/research_projects/lxmert (

    …huggingface#30644)
    
    Bumps [tqdm](https://github.com/tqdm/tqdm) from 4.48.2 to 4.66.3.
    - [Release notes](https://github.com/tqdm/tqdm/releases)
    - [Commits](tqdm/tqdm@v4.48.2...v4.66.3)
    
    ---
    updated-dependencies:
    - dependency-name: tqdm
      dependency-type: direct:production
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
    dependabot[bot] authored May 7, 2024
    Configuration menu
    Copy the full SHA
    4a17200 View commit details
    Browse the repository at this point in the history
  12. Separate tokenizer tests (huggingface#30675)

    * nit
    
    * better filter
    
    * pipeline tests should only be models/xxx not anything else
    
    * nit to better see filtering of the files that are passed to test torch
    
    * oups
    ArthurZucker authored May 7, 2024
    Configuration menu
    Copy the full SHA
    4208c42 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    05ec950 View commit details
    Browse the repository at this point in the history
  14. Rename artifact name prev_ci_results to ci_results (huggingface#3…

    …0697)
    
    * rename
    
    * update
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored May 7, 2024
    Configuration menu
    Copy the full SHA
    884e3b1 View commit details
    Browse the repository at this point in the history
  15. Add safetensors to model not found error msg for default use_safetens…

    …ors value (huggingface#30602)
    
    * add safetensors to model not found error for default use_safetensors=None case
    
    * format code w/ ruff
    
    * fix assert true typo
    davidgxue authored May 7, 2024
    Configuration menu
    Copy the full SHA
    cf7bed9 View commit details
    Browse the repository at this point in the history
  16. Pin deepspeed (huggingface#30701)

    pin ds
    muellerzr authored May 7, 2024
    Configuration menu
    Copy the full SHA
    5b7a225 View commit details
    Browse the repository at this point in the history

Commits on May 8, 2024

  1. Patch CLIP image preprocessor (huggingface#30698)

    * patch clip preprocessor
    
    * Update image_processing_clip.py
    
    * Update src/transformers/models/clip/image_processing_clip.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    rootonchair and amyeroberts authored May 8, 2024
    Configuration menu
    Copy the full SHA
    508c0bf View commit details
    Browse the repository at this point in the history
  2. Add examples for detection models finetuning (huggingface#30422)

    * Training script for object detection
    
    * Evaluation script for object detection
    
    * Training script for object detection with eval loop outside trainer
    
    * Trainer DETR finetuning
    
    * No trainer DETR finetuning
    
    * Eval script
    
    * Refine object detection example with trainer
    
    * Remove commented code and enable telemetry
    
    * No trainer example
    
    * Add requirements for object detection examples
    
    * Add test for trainer example
    
    * Readme draft
    
    * Fix uploading to HUB
    
    * Readme improvements
    
    * Update eval script
    
    * Adding tests for object-detection examples
    
    * Add object-detection example
    
    * Add object-detection resources to docs
    
    * Update README with custom dataset instructions
    
    * Update year
    
    * Replace valid with validation
    
    * Update instructions for custom dataset
    
    * Remove eval script
    
    * Remove use_auth_token
    
    * Add copied from and telemetry
    
    * Fixup
    
    * Update readme
    
    * Fix id2label
    
    * Fix links in docs
    
    * Update examples/pytorch/object-detection/run_object_detection.py
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Update examples/pytorch/object-detection/run_object_detection.py
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Move description to the top
    
    * Fix Trainer example
    
    * Update no trainer example
    
    * Update albumentations version
    
    ---------
    
    Co-authored-by: NielsRogge <[email protected]>
    qubvel and NielsRogge authored May 8, 2024
    Configuration menu
    Copy the full SHA
    998dbe0 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    1872bde View commit details
    Browse the repository at this point in the history
  4. Llava: remove dummy labels (huggingface#30706)

    remove labels from llavas
    zucchini-nlp authored May 8, 2024
    Configuration menu
    Copy the full SHA
    467164e View commit details
    Browse the repository at this point in the history
  5. Add installation of examples requirements in CI (huggingface#30708)

    * Add installation of examples requirements in CI
    
    * Update .circleci/create_circleci_config.py
    
    Co-authored-by: Arthur <[email protected]>
    
    ---------
    
    Co-authored-by: Arthur <[email protected]>
    qubvel and ArthurZucker authored May 8, 2024
    Configuration menu
    Copy the full SHA
    e7a5f45 View commit details
    Browse the repository at this point in the history
  6. Update object detection guide (huggingface#30683)

    * Object detection guide
    
    * Minor update
    
    * Minor updates, links
    
    * Fix typo
    
    * Wording, add albu space
    
    * Add missing part
    
    * Update docs/source/en/tasks/object_detection.md
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update docs/source/en/tasks/object_detection.md
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Update docs/source/en/tasks/object_detection.md
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Fix device, add imports for inference
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    Co-authored-by: NielsRogge <[email protected]>
    3 people authored May 8, 2024
    Configuration menu
    Copy the full SHA
    5962d62 View commit details
    Browse the repository at this point in the history
  7. Immutability for data collators (huggingface#30603)

    * immutability fix for seq2seq as well as immutability tests for the collators
    
    * ensure we don't act on none labels and formatting
    
    * remove tf/pt in respective tests as they are not required
    
    * more type error fixes tf/np
    
    * remove todo
    
    * apply suggestions from code review
    
    * formatting / style
    vasqu authored May 8, 2024
    Configuration menu
    Copy the full SHA
    71c1985 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    f26e407 View commit details
    Browse the repository at this point in the history

Commits on May 9, 2024

  1. Add dynamic resolution input/interpolate position embedding to SigLIP (

    …huggingface#30719)
    
    * Add interpolate positional encoding to siglip
    
    * Change # of patches for siglip interpolation test
    
    * fix formatting
    
    * Apply nit suggestions from code review
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    davidgxue and amyeroberts authored May 9, 2024
    Configuration menu
    Copy the full SHA
    60293bd View commit details
    Browse the repository at this point in the history
  2. Enable dynamic resolution for vivit (huggingface#30630)

    * feat: enable dynamic resolution for vivit
    
    * fix: formatting
    
    * remove: print statement for testing
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update tests/models/vivit/test_modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update tests/models/vivit/test_modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update tests/models/vivit/test_modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * Update src/transformers/models/vivit/modeling_vivit.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * fix: style check
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    jla524 and amyeroberts authored May 9, 2024
    Configuration menu
    Copy the full SHA
    8c5b3c1 View commit details
    Browse the repository at this point in the history
  3. Removal of deprecated maps (huggingface#30576)

    * [test_all] Remove all imports
    
    Remove remaining ARCHIVE MAPS
    
    Remove remaining PRETRAINED maps
    
    * review comments
    
    * [test_all] empty commit to trigger tests
    LysandreJik authored May 9, 2024
    Configuration menu
    Copy the full SHA
    297b732 View commit details
    Browse the repository at this point in the history
  4. Generate: add min_p sampling (huggingface#30639)

    * min_p
    
    * more relaxed test to avoid numerical issues
    
    * Update src/transformers/generation/logits_process.py
    
    Co-authored-by: menhguin <[email protected]>
    
    * Update src/transformers/generation/configuration_utils.py
    
    Co-authored-by: menhguin <[email protected]>
    
    * docstring clarifications
    
    * PR comments
    
    * Update tests/generation/test_logits_process.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    * make fixup
    
    ---------
    
    Co-authored-by: menhguin <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    3 people authored May 9, 2024
    Configuration menu
    Copy the full SHA
    df53c6e View commit details
    Browse the repository at this point in the history
  5. Fix image post-processing for OWLv2 (huggingface#30686)

    * feat: add note about owlv2
    
    * fix: post processing coordinates
    
    * remove: workaround document
    
    * fix: extra quotes
    
    * update: owlv2 docstrings
    
    * fix: copies check
    
    * feat: add unit test for resize
    
    * Update tests/models/owlv2/test_image_processor_owlv2.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    jla524 and amyeroberts authored May 9, 2024
    Configuration menu
    Copy the full SHA
    218f441 View commit details
    Browse the repository at this point in the history
  6. KV cache is no longer a model attribute (huggingface#30730)

    kv_cache is no longer a model attribute
    zucchini-nlp authored May 9, 2024
    Configuration menu
    Copy the full SHA
    5413b89 View commit details
    Browse the repository at this point in the history
  7. Generate: consistently handle special tokens as tensors (huggingface#…

    …30624)
    
    * tmp commit
    
    * [test_all] mvp
    
    * missing not
    
    * [test_all] final test fixes
    
    * fix musicgen_melody and rag
    
    * [test_all] empty commit
    
    * PR comments
    
    * Update src/transformers/generation/utils.py
    
    Co-authored-by: Arthur <[email protected]>
    
    ---------
    
    Co-authored-by: Arthur <[email protected]>
    gante and ArthurZucker authored May 9, 2024
    Configuration menu
    Copy the full SHA
    7130a22 View commit details
    Browse the repository at this point in the history
  8. Update CodeLlama references (huggingface#30218)

    * Update CodeLlama references
    
    * Update slow_documentation_tests.txt
    
    * Update slow_documentation_tests.txt
    osanseviero authored May 9, 2024
    Configuration menu
    Copy the full SHA
    c99d88e View commit details
    Browse the repository at this point in the history
  9. [docs] Update es/pipeline_tutorial.md (huggingface#30684)

    * copy en/ contect to es/
    
    * translate first section
    
    * translate the doc
    
    * fix typos
    
    * run make style
    aaronjimv authored May 9, 2024
    Configuration menu
    Copy the full SHA
    47735f5 View commit details
    Browse the repository at this point in the history

Commits on May 10, 2024

  1. Update llama3.md, fix typo (huggingface#30739)

    Update llama3.md
    
    fix typo again
    mimbres authored May 10, 2024
    Configuration menu
    Copy the full SHA
    73fcfb2 View commit details
    Browse the repository at this point in the history
  2. mlp_only_layers is more flexible than decoder_sparse_step (huggingfac…

    …e#30552)
    
    * force back to commit ba40a21 and fix workflow errors
    
    * match the review suggestions
    
    * fix ci errors
    
    * fix CI
    
    * fix ci, format code
    
    * fix ci, ruff format
    
    * fix ci, ruff format again
    
    * Update src/transformers/models/qwen2_moe/configuration_qwen2_moe.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/models/qwen2_moe/configuration_qwen2_moe.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * Update src/transformers/models/qwen2_moe/configuration_qwen2_moe.py
    
    Co-authored-by: Arthur <[email protected]>
    
    * solve this warning: Default Argument Value is mutable
    
    ---------
    
    Co-authored-by: Arthur <[email protected]>
    eigen2017 and ArthurZucker authored May 10, 2024
    Configuration menu
    Copy the full SHA
    1c52cb7 View commit details
    Browse the repository at this point in the history
  3. PEFT / Trainer: Make use of model.active_adapters() instead of depr…

    …ecated `model.active_adapter` whenever possible (huggingface#30738)
    
    * Update trainer.py
    
    * Update src/transformers/trainer.py
    
    * Update src/transformers/trainer.py
    
    * Update src/transformers/trainer.py
    
    * style
    
    * Update src/transformers/trainer.py
    
    * Update src/transformers/trainer.py
    younesbelkada authored May 10, 2024
    Configuration menu
    Copy the full SHA
    2d1602a View commit details
    Browse the repository at this point in the history
  4. [docs] Update link in es/pipeline_webserver.md (huggingface#30745)

    * update link
    
    * run make style
    aaronjimv authored May 10, 2024
    Configuration menu
    Copy the full SHA
    8ce4fef View commit details
    Browse the repository at this point in the history
  5. hqq - fix weight check in check_quantized_param (huggingface#30748)

    * hqq - fix weight check in check_quantized_param
    
    * ruff format
    mobicham authored May 10, 2024
    Configuration menu
    Copy the full SHA
    e0c3cee View commit details
    Browse the repository at this point in the history

Commits on May 13, 2024

  1. [awq] replace scale when we have GELU (huggingface#30074)

    * fix awq test
    
    * style
    
    * add log
    
    * new fix
    
    * style
    
    * only modifying impacted model in the end
    
    * rename function
    SunMarc authored May 13, 2024
    Configuration menu
    Copy the full SHA
    de6e0db View commit details
    Browse the repository at this point in the history
  2. Workflow: Replace actions/post-slack with centrally defined workflow (

    huggingface#30737)
    
    * Remove commit details
    
    * remove old workflow
    younesbelkada authored May 13, 2024
    Configuration menu
    Copy the full SHA
    a4e530e View commit details
    Browse the repository at this point in the history
  3. Blip dynamic input resolution (huggingface#30722)

    * blip with interpolated pos encoding
    
    * feat: Add interpolate_pos_encoding option to other models from `BLIP` family.
    
    * include check for textual generated content in tests
    zafstojano authored May 13, 2024
    Configuration menu
    Copy the full SHA
    f63d822 View commit details
    Browse the repository at this point in the history
  4. Support for Falcon2-11B (huggingface#30771)

    * remove unrelated changes
    
    * remove unrelated changes on phi and stable LM
    
    * add: Test for Falcon 10B
    
    * fix: formatting
    
    * fix: loading the falcon 10B in 8 bit precision using bitsanbytes.
    
    * fix: device placement
    
    * fix: broken tests.
    
    * fix: backwards compatibility for falcon 1B architecture.
    
    * chore: updated test.
    
    * chore: test_modeling_falcon.py to use the 11B model.
    
    * chore: minor edit
    
    * chore: formating.
    
    ---------
    
    Co-authored-by: Pablo Montalvo <[email protected]>
    Co-authored-by: ArthurZucker <[email protected]>
    3 people authored May 13, 2024
    Configuration menu
    Copy the full SHA
    e52741f View commit details
    Browse the repository at this point in the history
  5. [GroundingDino] Adding ms_deform_attn kernels (huggingface#30768)

    * Adding ms_deform_attn kernels to GroundingDino
    
    * Pointing to deformable detr kernels
    EduardoPach authored May 13, 2024
    Configuration menu
    Copy the full SHA
    453893e View commit details
    Browse the repository at this point in the history
  6. Llama: fix custom 4D masks, v2 (huggingface#30348)

    * 4d mask fixes
    
    * Update custom 4D mask logic
    
    * test moved to mixin
    
    * extra tests 4d mask
    
    * upd 4d mask and StaticCache handling
    
    * added Mask4DTestHard to mistral tests
    
    * post-rebase fixes
    
    * test fixes for StaticCache
    
    * make fix-copies
    
    * upd 1 after huggingface#30476
    
    * fix common tests
    
    * rm elif attention_mask.dim() == 4:
    
    * tests combined, fixed, mixtral supported
    
    * bigbird style chg reverted
    
    * rm if attention_mask.dim() == 2
    
    * modeling_llama formatting chg
    
    ---------
    
    Co-authored-by: Joao Gante <[email protected]>
    poedator and gante authored May 13, 2024
    Configuration menu
    Copy the full SHA
    a0779b9 View commit details
    Browse the repository at this point in the history
  7. Generation / FIX: Fix multi-device generation (huggingface#30746)

    * attempt to fix multi-device generation
    
    * fix
    
    * final fix
    
    * final fix
    
    * fix
    
    * fix
    
    * fix
    
    * fix
    
    * add joao suggestion
    
    * fix
    younesbelkada authored May 13, 2024
    Configuration menu
    Copy the full SHA
    f823fec View commit details
    Browse the repository at this point in the history
  8. Qwen: incorrect setup flag (huggingface#30776)

    qwen does not support the new cache classes
    gante authored May 13, 2024
    Configuration menu
    Copy the full SHA
    f4dc26d View commit details
    Browse the repository at this point in the history
  9. enable Pipeline to get device from model (huggingface#30534)

    * check model.device
    
    * fix
    
    * style fix
    
    * move model device
    
    * remove print
    
    * add comment
    
    * fix
    
    * add unit test
    
    * optimize
    
    * change test names and add more cases
    
    * Update tests/pipelines/test_pipelines_common.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    faaany and amyeroberts authored May 13, 2024
    Configuration menu
    Copy the full SHA
    69d9bca View commit details
    Browse the repository at this point in the history
  10. [Object detection pipeline] Lower threshold (huggingface#30710)

    * Lower threshold
    
    * Address comment
    NielsRogge authored May 13, 2024
    Configuration menu
    Copy the full SHA
    ce87dca View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    de2f722 View commit details
    Browse the repository at this point in the history
  12. Port IDEFICS to tensorflow (huggingface#26870)

    * Initial commit
    
    * Just a copy of modeling_idefics.py that will be ported to TF
    
    * - Prepend TF to the name of all classes
    - Convert pytorch ops to TF (not all operations are converted yet)
    
    * Add TF imports
    
    * Add autotranslated files
    
    * Add TF classes to model_tf_auto.py
    
    * Add the TF classes in model_doc
    
    * include auto-translated code
    
    * Adopted from auto-translated version
    
    * Add a forgotten super().build
    
    * Add test code for TF version.
    
    * Fix indentation and load pytorch weights for now
    
    * Some fixes. Many tests are still failing but some are passing now.
    
    - I have added TODO's for some of the hacks I made to unblock me
      and I will address them soon
    - I have the processing_idefics.py hacked in my view to support TF temporarily
    
    * Add ALL_LAYERNORM_LAYERS to match pytorch
    
    * Revert "Add ALL_LAYERNORM_LAYERS to match pytorch"
    
    This reverts commit 7e0a351 as it
    is not needed in the tf implementation.
    
    * Fix freeze_relevant_params()
    
    * Some more fixes
    
    * Fix test_attention_outputs
    
    * Add tf stuff to processing_idefics.py
    
    processing_idefics.py supports both pytorch and tf now.
    
    test_processor_idefics.py for pytorch is passing, so i didn't break anything
    but still some issues with tf. I also need to add tf tests in
    test_processor_idefics.py.
    
    * Pass return_tensors to image processing code and fix test
    
    * Pass return_tensors to the image processor __init__
    
    * Fix several test cases
    
    - Make input to some of the forward pass of type `TFModelInputType`
    - Decorate main layer forward pass with `@unpack_inputs`
    - Decorate main layer with `@keras_serializable`
    - Pass `inputs` to TFIdeficsModel
    
    * Some more fixes forgotten in last commit
    
    * Fix processing code and vision_tf.py
    
    * Fix perceiver bug
    
    * Import from
    
    * Auto-add build() methods + style pass
    
    * Fix build() errors due to `None` being passed as shape to some layers
    
    * Change name in TFIdeficsForVisionText2Text to attribute in IdeficsForVisionText2Text
    
    * Fix pytorch weights load for tf2
    
    There were a lot of `name=` missing in weight initialization code.
    
    * Attempt to fix CI
    
    * Add back accidently removed line
    
    * Remove torch-specific stuff from the TF test file
    
    * make fix-copies, make style, remove autotranslated files
    
    * Fixes to imports/docstrings
    
    * Let's try the from future import in desperation
    
    * Fix the core random_attention_mask fn to match the torch/flax behaviour
    
    * Clean random_attention_mask up correctly
    
    * Remove torch-only test
    
    * Fix loss shape, couple of nits
    
    * make style
    
    * Don't test for OOB embeddings because IDEFICS uses those deliberately
    
    * Fix loss computation to handle masking
    
    * Fix test failures when flattening
    
    * Fix some test failures
    
    - Add cross attention gate which was missing and wasn't being passed arround
    - Fix overwriting of image_attention_mask due to hack I had for dummy inputs
    
    * Add a proper stateless scaled_dot_product_attention
    
    * make style
    
    * Adding missing attribute from the PyTorch version
    
    * Small cleanups to decoupledlinearlayer in case that helps
    
    * Pass epsilon to LayerNormalization
    
    * Attemp to fix pytorch weight cross-loading for TFIdeficsEmbedding
    
    * Fix a bug in TFIdeficsGatedCrossAttentionLayer
    
    * Patching up build() methods
    
    * Constant self.inv_freq
    
    * Constant self.inv_freq
    
    * First working version
    
    The TF implementation works now, there was a bug in the TFIdeficsDecoupledLinear
    where the weights were mis-intialized (in_features,out_features)
    when it should be: (out_features, in_features)
    
    I have tested this so far with tiny-random and idefics-9b-instruct
    and gives correct output.
    
    I also dumped the final outputs for both pytorch and TF
    and they are identical.
    
    * Fix some test failures
    
    * remove print statement
    
    * Fix return_tensors
    
    * Fix CI test failure check_code_quality
    
    * Attempt to fix CI failures by running `make fixup`
    
    The hardcoded IDs in test_modeling_tf_idefics.py are for the integration
    test and makes that file unreadable and should probably be moved to a seperate file.
    
    * Attempt to fix tests_pr_documentation_tests
    
    * Fix a test failure in test_image_processing_idefics.py
    
    * Fix test test_pt_tf_model_equivalence
    
    * Fix a few failures
    
    * Tiny fix
    
    * Some minor fixes
    
    * Remove a duplicate test
    
    * Override a few test failures for IDEFICS
    
    - `test_keras_save_load` is passing now
    - `test_compile_tf_model` is still failing
    
    * Fix processing_idefics.py after rebase
    
    * Guard import keras with is_tf_available
    
    * fix check code quality
    
    * fix check code quality
    
    * Minor fixes
    
    * Skip test_save_load temporarily
    
    This test passed on my local box but fails on the CI, skipping
    for now to see if there are other remaining failures on the CI.
    
    * Run `ruff format tests src utils`
    
    * Fix last failing test, `test_compile_tf_model`
    
    * Add fixes for vision_tf.py
    
    I forgot to add this file in last commit.
    
    * Minor fixes
    
    * Replace "<<<" with "<<" for doc tests
    
    IDEFICS-9B is too big for doctest runner, so don't run it there
    
    * Make code more readable
    
    * Fix bug after code review
    
    I added a layer_norm_eps to IdeficsConfig but I don't even need it
    since the vision config has a layer_norm_eps.
    
    * Fix after code review
    
    Use original code tokenizer.convert_tokens_to_ids
    
    * Keep PyTorch as the default return_tensors
    
    * Fixes to modeling_tf after code review
    
    * Fixes from code review
    
    - Remove all references of `TF_IDEFICS_PRETRAINED_MODEL_ARCHIVE_LIST`
    - Pass 1e-5 to LayerNormalization in perceiver
    
    * Run ruff
    
    * Undo a change
    
    * Refactor processing code after Matt's suggestion
    
    * Remove TODO's that aren't needed anymore
    
    * For pytorch, Use original pytorch processing code from main
    
    Since this PR is a TF port it shouldn't make any modifications
    to pytorch IDEFICS code. This changes undo's the pytorch processing
    modifications I made and uses original code from main.
    
    * Update tests/models/idefics/test_modeling_idefics.py
    
    * Update tests/models/idefics/test_modeling_tf_idefics.py
    
    * Add missing imports for is_pt_tf_cross_test
    
    * [DO NOT MERGE]: This is a commit for debugging and will be reverted
    
    The cross test `test_pt_tf_model_equivalence` passes locally but
    fails when running on the CI. This commit is to help debug that
    and will be reverted.
    
    * Revert "[DO NOT MERGE]: This is a commit for debugging and will be reverted"
    
    This reverts commit 8f0d709.
    
    * [DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted
    
    * [DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted
    
    * Revert "[DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted"
    
    This reverts commit 998cc38.
    
    * Revert "[DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted"
    
    This reverts commit 1c695ac.
    
    * Don't skip test_save_load
    
    IIRC test_save_load was also failing on the CI but not on my local
    box, it might be easier to debug that on the CI first than the cross tests
    
    * Debugging commit, will be reverted
    
    * Revert "Debugging commit, will be reverted"
    
    This reverts commit 8eafc8e.
    
    * Override `test_save_load` and push model to save
    
    Maybe this will help me repro this weird bug
    
    * pass my repo_id
    
    * add endpoint
    
    * Pass a temp (write) token just for this CI
    
    * Undo last few commits, still pushing to hub for model debugging
    
    The issue seems to be with save_pretrained(),  when I looked at the model saved
    from the CI test failure it is basically empty and has no weights.
    `self.save_weights(..)` seems to be failing in save_pretrained but needs
    more debugging
    
    * Add logging to modeling tf utils, will be reverted just for debugging
    
    * Debugging, will revert
    
    * Revert "Debugging, will revert"
    
    This reverts commit 9d0d307.
    
    * Revert "Add logging to modeling tf utils, will be reverted just for debugging"
    
    This reverts commit 774b6b7.
    
    * Remove `test_save_load`
    
    The CI failures are gone after my latest rebase, no idea why
    but I was still saving the model to my hub on HF and the tf_model.h5
    file now has everything.
    
    * Run make fix-copies
    
    * Run ruff format tests src utils
    
    * Debugging commit, will be reverted
    
    * Run ruff, also trigger CI run
    
    * Run ruff again
    
    * Undo debugging commit
    
    ---------
    
    Co-authored-by: Matt <[email protected]>
    Co-authored-by: Matt <[email protected]>
    3 people authored May 13, 2024
    Configuration menu
    Copy the full SHA
    9430635 View commit details
    Browse the repository at this point in the history
  13. Generate: assistant should be greedy in assisted decoding (huggingfac…

    …e#30778)
    
    * assistant should be greedy
    
    * better comment
    
    * Update src/transformers/generation/candidate_generator.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: amyeroberts <[email protected]>
    gante and amyeroberts authored May 13, 2024
    Configuration menu
    Copy the full SHA
    2e27291 View commit details
    Browse the repository at this point in the history
  14. Save other CI jobs' result (torch/tf pipeline, example, deepspeed etc) (

    huggingface#30699)
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * update
    
    * Update utils/notification_service.py
    
    Co-authored-by: amyeroberts <[email protected]>
    
    ---------
    
    Co-authored-by: ydshieh <[email protected]>
    Co-authored-by: amyeroberts <[email protected]>
    3 people authored May 13, 2024
    Configuration menu
    Copy the full SHA
    82c1625 View commit details
    Browse the repository at this point in the history
  15. Deprecate models script (huggingface#30184)

    * Add utility for finding candidate models for deprecation
    
    * Update model init
    
    * Make into configurable script
    
    * Fix path
    
    * Add sorting of base object alphabetically
    
    * Tidy
    
    * Refactor __init__ alpha ordering
    
    * Update script with logging
    
    * fix import
    
    * Fix logger
    
    * Fix logger
    
    * Get config file before moving files
    
    * Take models from CLI
    
    * Split models into lines to make easier to feed to deprecate_models script
    
    * Update
    
    * Use posix path
    
    * Print instead
    
    * Add example in module docstring
    
    * Fix up
    
    * Add clarifying comments; add models to DEPRECATE_MODELS
    
    * Address PR comments
    
    * Don't update relative paths on the same level
    amyeroberts authored May 13, 2024
    Configuration menu
    Copy the full SHA
    0f8fefd View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    539ed75 View commit details
    Browse the repository at this point in the history
  17. CI: update to ROCm 6.0.2 and test MI300 (huggingface#30266)

    * update to ROCm 6.0.2 and test MI300
    
    * add callers for mi300
    
    * update dockerfile
    
    * fix trainer tests
    
    * remove apex
    
    * style
    
    * Update tests/trainer/test_trainer_seq2seq.py
    
    * Update tests/trainer/test_trainer_seq2seq.py
    
    * Update tests/trainer/test_trainer_seq2seq.py
    
    * Update tests/trainer/test_trainer_seq2seq.py
    
    * update to torch 2.3
    
    * add workflow dispatch target
    
    * we may need branches: mi300-ci after all
    
    * nit
    
    * fix docker build
    
    * nit
    
    * add check runner
    
    * remove docker-gpu
    
    * fix issues
    
    * fix
    
    ---------
    
    Co-authored-by: Yih-Dar <[email protected]>
    Co-authored-by: ydshieh <[email protected]>
    3 people authored May 13, 2024
    Configuration menu
    Copy the full SHA
    37bba2a View commit details
    Browse the repository at this point in the history

Commits on May 14, 2024

  1. Fix OWLv2 Doc (huggingface#30794)

    fix: owlv2 doc
    jla524 authored May 14, 2024
    Configuration menu
    Copy the full SHA
    449894d View commit details
    Browse the repository at this point in the history
  2. Fix cache type in Idefics2 (huggingface#30729)

    standardize cache in idefics2
    zucchini-nlp authored May 14, 2024
    Configuration menu
    Copy the full SHA
    c02d302 View commit details
    Browse the repository at this point in the history
  3. PEFT: Access active_adapters as a property in Trainer (huggingface#30790

    )
    
    Access active_adapters as a property
    pashminacameron authored May 14, 2024
    Configuration menu
    Copy the full SHA
    65ea190 View commit details
    Browse the repository at this point in the history
  4. Add Watermarking LogitsProcessor and WatermarkDetector (huggingface#2…

    …9676)
    
    * add watermarking processor
    
    * remove the other hashing (context width=1 always)
    
    * make style
    
    * Update src/transformers/generation/logits_process.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/logits_process.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/logits_process.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/configuration_utils.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * update watermarking process
    
    * add detector
    
    * update tests to use detector
    
    * fix failing tests
    
    * rename `input_seq`
    
    * make style
    
    * doc for processor
    
    * minor fixes
    
    * docs
    
    * make quality
    
    * Update src/transformers/generation/configuration_utils.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/logits_process.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/watermarking.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/watermarking.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/watermarking.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * add PR suggestions
    
    * let's use lru_cache's default max size (128)
    
    * import processor if torch available
    
    * maybe like this
    
    * lets move the config to torch independet file
    
    * add docs
    
    * tiny docs fix to make the test happy
    
    * Update src/transformers/generation/configuration_utils.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * Update src/transformers/generation/watermarking.py
    
    Co-authored-by: Joao Gante <[email protected]>
    
    * PR suggestions
    
    * add docs
    
    * fix test
    
    * fix docs
    
    * address pr comments
    
    * style
    
    * Revert "style"
    
    This reverts commit 7f33cc3.
    
    * correct style
    
    * make doctest green
    
    ---------
    
    Co-authored-by: Joao Gante <[email protected]>
    zucchini-nlp and gante authored May 14, 2024
    Configuration menu
    Copy the full SHA
    5ad960f View commit details
    Browse the repository at this point in the history
  5. fix merge

    younesbelkada committed May 14, 2024
    Configuration menu
    Copy the full SHA
    3ed384f View commit details
    Browse the repository at this point in the history
  6. fix unconsistent type

    younesbelkada committed May 14, 2024
    Configuration menu
    Copy the full SHA
    55eb860 View commit details
    Browse the repository at this point in the history
  7. more

    younesbelkada committed May 14, 2024
    Configuration menu
    Copy the full SHA
    f754335 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    d8f8a9c View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    a449078 View commit details
    Browse the repository at this point in the history

Commits on May 15, 2024

  1. fix tokenizer

    younesbelkada committed May 15, 2024
    Configuration menu
    Copy the full SHA
    3bdbb2e View commit details
    Browse the repository at this point in the history
  2. Update src/transformers/modeling_utils.py

    Co-authored-by: amyeroberts <[email protected]>
    younesbelkada and amyeroberts authored May 15, 2024
    Configuration menu
    Copy the full SHA
    0ab79f6 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    65433c4 View commit details
    Browse the repository at this point in the history
  4. from_gguf -> gguf_file

    younesbelkada committed May 15, 2024
    Configuration menu
    Copy the full SHA
    1b5ae54 View commit details
    Browse the repository at this point in the history
  5. replace on docs too

    younesbelkada committed May 15, 2024
    Configuration menu
    Copy the full SHA
    d6b67c6 View commit details
    Browse the repository at this point in the history