-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Loading GGUF support #2
base: main
Are you sure you want to change the base?
Commits on Apr 19, 2024
-
Fix config + attn_implementation in AutoModelForCausalLM.from_pretrai…
…ned (huggingface#30299) * Update modeling_utils.py * Update test_modeling_utils.py * Update test_modeling_utils.py * Update test_modeling_utils.py
Configuration menu - View commit details
-
Copy full SHA for 21c912e - Browse repository at this point
Copy the full SHA 21c912eView commit details -
Add TF swiftformer (huggingface#23342)
* Duplicate swiftformer * Convert SwiftFormerPatchEmbedding * Convert SwiftFormerEmbeddings * Convert TFSwiftFormerMlp * Convert TFSwiftFormerConvEncoder * Convert TFSwiftFormerLocalRepresentation * convert TFSwiftFormerEncoderBlock * Convert SwiftFormerStage * Convert SwiftFormerEncoder * Add TFSWiftFormerPreTrainedModel * Convert SwiftFormerForImageClassification * Add kwargs and start drop path * Fix syntax * Change Model class name * Add TFSwiftFormer to __init__ * Duplicate test_modeling_swiftformer * First test conversions * Change require_torch to require_tf * Add exports to swiftformer __init__ * Add TFSwiftFormerModel wrapper * Fix __init__ and run black * Remove docstring from MainLayer, fix padding * Use keras.layers.Activation on keras.Sequential * Fix swiftformer exports * Fix activation layer from config * Remove post_inits * Use tf.keras.layers.ZeroPadding2D * Convert torch normalize * Change tf test input shape * Fix softmax and reduce_sum * Convert expand_dims and repeat * Add missing reshape and tranpose * Simplify TFSwiftFormerEncoderBlock.call * Fix mismatch in patch embeddings * Fix expected output shape to match channels last * Fix swiftformer typo * Disable test_onnx * Fix TFSwiftFormerForImageClassification call * Add unpack inputs * Convert flatten(2).mean(-1) * Change vision dummy inputs (to be reviewed) * Change test_forward_signature to use .call * Fix @unpack_inputs * Set return_tensors="tf" and rename class * Rename wrongly named patch_embeddings layer * Add serving_output and change dummy_input shape * Make dimensions BCHW and transpose inside embedding layer * Change SwiftFormerEncoderBlock * Fix ruff problems * Add image size to swiftformer config * Change tranpose to MainLayer and use -1 for reshape * Remove serving_outputs and dummy_inputs * Remove test_initialization test from tf model * Make Sequential component a separate layer * Fix layers' names * Tranpose encoder outputs * Fix tests and check if hidden states is not None * Fix TFSwiftFormerForImageClassification * Run make fixup * Run make fix-copies * Update modeling_tf_auto * Update docs * Fix modeling auto mapping * Update modelint_tf_swiftformer docs * Fill image_size doc and type * Add reduction=None to loss computation * Update docs * make style * Debug: Delete the tip to see if that changes anything * Re-add tip * Remove add_code_sample_docstrings * Remove unused import * Get the debug to actually tell us the problem it has with the docs * Try a substitution to match the PyTorch file? * Add swiftformer to ignore list * Add build() methods * Update copyright year Co-authored-by: amyeroberts <[email protected]> * Remove FIXME comment * Remove from_pt * Update copyright year Co-authored-by: amyeroberts <[email protected]> * Rename one-letter variables * Remove FIXMEs related to momentum * Remove old TODO comment * Remove outstanding FIXME comments * Get dropout rate from config * Add specific dropout config for MLP * Add convencoder dropout to config * Pass config to SwiftFormerDropPath layer * Fix drop_path variable name and add Adapted from comment * Run ruff * Removed copied from comment * Run fix copies * Change drop_path to identity to match pt * Cleanup build() methods and move to new keras imports * Update docs/source/en/model_doc/swiftformer.md Co-authored-by: Matt <[email protected]> * Raise error if drop_path_rate > 0.0 * Apply suggestions from code review Replace (self.dim), with self.dim, Co-authored-by: Matt <[email protected]> * Remove drop_path function * Add training to TFSwiftFormerEncoder * Set self.built = True last Co-authored-by: amyeroberts <[email protected]> * Should have been added to previous commit Co-authored-by: amyeroberts <[email protected]> * Apply suggestions from code review Co-authored-by: amyeroberts <[email protected]> * Change default_feature_extractor to default_image_processor Co-authored-by: amyeroberts <[email protected]> * Import Keras from modeling_tf_utils * Remove relative import * Run ruff --fix * Move import keras to tf_available * Add copied from comment to test_forward_signature * Reduce batch size and num_labels * Extract loss logic to hf_compute_loss * Run ruff format --------- Co-authored-by: Matt <[email protected]> Co-authored-by: amyeroberts <[email protected]> Co-authored-by: Matt <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for d2cec09 - Browse repository at this point
Copy the full SHA d2cec09View commit details -
[Grounding DINO] Add resources (huggingface#30232)
* Add resources * Address comments * Apply suggestions from code review Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 8c12690 - Browse repository at this point
Copy the full SHA 8c12690View commit details
Commits on Apr 22, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 81e4324 - Browse repository at this point
Copy the full SHA 81e4324View commit details -
Configuration menu - View commit details
-
Copy full SHA for 8a0d5b8 - Browse repository at this point
Copy the full SHA 8a0d5b8View commit details -
Nits for model docs (huggingface#29795)
* Update llava_next.md * Update seggpt.md
Configuration menu - View commit details
-
Copy full SHA for b20b017 - Browse repository at this point
Copy the full SHA b20b017View commit details -
Enable multi-device for more models (huggingface#30379)
* feat: support for vitmatte * feat: support for vivit * feat: support for beit * feat: support for blip :D * feat: support for data2vec
Configuration menu - View commit details
-
Copy full SHA for 8b02bb6 - Browse repository at this point
Copy the full SHA 8b02bb6View commit details -
GenerationConfig: warn if pad token is negative (huggingface#30187)
* warn if pad token is negative * Update src/transformers/generation/configuration_utils.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/generation/configuration_utils.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/configuration_utils.py Co-authored-by: Joao Gante <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]> Co-authored-by: Joao Gante <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9138935 - Browse repository at this point
Copy the full SHA 9138935View commit details -
Add FSDP config for CPU RAM efficient loading through accelerate (hug…
…gingface#30002) * Add FSDP config for CPU RAM efficient loading * Style fix * Update src/transformers/training_args.py Co-authored-by: Zach Mueller <[email protected]> * Update src/transformers/training_args.py Co-authored-by: amyeroberts <[email protected]> * Add sync_module_states and cpu_ram_efficient_loading validation logic * Update src/transformers/training_args.py Co-authored-by: amyeroberts <[email protected]> * Style --------- Co-authored-by: Zach Mueller <[email protected]> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for f16caf4 - Browse repository at this point
Copy the full SHA f16caf4View commit details -
Configuration menu - View commit details
-
Copy full SHA for 08534f3 - Browse repository at this point
Copy the full SHA 08534f3View commit details -
Configuration menu - View commit details
-
Copy full SHA for ebd9944 - Browse repository at this point
Copy the full SHA ebd9944View commit details -
Llama
family, fixuse_cache=False
generation (huggingface#30380)* nit to make sure cache positions are not sliced * fix other models * nit * style
Configuration menu - View commit details
-
Copy full SHA for 2d92db8 - Browse repository at this point
Copy the full SHA 2d92db8View commit details -
Update docstrings for text generation pipeline (huggingface#30343)
* Update docstrings for text generation pipeline * Fix docstring arg * Update docstring to explain chat mode * Fix doctests * Fix doctests
Configuration menu - View commit details
-
Copy full SHA for 0e9d44d - Browse repository at this point
Copy the full SHA 0e9d44dView commit details -
Terminator strings for generate() (huggingface#28932)
* stash commit (will discard all of this) * stash commit * First commit - needs a lot of testing! * Add a test * Fix imports and make the tests actually test something * Tests pass! * Rearrange test * Add comments (but it's still a bit confusing) * Stop storing the tokenizer * Comment fixup * Fix for input_ids with a single sequence * Update tests to test single sequences * make fixup * Fix incorrect use of isin() * Expand tests to catch more cases * Expand tests to catch more cases * make fixup * Fix length calculation and update tests * Handle Ġ as a space replacement too * Update src/transformers/generation/stopping_criteria.py Co-authored-by: Joao Gante <[email protected]> * Add optimizations from Joao's suggestion * Remove TODO * Update src/transformers/generation/stopping_criteria.py Co-authored-by: Joao Gante <[email protected]> * Update tests/generation/test_stopping_criteria.py Co-authored-by: Joao Gante <[email protected]> * make fixup * Rename some variables and remove some debugging clauses for clarity * Add tests for the sub-methods * Clarify one test slightly * Add stop_strings to GenerationConfig * generate() supports stop_string arg, asks for tokenizer if not provided * make fixup * Cleanup code and rename variables for clarity * Update tokenizer error * Update tokenizer passing, handle generation on GPU * Slightly more explanation cleanup * More comment cleanup * Factor out the token cleanup so it's more obvious what we're doing, and we can change it later * Careful with that cleanup! * Cleanup + optimizations to _get_matching_positions * More minor performance tweaks * Implement caching and eliminate some expensive ops (startup time: 200ms -> 9ms) * Remove the pin_memory call * Parallelize across all stop strings! * Quick fix for tensor devices * Update embeddings test for the new format * Fix test imports * Manual patching for BERT-like tokenizers * Return a bool vector instead of a single True/False * Better comment * Better comment * Add tests from @zucchini-nlp * Amy's list creation nit * tok_list -> token_list * Push a big expanded docstring (should we put it somewhere else?) * Expand docstrings * Docstring fixups * Rebase * make fixup * Make a properly general method for figuring out token strings * Fix naming throughout the functions * Move cache, refactor, fix tests * Add comment * Remove finished TODO * Remove finished TODO * make fixup * Update src/transformers/generation/stopping_criteria.py Co-authored-by: amyeroberts <[email protected]> * Update and shorten docstring * Update tests to be shorter/clearer and test specific cases --------- Co-authored-by: Joao Gante <[email protected]> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 0d84901 - Browse repository at this point
Copy the full SHA 0d84901View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5c913ec - Browse repository at this point
Copy the full SHA 5c913ecView commit details -
Fix layerwise GaLore optimizer hard to converge with warmup scheduler (…
…huggingface#30372) Update optimization.py
Configuration menu - View commit details
-
Copy full SHA for f3b3533 - Browse repository at this point
Copy the full SHA f3b3533View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6c7335e - Browse repository at this point
Copy the full SHA 6c7335eView commit details -
Fix DETA save_pretrained (huggingface#30326)
* Add class_embed to tied weights for DETA * Fix test_tied_weights_keys for DETA model * Replace error raise with assert statement
Configuration menu - View commit details
-
Copy full SHA for 13b3b90 - Browse repository at this point
Copy the full SHA 13b3b90View commit details -
FIX / PEFT: Pass device correctly to peft (huggingface#30397)
pass device correctly to peft
Configuration menu - View commit details
-
Copy full SHA for 367a0db - Browse repository at this point
Copy the full SHA 367a0dbView commit details -
Configuration menu - View commit details
-
Copy full SHA for 8b81bfb - Browse repository at this point
Copy the full SHA 8b81bfbView commit details -
Configuration menu - View commit details
-
Copy full SHA for c49f1a8 - Browse repository at this point
Copy the full SHA c49f1a8View commit details -
Add sdpa and fa2 the Wav2vec2 family. (huggingface#30121)
* add sdpa to wav2vec. Co-authored-by: kamilakesbi <[email protected]> Co-authored-by: jp1924 <[email protected]> * add fa2 to wav2vec2 * add tests * fix attention_mask compatibility with fa2 * minor dtype fix * replace fa2 slow test * fix fa2 slow test * apply code review + add fa2 batch test * add sdpa and fa2 to hubert * sdpa and fa2 to data2vec_audio * sdpa and fa2 to Sew * sdpa to unispeech + unispeech sat * small fix * attention mask in tests Co-authored-by: Sanchit Gandhi <[email protected]> * add_speedup_benchmark_to_doc --------- Co-authored-by: [email protected] <[email protected]> Co-authored-by: Sanchit Gandhi <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 569743f - Browse repository at this point
Copy the full SHA 569743fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7fa538b - Browse repository at this point
Copy the full SHA 7fa538bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5485327 - Browse repository at this point
Copy the full SHA 5485327View commit details -
[FEAT]: EETQ quantizer support (huggingface#30262)
* [FEAT]: EETQ quantizer support * Update quantization.md * Update docs/source/en/main_classes/quantization.md Co-authored-by: Marc Sun <[email protected]> * Update docs/source/en/quantization.md Co-authored-by: Marc Sun <[email protected]> * Update docs/source/en/quantization.md Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/integrations/__init__.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/integrations/__init__.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/integrations/eetq.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/integrations/eetq.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/integrations/eetq.py Co-authored-by: Marc Sun <[email protected]> * Update tests/quantization/eetq_integration/test_eetq.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/quantizers/auto.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/quantizers/auto.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/quantizers/auto.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/quantizers/quantizer_eetq.py Co-authored-by: Marc Sun <[email protected]> * Update tests/quantization/eetq_integration/test_eetq.py Co-authored-by: Marc Sun <[email protected]> * Update src/transformers/quantizers/quantizer_eetq.py Co-authored-by: Marc Sun <[email protected]> * Update tests/quantization/eetq_integration/test_eetq.py Co-authored-by: Marc Sun <[email protected]> * Update tests/quantization/eetq_integration/test_eetq.py Co-authored-by: Marc Sun <[email protected]> * [FEAT]: EETQ quantizer support * [FEAT]: EETQ quantizer support * remove whitespaces * update quantization.md * style * Update docs/source/en/quantization.md Co-authored-by: Younes Belkada <[email protected]> * add copyright * Update quantization.md * Update docs/source/en/quantization.md Co-authored-by: amyeroberts <[email protected]> * Update docs/source/en/quantization.md Co-authored-by: amyeroberts <[email protected]> * Address the comments by amyeroberts * style --------- Co-authored-by: Marc Sun <[email protected]> Co-authored-by: Marc Sun <[email protected]> Co-authored-by: Younes Belkada <[email protected]> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for b4c18a8 - Browse repository at this point
Copy the full SHA b4c18a8View commit details -
[docs] LLM inference (huggingface#29791)
* first draft * feedback * static cache snippet * feedback * feedback
Configuration menu - View commit details
-
Copy full SHA for e74d793 - Browse repository at this point
Copy the full SHA e74d793View commit details
Commits on Apr 23, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 6c1295a - Browse repository at this point
Copy the full SHA 6c1295aView commit details -
Add inputs embeds in generation (huggingface#30269)
* Add inputs embeds in generation * always scale embeds * fix-copies * fix failing test * fix copies once more * remove embeds for models with scaling * second try to revert * codestyle
Configuration menu - View commit details
-
Copy full SHA for 408453b - Browse repository at this point
Copy the full SHA 408453bView commit details -
[Grounding DINO] Add support for cross-attention in GroundingDinoMult…
…iHeadAttention (huggingface#30364) * Added cross attention support * Fixed dtypes * Fixed assumption * Moved to decoder
Configuration menu - View commit details
-
Copy full SHA for c651ea9 - Browse repository at this point
Copy the full SHA c651ea9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 179ab09 - Browse repository at this point
Copy the full SHA 179ab09View commit details -
fix: link to HF repo/tree/revision when a file is missing (huggingfac…
…e#30406) fix: link to HF repo tree when a file is missing
Configuration menu - View commit details
-
Copy full SHA for 04ac324 - Browse repository at this point
Copy the full SHA 04ac324View commit details -
[tests] add
require_torch_sdpa
for test that needs sdpa support (hu……ggingface#30408) * add cuda flag * check for sdpa * add bitsandbytes
Configuration menu - View commit details
-
Copy full SHA for 2d61823 - Browse repository at this point
Copy the full SHA 2d61823View commit details -
Configuration menu - View commit details
-
Copy full SHA for 31921d8 - Browse repository at this point
Copy the full SHA 31921d8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 074f05e - Browse repository at this point
Copy the full SHA 074f05eView commit details -
Configuration menu - View commit details
-
Copy full SHA for ca8363e - Browse repository at this point
Copy the full SHA ca8363eView commit details -
Fix on "cache position" for assisted generation (huggingface#30068)
* clean commit history I hope * get kv seq length correctly * PR suggestions * Update src/transformers/testing_utils.py Co-authored-by: Joao Gante <[email protected]> * add comment * give gpt bigcode it's own overriden method * remove code --------- Co-authored-by: Joao Gante <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 77b59dc - Browse repository at this point
Copy the full SHA 77b59dcView commit details -
fix for itemsize => element_size() for torch backwards compat (huggin…
…gface#30133) * fix for itemsize => element_size() for torch backwards compat * improve handling of element counting * Update src/transformers/modeling_utils.py * fixup * Update src/transformers/modeling_utils.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: Younes Belkada <[email protected]> Co-authored-by: Younes Belkada <[email protected]> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 57fc00f - Browse repository at this point
Copy the full SHA 57fc00fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 4b63d01 - Browse repository at this point
Copy the full SHA 4b63d01View commit details -
FIX: re-add bnb on docker image (huggingface#30427)
Update Dockerfile
Configuration menu - View commit details
-
Copy full SHA for d179b9d - Browse repository at this point
Copy the full SHA d179b9dView commit details -
Fix LayoutLMv2 init issue and doctest (huggingface#30278)
* fix * try suggestion * update --------- Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 416fdba - Browse repository at this point
Copy the full SHA 416fdbaView commit details -
Remove old TF port docs (huggingface#30426)
* Remove old TF port guide * repo-consistency * Remove some translations as well for consistency * Remove some translations as well for consistency
Configuration menu - View commit details
-
Copy full SHA for 696eded - Browse repository at this point
Copy the full SHA 696ededView commit details -
Rename torch.run to torchrun (huggingface#30405)
torch.run does not exist anywhere as far as I can tell.
Configuration menu - View commit details
-
Copy full SHA for b8b1e44 - Browse repository at this point
Copy the full SHA b8b1e44View commit details -
Fix use_cache for xla fsdp (huggingface#30353)
* Fix use_cache for xla fsdp * Fix linters
Configuration menu - View commit details
-
Copy full SHA for 12c39e5 - Browse repository at this point
Copy the full SHA 12c39e5View commit details -
[
LlamaTokenizerFast
] Refactor default llama (huggingface#28881)* push legacy to fast as well * super strange * Update src/transformers/convert_slow_tokenizer.py * make sure we are BC * fix Llama test * nit * revert * more test * style * update * small update w.r.t tokenizers * nit * don't split * lol * add a test for `add_prefix_space=False` * fix gemma tokenizer as well * update * fix gemma * nicer failures * fixup * update * fix the example for legacy = False * use `huggyllama/llama-7b` for the PR doctest * nit * use from_slow * fix llama
Configuration menu - View commit details
-
Copy full SHA for e34da3e - Browse repository at this point
Copy the full SHA e34da3eView commit details
Commits on Apr 24, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 0eb8fbc - Browse repository at this point
Copy the full SHA 0eb8fbcView commit details -
Remove add-new-model in favor of add-new-model-like (huggingface#30424)
* Remove add-new-model in favor of add-new-model-like * nits
Configuration menu - View commit details
-
Copy full SHA for d4e92f1 - Browse repository at this point
Copy the full SHA d4e92f1View commit details -
Remove mentions of models in the READMEs and link to the documentatio…
…n page in which they are featured. (huggingface#30420) * REAMDEs * REAMDEs v2
Configuration menu - View commit details
-
Copy full SHA for c6bba94 - Browse repository at this point
Copy the full SHA c6bba94View commit details -
New model PR needs green (slow tests) CI (huggingface#30341)
* You should not pass Co-authored-by: Arthur <[email protected]> --------- Co-authored-by: ydshieh <[email protected]> Co-authored-by: Arthur <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for fc34f84 - Browse repository at this point
Copy the full SHA fc34f84View commit details -
Add llama3 (huggingface#30334)
* nuke * add co-author * add co-author * update card * fixup and fix copies to please our ci * nit fixup * super small nits * remove tokenizer_path from call to `write_model` * always safe serialize by default --------- Co-authored-by: pcuenca <[email protected]> Co-authored-by: xenova <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 89c510d - Browse repository at this point
Copy the full SHA 89c510dView commit details -
Fix YOLOS image processor resizing (huggingface#30436)
* Add test for square image that fails * Fix for square images * Extend test cases * Fix resizing in tests * Style fixup
Configuration menu - View commit details
-
Copy full SHA for 767e351 - Browse repository at this point
Copy the full SHA 767e351View commit details -
[
Llava
] + CIs fix red cis and llava integration tests (huggingface#……30440) * nit * nit and fmt skip * fixup * Update src/transformers/convert_slow_tokenizer.py Co-authored-by: amyeroberts <[email protected]> * set to true --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9a4a119 - Browse repository at this point
Copy the full SHA 9a4a119View commit details -
[tests] make test device-agnostic (huggingface#30444)
* make device-agnostic * clean code
Configuration menu - View commit details
-
Copy full SHA for 16c8e17 - Browse repository at this point
Copy the full SHA 16c8e17View commit details -
fix uncaught init of linear layer in clip's/siglip's for image classi…
…fication models (huggingface#30435) * fix clip's/siglip's _init_weights to reflect linear layers in "for image classification" * trigger slow tests
Configuration menu - View commit details
-
Copy full SHA for 5d64ae9 - Browse repository at this point
Copy the full SHA 5d64ae9View commit details -
fix jamba slow foward for multi-gpu (huggingface#30418)
* fix jamba slow foward for multi-gpu * remove comm * oups * style
Configuration menu - View commit details
-
Copy full SHA for 37fa1f6 - Browse repository at this point
Copy the full SHA 37fa1f6View commit details -
[SegGPT] Fix loss calculation (huggingface#30421)
* Fixed main train issues * Added loss test * Update src/transformers/models/seggpt/modeling_seggpt.py Co-authored-by: amyeroberts <[email protected]> * Added missing labels arg in SegGptModel forward * Fixed typo * Added slow test to test loss calculation --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for d26c141 - Browse repository at this point
Copy the full SHA d26c141View commit details -
Add
paths
filter to avoid the chance of being triggered (huggingfac……e#30453) * trigger * remove the last job --------- Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 42fed15 - Browse repository at this point
Copy the full SHA 42fed15View commit details -
* chore(root): Initial commit of Phi-3 files. * fix(root): Fixes Phi-3 missing on readme. * fix(root): Ensures files are consistent. * fix(phi3): Fixes unit tests. * fix(tests): Fixes style of phi-3 test file. * chore(tests): Adds integration tests for Phi-3. * fix(phi3): Removes additional flash-attention usage, .e.g, swiglu and rmsnorm. * fix(phi3): Fixes incorrect docstrings. * fix(phi3): Fixes docstring typos. * fix(phi3): Adds support for Su and Yarn embeddings. * fix(phi3): Improves according first batch of reviews. * fix(phi3): Uses up_states instead of y in Phi3MLP. * fix(phi3): Uses gemma rotary embedding to support torch.compile. * fix(phi3): Improves how rotary embedding classes are defined. * fix(phi3): Fixes inv_freq not being re-computed for extended RoPE. * fix(phi3): Adds last suggestions to modeling file. * fix(phi3): Splits inv_freq calculation in two lines.
Configuration menu - View commit details
-
Copy full SHA for c9693db - Browse repository at this point
Copy the full SHA c9693dbView commit details -
Fix wrong indent in
utils/check_if_new_model_added.py
(huggingface#……30456) fix Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for d0d430f - Browse repository at this point
Copy the full SHA d0d430fView commit details -
[
research_project
] Most of the security issues come from this requi……rement.txt (huggingface#29977) update most of decision transformers research project
Configuration menu - View commit details
-
Copy full SHA for 661190b - Browse repository at this point
Copy the full SHA 661190bView commit details -
Neuron: When save_safetensor=False, no need to move model to CPU (hug…
…gingface#29703) save_safetensor=True is default as of release 4.35.0, which then required TPU hotfix huggingface#27799 (issue huggingface#27578). However, when the flag save_safetensor is set to False (compatibility mode), moving the model to CPU causes generation of too many graphs during checkpoint huggingface#28438. This PR disable moving of model to CPU when save_safetensor=False.
Configuration menu - View commit details
-
Copy full SHA for d1d94d7 - Browse repository at this point
Copy the full SHA d1d94d7View commit details -
Enable fp16 on CPU (huggingface#30459)
* Check removing flag for torch * LLM oops * Getting there... * More discoveries * Change * Clean up and prettify * Logic check * Not
Configuration menu - View commit details
-
Copy full SHA for 5c57463 - Browse repository at this point
Copy the full SHA 5c57463View commit details -
Non blocking support to torch DL's (huggingface#30465)
* Non blocking support * Check for optimization * Doc
Configuration menu - View commit details
-
Copy full SHA for 6ad9c8f - Browse repository at this point
Copy the full SHA 6ad9c8fView commit details -
consistent job / pytest report / artifact name correspondence (huggin…
…gface#30392) * better names * run better names * update * update --------- Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for fbb41cd - Browse repository at this point
Copy the full SHA fbb41cdView commit details
Commits on Apr 25, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 2a0c9b0 - Browse repository at this point
Copy the full SHA 2a0c9b0View commit details -
Workflow / ENH: Add SSH into our runners workflow (huggingface#30425)
* add SSH into our runners workflow * fix * fix * fix * use our previous approaches * forward contrib credits from discussions --------- Co-authored-by: Yih-Dar <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for cebb072 - Browse repository at this point
Copy the full SHA cebb072View commit details -
FIX / Workflow: Change tailscale trigger condition (huggingface#30471)
Update push-important-models.yml
Configuration menu - View commit details
-
Copy full SHA for cd0cd12 - Browse repository at this point
Copy the full SHA cd0cd12View commit details -
FIX / Workflow: Fix SSH workflow bug (huggingface#30474)
Update ssh-runner.yml
Configuration menu - View commit details
-
Copy full SHA for e9b1635 - Browse repository at this point
Copy the full SHA e9b1635View commit details -
[fix codellama conversion] (huggingface#30472)
* fix codellama conversion * nit
Configuration menu - View commit details
-
Copy full SHA for c60749d - Browse repository at this point
Copy the full SHA c60749dView commit details -
Script for finding candidate models for deprecation (huggingface#29686)
* Add utility for finding candidate models for deprecation * Better model filtering * Update * Add warning tip * Fix up * Review comments * Filter requests based on tags * Add copyright header
Configuration menu - View commit details
-
Copy full SHA for 30ee508 - Browse repository at this point
Copy the full SHA 30ee508View commit details -
Configuration menu - View commit details
-
Copy full SHA for fac7bb3 - Browse repository at this point
Copy the full SHA fac7bb3View commit details -
Fix SigLip classification doctest (huggingface#30475)
* Fix SigLip classification doctest * Remove extra line * Update src/transformers/models/siglip/modeling_siglip.py
Configuration menu - View commit details
-
Copy full SHA for 4fed29e - Browse repository at this point
Copy the full SHA 4fed29eView commit details -
Configuration menu - View commit details
-
Copy full SHA for aca4a10 - Browse repository at this point
Copy the full SHA aca4a10View commit details -
Prevent crash with
WandbCallback
with third parties (huggingface#30477Configuration menu - View commit details
-
Copy full SHA for ce5ae5a - Browse repository at this point
Copy the full SHA ce5ae5aView commit details -
🚨 Add training compatibility for Musicgen-like models (huggingface#29802
) * first modeling code * make repository * still WIP * update model * add tests * add latest change * clean docstrings and copied from * update docstrings md and readme * correct chroma function * correct copied from and remove unreleated test * add doc to toctree * correct imports * add convert script to notdoctested * Add suggestion from Sanchit Co-authored-by: Sanchit Gandhi <[email protected]> * correct get_uncoditional_inputs docstrings * modify README according to SANCHIT feedback * add chroma to audio utils * clean librosa and torchaudio hard dependencies * fix FE * refactor audio decoder -> audio encoder for consistency with previous musicgen * refactor conditional -> encoder * modify sampling rate logics * modify license at the beginning * refactor all_self_attns->all_attentions * remove ignore copy from causallm generate * add copied from for from_sub_models * fix make copies * add warning if audio is truncated * add copied from where relevant * remove artefact * fix convert script * fix torchaudio and FE * modify chroma method according to feedback-> better naming * refactor input_values->input_features * refactor input_values->input_features and fix import fe * add input_features to docstrigs * correct inputs_embeds logics * remove dtype conversion * refactor _prepare_conditional_hidden_states_kwargs_for_generation ->_prepare_encoder_hidden_states_kwargs_for_generation * change warning for chroma length * Update src/transformers/models/musicgen_melody/convert_musicgen_melody_transformers.py Co-authored-by: Sanchit Gandhi <[email protected]> * change way to save wav, using soundfile * correct docs and change to soundfile * fix import * fix init proj layers * add draft training * fix cross entropy * clean loss computation * fix labels * remove line breaks from md * fix issue with docstrings * add FE suggestions * improve is in logics and remove useless imports * remove custom from_pretrained * simplify docstring code * add suggestions for modeling tests * make style * update converting script with sanity check * remove encoder attention mask from conditional generation * replace musicgen melody checkpoints with official orga * rename ylacombe->facebook in checkpoints * fix copies * remove unecessary warning * add shape in code docstrings * add files to slow doc tests * fix md bug and add md to not_tested * make fix-copies * fix hidden states test and batching * update training code * add training tests for melody * add training for o.g musicgen * fix copied from * remove final todos * make style * fix style * add suggestions from review * add ref to the original loss computation code * rename method + fix labels in tests * make style --------- Co-authored-by: Sanchit Gandhi <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 90cb55b - Browse repository at this point
Copy the full SHA 90cb55bView commit details -
Add WSD scheduler (huggingface#30231)
* Added WSD scheduler. * Added tests. * Fixed errors. * Fix formatting. * CI fixes.
Configuration menu - View commit details
-
Copy full SHA for 7b1170b - Browse repository at this point
Copy the full SHA 7b1170bView commit details -
Fix Issue huggingface#29817 Video Classification Task Guide Using Un…
…declared Variables (huggingface#30457) * Fix issue huggingface#29817 Video Classification Task Guide Using Undeclared Variables * Update docs/source/en/tasks/video_classification.md updated with review comments Co-authored-by: amyeroberts <[email protected]> * Fix issue huggingface#29817 Add line space following PR comments --------- Co-authored-by: manju-rangam <Manju1@Git> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9283313 - Browse repository at this point
Copy the full SHA 9283313View commit details -
Make accelerate install non-torch dependent (huggingface#30463)
* Pin accelerate w/o eager * Eager * Update .circleci/create_circleci_config.py Co-authored-by: amyeroberts <[email protected]> * Expound * Expound squared * PyTorch -> dependency --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 86f2569 - Browse repository at this point
Copy the full SHA 86f2569View commit details -
Introduce Stateful Callbacks (huggingface#29666)
* Introduce saveable callbacks * Add note * Test for non-present and flag * Support early stopping and refusing to train further * Update docstring * More saving * Import oopsie * Apply suggestions from code review Co-authored-by: amyeroberts <[email protected]> * Make it go through TrainerArguments * Document * Fix test * Apply suggestions from code review Co-authored-by: amyeroberts <[email protected]> * Rework to allow for duplicates * CLean * Fix failing tests --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for ad697f1 - Browse repository at this point
Copy the full SHA ad697f1View commit details -
Configuration menu - View commit details
-
Copy full SHA for e60491a - Browse repository at this point
Copy the full SHA e60491aView commit details -
Do not use deprecated
SourceFileLoader.load_module()
in dynamic mod……ule loading (huggingface#30370)
Configuration menu - View commit details
-
Copy full SHA for bc274a2 - Browse repository at this point
Copy the full SHA bc274a2View commit details -
Add sidebar tutorial for chat models (huggingface#30401)
* Draft tutorial for talking to chat models * Reformat lists and text snippets * Cleanups and clarifications * Finish up remaining TODOs * Correct section link * Small fix * Add proper quantization examples * Add proper quantization examples * Add proper quantization examples * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/_toctree.yml Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Fix Text Generation Pipeline link and add a ref to the LLM inference guide * intelligent -> capable * Small intro cleanup * Small text cleanup * Small text cleanup * Clarification about system message * Clarification about system message --------- Co-authored-by: Steven Liu <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for f396271 - Browse repository at this point
Copy the full SHA f396271View commit details -
Quantization:
HfQuantizer
quant method update (huggingface#30484)ensure popular quant methods are supported
Configuration menu - View commit details
-
Copy full SHA for 26ddc58 - Browse repository at this point
Copy the full SHA 26ddc58View commit details -
[docs] Spanish translation of pipeline_tutorial.md (huggingface#30252)
* add pipeline_webserver to es/ * add pipeline_webserver to es/, translate first section * add comment for checking link * translate pipeline_webserver * edit pipeline_webserver * fix typo
Configuration menu - View commit details
-
Copy full SHA for a98c417 - Browse repository at this point
Copy the full SHA a98c417View commit details
Commits on Apr 26, 2024
-
FEAT: PEFT support for EETQ (huggingface#30449)
Update quantizer_eetq.py
Configuration menu - View commit details
-
Copy full SHA for 19cfdf0 - Browse repository at this point
Copy the full SHA 19cfdf0View commit details -
Fix the
bitsandbytes
error formatting ("Some modules are dispatched…… on ...") (huggingface#30494) Fix the `bitsandbytes` error when some modules are not properly offloaded.
Configuration menu - View commit details
-
Copy full SHA for 59e715f - Browse repository at this point
Copy the full SHA 59e715fView commit details -
Update
dtype_byte_size
to handle torch.float8_e4m3fn/float8_e5m2 ty……pes (huggingface#30488) * Update modeling_utils/dtype_byte_size to handle float8 types * Add a test for dtype_byte_size * Format * Fix bool
Configuration menu - View commit details
-
Copy full SHA for 20081c7 - Browse repository at this point
Copy the full SHA 20081c7View commit details -
Use the Keras set_random_seed in tests (huggingface#30504)
Use the Keras set_random_seed to ensure reproducible weight initialization
Configuration menu - View commit details
-
Copy full SHA for 2de5cb1 - Browse repository at this point
Copy the full SHA 2de5cb1View commit details -
[
BERT
] Add support for sdpa (huggingface#28802)* Adding SDPA support for BERT * Using the proper input name for testing model input in inference() * Adding documentation for SDPA in BERT model page * Use the stable link for the documentation * Adding a gate to only call .contiguous() for torch < 2.2.0 * Additions and fixes to the documentation * Minor updates to documentation * Adding extra requirements needed for the contiguous() bug * Adding "Adapted from" in plcae of the "Copied from" * Add benchmark speedup tables to the documentation * Minor fixes to the documentation * Use ClapText as a replacemenet for Bert in the Copied-From * Some more fixes for the fix-copies references * Overriding the test_eager_matches_sdpa_generate in bert tests to not load with low_cpu_mem_usage [test all] * Undo changes to separate test * Refactored SDPA self attention code for KV projections * Change use_sdpa to attn_implementation * Fix test_sdpa_can_dispatch_on_flash by preparing input (required for MultipleChoice models)
Configuration menu - View commit details
-
Copy full SHA for dfa7b58 - Browse repository at this point
Copy the full SHA dfa7b58View commit details -
Remove skipping logic now that set_epoch exists (huggingface#30501)
* Remove skipping logic now that set_epoch exists * Working version, clean
Configuration menu - View commit details
-
Copy full SHA for 77ff304 - Browse repository at this point
Copy the full SHA 77ff304View commit details -
[
DETR
] Remove timm hardcoded logic in modeling files (huggingface#2……9038) * Enable instantiating model with pretrained backbone weights * Clarify pretrained import * Use load_backbone instead * Add backbone_kwargs to config * Fix up * Add tests * Tidy up * Enable instantiating model with pretrained backbone weights * Update tests so backbone checkpoint isn't passed in * Clarify pretrained import * Update configs - docs and validation check * Update src/transformers/utils/backbone_utils.py Co-authored-by: Arthur <[email protected]> * Clarify exception message * Update config init in tests * Add test for when use_timm_backbone=True * Use load_backbone instead * Add use_timm_backbone to the model configs * Add backbone_kwargs to config * Pass kwargs to constructors * Draft * Fix tests * Add back timm - weight naming * More tidying up * Whoops * Tidy up * Handle when kwargs are none * Update tests * Revert test changes * Deformable detr test - don't use default * Don't mutate; correct model attributes * Add some clarifying comments * nit - grammar is hard --------- Co-authored-by: Arthur <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for aafa7ce - Browse repository at this point
Copy the full SHA aafa7ceView commit details -
[examples] update whisper fine-tuning (huggingface#29938)
* [examples] update whisper fine-tuning * deprecate forced/suppress tokens * item assignment * update readme * final fix
Configuration menu - View commit details
-
Copy full SHA for 38b53da - Browse repository at this point
Copy the full SHA 38b53daView commit details -
Fix GroundingDINO, DPR after BERT SDPA update (huggingface#30506)
Fix GroundingDINO, DPR after BET SDPA update
Configuration menu - View commit details
-
Copy full SHA for e7d52a1 - Browse repository at this point
Copy the full SHA e7d52a1View commit details -
load_image - decode b64encode and encodebytes strings (huggingface#30192
Configuration menu - View commit details
-
Copy full SHA for c793b26 - Browse repository at this point
Copy the full SHA c793b26View commit details -
[SegGPT] Fix seggpt image processor (huggingface#29550)
* Fixed SegGptImageProcessor to handle 2D and 3D prompt mask inputs * Added new test to check prompt mask equivalence * New proposal * Better proposal * Removed unnecessary method * Updated seggpt docs * Introduced do_convert_rgb * nits
Configuration menu - View commit details
-
Copy full SHA for 6d4cabd - Browse repository at this point
Copy the full SHA 6d4cabdView commit details -
Configuration menu - View commit details
-
Copy full SHA for 73014b5 - Browse repository at this point
Copy the full SHA 73014b5View commit details
Commits on Apr 29, 2024
-
Allow boolean FSDP options in fsdp_config (huggingface#30439)
* Allow boolean FSDP options in fsdp_config * Use lower() to be safe
Configuration menu - View commit details
-
Copy full SHA for 80126f9 - Browse repository at this point
Copy the full SHA 80126f9View commit details -
Pass attn_implementation when using AutoXXX.from_config (huggingface#…
…30507) * Pass attn_implementation when using AutoXXX.from_config * Fix
Configuration menu - View commit details
-
Copy full SHA for e8acb70 - Browse repository at this point
Copy the full SHA e8acb70View commit details -
Fix broken link to Transformers notebooks (huggingface#30512)
Co-authored-by: Clint Adams <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for bdbe166 - Browse repository at this point
Copy the full SHA bdbe166View commit details -
Update runner tag for PR slow CI (huggingface#30535)
fix Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for c024218 - Browse repository at this point
Copy the full SHA c024218View commit details -
Fix repo. fetch/checkout in PR slow CI job (huggingface#30537)
fix Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 87be06c - Browse repository at this point
Copy the full SHA 87be06cView commit details -
Reenable SDPA's FA2 During Training with torch.compile (huggingface#3…
…0442) * Reenable SDPA's FA2 during training with torch.compile * fix Olmo's SDPA FA2 dispatching too * update formatting * improved SDPA comment * formatting and explanatory comment * is_causal if statement to one-liner
Configuration menu - View commit details
-
Copy full SHA for 9df8b30 - Browse repository at this point
Copy the full SHA 9df8b30View commit details -
Include safetensors as part of
_load_best_model
(huggingface#30553)* Include safetensors * Cleanup
Configuration menu - View commit details
-
Copy full SHA for a3aabc7 - Browse repository at this point
Copy the full SHA a3aabc7View commit details
Commits on Apr 30, 2024
-
Pass
use_cache
in kwargs for GPTNeoX (huggingface#30538)pass use_cache in kwargs
Configuration menu - View commit details
-
Copy full SHA for c712d05 - Browse repository at this point
Copy the full SHA c712d05View commit details -
Enable multi-device for more models (huggingface#30409)
* feat: support for dinov2 * feat: support for depth_anything * feat: support for efficientformer * feat: support for bert (is this right?) * update: embedding split * remove: empty string * feat: support for align * fix: copies * fix: QAQBertEmbeddings * fix: more consistency issues * revert: support for effientformer * feat: support for altclip * feat: support for blip_text * support for ChineseCLIP * feat: support for depth anything * feat: support for dpt * feat: support for dpt * feat: support for git * feat: support for groupvit * update: format * fix: support for clip * fix: consistency * feat: support for pvt * feat: support for vit_msn * fix: consistency * fix: other copies * remove: device transfer * revert: in-place add * update: support for align * update: support for bert * update: support for Chinese CLIP * revert: changes to efficientformer * update: support for dpt * update: support for efficientformer * revert: changes to git * revert: changes to groupvit * revert: changes to roc_bert * update: support for vit_msn * revert: changes to dpt * remove: extra space * style: extra space
Configuration menu - View commit details
-
Copy full SHA for 0ae789e - Browse repository at this point
Copy the full SHA 0ae789eView commit details -
Configuration menu - View commit details
-
Copy full SHA for 45983db - Browse repository at this point
Copy the full SHA 45983dbView commit details -
Configuration menu - View commit details
-
Copy full SHA for e6c6f6c - Browse repository at this point
Copy the full SHA e6c6f6cView commit details -
Configuration menu - View commit details
-
Copy full SHA for a6cd08c - Browse repository at this point
Copy the full SHA a6cd08cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 6611877 - Browse repository at this point
Copy the full SHA 6611877View commit details -
Update src/transformers/models/auto/tokenization_auto.py
Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 455163b - Browse repository at this point
Copy the full SHA 455163bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 42d5815 - Browse repository at this point
Copy the full SHA 42d5815View commit details -
Update src/transformers/modeling_utils.py
Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 1d3acec - Browse repository at this point
Copy the full SHA 1d3acecView commit details -
Configuration menu - View commit details
-
Copy full SHA for af3c42c - Browse repository at this point
Copy the full SHA af3c42cView commit details -
Configuration menu - View commit details
-
Copy full SHA for a27db0c - Browse repository at this point
Copy the full SHA a27db0cView commit details -
Apply suggestions from code review
Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 14ad10c - Browse repository at this point
Copy the full SHA 14ad10cView commit details -
Update src/transformers/modeling_gguf_pytorch_utils.py
Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for ab621a7 - Browse repository at this point
Copy the full SHA ab621a7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 207820a - Browse repository at this point
Copy the full SHA 207820aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 1fef8ad - Browse repository at this point
Copy the full SHA 1fef8adView commit details -
Configuration menu - View commit details
-
Copy full SHA for 9ae7363 - Browse repository at this point
Copy the full SHA 9ae7363View commit details -
Configuration menu - View commit details
-
Copy full SHA for 75bbfd5 - Browse repository at this point
Copy the full SHA 75bbfd5View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1bff6a0 - Browse repository at this point
Copy the full SHA 1bff6a0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 78a57c5 - Browse repository at this point
Copy the full SHA 78a57c5View commit details -
Fix seq2seq collator padding (huggingface#30556)
* fix seq2seq data collator to respect the given padding strategy further added tests for the seq2seq data collator in the style of the `data_collator_for_token_classification` (pt, tf, np) * formatting and change bool equals "==" to "is" * add missed return types in tests * update numpy test as it can handle unequal shapes, not like pt or tf
Configuration menu - View commit details
-
Copy full SHA for 9112520 - Browse repository at this point
Copy the full SHA 9112520View commit details -
BlipModel: get_multimodal_features method (huggingface#30438)
* add_blip_get_multimodal_feautres * Fix docstring error * reimplement get_multimodal_features * fix error * recheck code quality * add new necessary tests
Configuration menu - View commit details
-
Copy full SHA for 0cdb6b3 - Browse repository at this point
Copy the full SHA 0cdb6b3View commit details -
Add chat templating support for KeyDataset in text-generation pipeline (
huggingface#30558) * added chat templating support for keydataset in generation pipeline * fixed and improved test * fix formatting test failures * Fix tests * Fix tests
Configuration menu - View commit details
-
Copy full SHA for 2ecefc3 - Browse repository at this point
Copy the full SHA 2ecefc3View commit details -
Fix generation doctests (huggingface#30263)
* fix doctest * fix torch doctest * make CI happy * raise error * make fixup
Configuration menu - View commit details
-
Copy full SHA for b8ac4d0 - Browse repository at this point
Copy the full SHA b8ac4d0View commit details -
General PR slow CI (huggingface#30540)
* More general PR slow CI * Update utils/pr_slow_ci_models.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: ydshieh <[email protected]> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 87927b2 - Browse repository at this point
Copy the full SHA 87927b2View commit details -
Remove
use_square_size
after loading (huggingface#30567)* fix * add test --------- Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 78fdd64 - Browse repository at this point
Copy the full SHA 78fdd64View commit details
Commits on May 1, 2024
-
Use text config's vocab size in testing models (huggingface#30568)
use text config's vocab size
Configuration menu - View commit details
-
Copy full SHA for 9d31b32 - Browse repository at this point
Copy the full SHA 9d31b32View commit details -
Encoder-decoder models: move embedding scale to nn.Module (huggingfac…
…e#30410) * move scaling to nn.Module * let the test be here for now (need to fix) * failing tests * last failing models * Revert commit 4c14817 * clean-up * oops forgot * codestyle * raise NotImplemented when possible * Update tests/test_modeling_common.py Co-authored-by: amyeroberts <[email protected]> * skip tests in respective modeling files --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 38a4bf7 - Browse repository at this point
Copy the full SHA 38a4bf7View commit details -
Fix Marian model conversion (huggingface#30173)
* fix marian model coversion * uncomment that line * remove unnecessary code * revert tie_weights, doesn't hurt
Configuration menu - View commit details
-
Copy full SHA for 4bc9cb3 - Browse repository at this point
Copy the full SHA 4bc9cb3View commit details -
Refactor default chat template warnings (huggingface#30551)
* Temporarily silence warnings in apply_chat_template until we can properly deprecate default chat templates * make fixup * Move the default chat template warning into apply_chat_template itself * make fixup
Configuration menu - View commit details
-
Copy full SHA for 4b4da18 - Browse repository at this point
Copy the full SHA 4b4da18View commit details -
Fix QA example (huggingface#30580)
* Handle cases when CLS token is absent * Use BOS token as a fallback
Configuration menu - View commit details
-
Copy full SHA for 1e05671 - Browse repository at this point
Copy the full SHA 1e05671View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3c69d81 - Browse repository at this point
Copy the full SHA 3c69d81View commit details -
Fix canonical model --model_type in examples (huggingface#30480)
Fix --model_type in examples
Configuration menu - View commit details
-
Copy full SHA for bbaa8ce - Browse repository at this point
Copy the full SHA bbaa8ceView commit details -
Gemma: update activation warning (huggingface#29995)
* Gemma: only display act. warning when necessary This is a nit PR, but I was confused. I got the warning even after I had changed `hidden_act` to `gelu_pytorch_tanh`, telling me that I was using the "legacy" `gelu_pytorch_tanh`. Another option is to keep the warning but change the message to say something like "`hidden_act` is ignored, please use `hidden_activation` instead. Setting Gemma's activation function to `gelu_pytorch_tanh`". * Change message, and set `config.hidden_activation`
Configuration menu - View commit details
-
Copy full SHA for f4f18af - Browse repository at this point
Copy the full SHA f4f18afView commit details -
Bump gitpython from 3.1.32 to 3.1.41 in /examples/research_projects/d…
…ecision_transformer (huggingface#30587) Bump gitpython in /examples/research_projects/decision_transformer Bumps [gitpython](https://github.com/gitpython-developers/GitPython) from 3.1.32 to 3.1.41. - [Release notes](https://github.com/gitpython-developers/GitPython/releases) - [Changelog](https://github.com/gitpython-developers/GitPython/blob/main/CHANGES) - [Commits](gitpython-developers/GitPython@3.1.32...3.1.41) --- updated-dependencies: - dependency-name: gitpython dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for b71f512 - Browse repository at this point
Copy the full SHA b71f512View commit details -
Bump grpcio from 1.44.0 to 1.53.2 in /examples/research_projects/deci…
…sion_transformer (huggingface#30585) Bump grpcio in /examples/research_projects/decision_transformer Bumps [grpcio](https://github.com/grpc/grpc) from 1.44.0 to 1.53.2. - [Release notes](https://github.com/grpc/grpc/releases) - [Changelog](https://github.com/grpc/grpc/blob/master/doc/grpc_release_schedule.md) - [Commits](grpc/grpc@v1.44.0...v1.53.2) --- updated-dependencies: - dependency-name: grpcio dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for ff8f624 - Browse repository at this point
Copy the full SHA ff8f624View commit details -
Bump gitpython from 3.1.32 to 3.1.41 in /examples/research_projects/d…
…istillation (huggingface#30586) Bump gitpython in /examples/research_projects/distillation Bumps [gitpython](https://github.com/gitpython-developers/GitPython) from 3.1.32 to 3.1.41. - [Release notes](https://github.com/gitpython-developers/GitPython/releases) - [Changelog](https://github.com/gitpython-developers/GitPython/blob/main/CHANGES) - [Commits](gitpython-developers/GitPython@3.1.32...3.1.41) --- updated-dependencies: - dependency-name: gitpython dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 7164171 - Browse repository at this point
Copy the full SHA 7164171View commit details -
Bump pyarrow from 7.0.0 to 15.0.0 in /examples/research_projects/deci…
…sion_transformer (huggingface#30582) Bump pyarrow in /examples/research_projects/decision_transformer Bumps [pyarrow](https://github.com/apache/arrow) from 7.0.0 to 15.0.0. - [Commits](apache/arrow@go/v7.0.0...go/v15.0.0) --- updated-dependencies: - dependency-name: pyarrow dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for aefbdfe - Browse repository at this point
Copy the full SHA aefbdfeView commit details -
Bump pyarrow from 1.0.1 to 15.0.0 in /examples/research_projects/visu…
…al_bert (huggingface#30583) Bump pyarrow in /examples/research_projects/visual_bert Bumps [pyarrow](https://github.com/apache/arrow) from 1.0.1 to 15.0.0. - [Commits](apache/arrow@apache-arrow-1.0.1...go/v15.0.0) --- updated-dependencies: - dependency-name: pyarrow dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 0ec3003 - Browse repository at this point
Copy the full SHA 0ec3003View commit details -
Bump pyarrow from 1.0.1 to 15.0.0 in /examples/research_projects/lxme…
…rt (huggingface#30584) Bumps [pyarrow](https://github.com/apache/arrow) from 1.0.1 to 15.0.0. - [Commits](apache/arrow@apache-arrow-1.0.1...go/v15.0.0) --- updated-dependencies: - dependency-name: pyarrow dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for b33f01f - Browse repository at this point
Copy the full SHA b33f01fView commit details -
Bump torch from 1.6.0 to 1.13.1 in /examples/research_projects/lxmert (…
…huggingface#21174) Bumps [torch](https://github.com/pytorch/pytorch) from 1.6.0 to 1.13.1. - [Release notes](https://github.com/pytorch/pytorch/releases) - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md) - [Commits](pytorch/pytorch@v1.6.0...v1.13.1) --- updated-dependencies: - dependency-name: torch dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 7a29c57 - Browse repository at this point
Copy the full SHA 7a29c57View commit details -
Bump torch from 1.11.0 to 1.13.1 in /examples/research_projects/codep…
…arrot (huggingface#21170) Bump torch in /examples/research_projects/codeparrot Bumps [torch](https://github.com/pytorch/pytorch) from 1.11.0 to 1.13.1. - [Release notes](https://github.com/pytorch/pytorch/releases) - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md) - [Commits](pytorch/pytorch@v1.11.0...v1.13.1) --- updated-dependencies: - dependency-name: torch dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for ce66c0e - Browse repository at this point
Copy the full SHA ce66c0eView commit details -
Bump torch from 1.6.0 to 1.13.1 in /examples/research_projects/visual…
…_bert (huggingface#21172) Bump torch in /examples/research_projects/visual_bert Bumps [torch](https://github.com/pytorch/pytorch) from 1.6.0 to 1.13.1. - [Release notes](https://github.com/pytorch/pytorch/releases) - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md) - [Commits](pytorch/pytorch@v1.6.0...v1.13.1) --- updated-dependencies: - dependency-name: torch dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 6e0cba3 - Browse repository at this point
Copy the full SHA 6e0cba3View commit details -
Fix image segmentation example - don't reopen image (huggingface#30481)
Fix image segmentation example - don't repoen image
Configuration menu - View commit details
-
Copy full SHA for d2feb54 - Browse repository at this point
Copy the full SHA d2feb54View commit details -
Improve object detection task guideline (huggingface#29967)
* Add improvements * Address comment
Configuration menu - View commit details
-
Copy full SHA for dc401d3 - Browse repository at this point
Copy the full SHA dc401d3View commit details -
Configuration menu - View commit details
-
Copy full SHA for d57ffb4 - Browse repository at this point
Copy the full SHA d57ffb4View commit details -
Fix llava half precision and autocast issues (huggingface#29721)
* Ensure input_embeds and image_features are the same dtype in autocast * Fix nans in half precision llava-next and fix autocasting behavior. * Fix styling issues. * fix randn newline instantiation * fix broken slow llava test * Fix llava next init. * fix styling issues * [run-slow]llava,llava_next * fix styling issues
Configuration menu - View commit details
-
Copy full SHA for 5090ea3 - Browse repository at this point
Copy the full SHA 5090ea3View commit details -
Bump torch from 1.11.0 to 1.13.1 in /examples/research_projects/decis…
…ion_transformer (huggingface#21171) Bump torch in /examples/research_projects/decision_transformer Bumps [torch](https://github.com/pytorch/pytorch) from 1.11.0 to 1.13.1. - [Release notes](https://github.com/pytorch/pytorch/releases) - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md) - [Commits](pytorch/pytorch@v1.11.0...v1.13.1) --- updated-dependencies: - dependency-name: torch dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 6f465d4 - Browse repository at this point
Copy the full SHA 6f465d4View commit details -
Bump torch from 1.9.0+cpu to 1.13.1 in /examples/research_projects/ja…
…x-projects/hybrid_clip (huggingface#21167) Bump torch in /examples/research_projects/jax-projects/hybrid_clip Bumps [torch](https://github.com/pytorch/pytorch) from 1.9.0+cpu to 1.13.1. - [Release notes](https://github.com/pytorch/pytorch/releases) - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md) - [Commits](https://github.com/pytorch/pytorch/commits/v1.13.1) --- updated-dependencies: - dependency-name: torch dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 4f3c7af - Browse repository at this point
Copy the full SHA 4f3c7afView commit details -
Bump pillow from 10.0.1 to 10.2.0 in /examples/research_projects/deci…
…sion_transformer (huggingface#28655) Bump pillow in /examples/research_projects/decision_transformer Bumps [pillow](https://github.com/python-pillow/Pillow) from 10.0.1 to 10.2.0. - [Release notes](https://github.com/python-pillow/Pillow/releases) - [Changelog](https://github.com/python-pillow/Pillow/blob/main/CHANGES.rst) - [Commits](python-pillow/Pillow@10.0.1...10.2.0) --- updated-dependencies: - dependency-name: pillow dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 3a36597 - Browse repository at this point
Copy the full SHA 3a36597View commit details -
Bump torch from 1.9.0+cpu to 1.13.1 in /examples/flax/vision (hugging…
…face#21168) Bumps [torch](https://github.com/pytorch/pytorch) from 1.9.0+cpu to 1.13.1. - [Release notes](https://github.com/pytorch/pytorch/releases) - [Changelog](https://github.com/pytorch/pytorch/blob/master/RELEASE.md) - [Commits](https://github.com/pytorch/pytorch/commits/v1.13.1) --- updated-dependencies: - dependency-name: torch dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for c681b58 - Browse repository at this point
Copy the full SHA c681b58View commit details
Commits on May 2, 2024
-
Fix: failing CI after huggingface#30568 (huggingface#30599)
* failiing CI * no let's keep it intil full deprecation in v4.42
Configuration menu - View commit details
-
Copy full SHA for 5cf3e6b - Browse repository at this point
Copy the full SHA 5cf3e6bView commit details -
Configuration menu - View commit details
-
Copy full SHA for fbabd67 - Browse repository at this point
Copy the full SHA fbabd67View commit details -
Fix memory leak with CTC training script on Chinese languages (huggin…
…gface#30358) * Fix memory leak with CTC training script on Chinese languages * Fix lint
Configuration menu - View commit details
-
Copy full SHA for 12c5544 - Browse repository at this point
Copy the full SHA 12c5544View commit details -
🚨 Update image_processing_vitmatte.py (huggingface#30566)
* Update image_processing_vitmatte.py * add test * [run-slow]vitmatte
Configuration menu - View commit details
-
Copy full SHA for f953025 - Browse repository at this point
Copy the full SHA f953025View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4ad5ada - Browse repository at this point
Copy the full SHA 4ad5adaView commit details -
fix:missing
output_router_logits
in SwitchTransformers (huggingface……#30573) * fix:missing `output_router_logits` in SwitchTransformers * fix whitespace in blank line
Configuration menu - View commit details
-
Copy full SHA for a65da83 - Browse repository at this point
Copy the full SHA a65da83View commit details -
Use
contiguous()
in clip checkpoint conversion script (huggingface#……30613) * fix * fix --------- Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for f57f014 - Browse repository at this point
Copy the full SHA f57f014View commit details -
phi3 chat_template does not support system role (huggingface#30606)
* phi3 chat_template does not support system role * fix doc test error
Configuration menu - View commit details
-
Copy full SHA for 801894e - Browse repository at this point
Copy the full SHA 801894eView commit details -
Docs: fix
generate
-related rendering issues (huggingface#30600)* does this work? * like this? * fix the other generate links * missing these
Configuration menu - View commit details
-
Copy full SHA for aa55ff4 - Browse repository at this point
Copy the full SHA aa55ff4View commit details -
Docs: add missing
StoppingCriteria
autodocs (huggingface#30617)* add missing docstrings to docs * Update src/transformers/generation/stopping_criteria.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 66abe13 - Browse repository at this point
Copy the full SHA 66abe13View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9719202 - Browse repository at this point
Copy the full SHA 9719202View commit details -
Configuration menu - View commit details
-
Copy full SHA for 39359e5 - Browse repository at this point
Copy the full SHA 39359e5View commit details -
Output
None
as attention when layer is skipped (huggingface#30597)* Output `None` as attention when layer is skipped * Add test for output_attentions
Configuration menu - View commit details
-
Copy full SHA for 4c94093 - Browse repository at this point
Copy the full SHA 4c94093View commit details -
Add HQQ quantization support (huggingface#29637)
* update HQQ transformers integration * push import_utils.py * add force_hooks check in modeling_utils.py * fix | with Optional * force bias as param * check bias is Tensor * force forward for multi-gpu * review fixes pass * remove torch grad() * if any key in linear_tags fix * add cpu/disk check * isinstance return * add multigpu test + refactor tests * clean hqq_utils imports in hqq.py * clean hqq_utils imports in quantizer_hqq.py * delete hqq_utils.py * Delete src/transformers/utils/hqq_utils.py * ruff init * remove torch.float16 from __init__ in test * refactor test * isinstance -> type in quantizer_hqq.py * cpu/disk device_map check in quantizer_hqq.py * remove type(module) nn.linear check in quantizer_hqq.py * add BaseQuantizeConfig import inside HqqConfig init * remove hqq import in hqq.py * remove accelerate import from test_hqq.py * quant config.py doc update * add hqqconfig to main_classes doc * make style * __init__ fix * ruff __init__ * skip_modules list * hqqconfig format fix * hqqconfig doc fix * hqqconfig doc fix * hqqconfig doc fix * hqqconfig doc fix * hqqconfig doc fix * hqqconfig doc fix * hqqconfig doc fix * hqqconfig doc fix * hqqconfig doc fix * test_hqq.py remove mistral comment * remove self.using_multi_gpu is False * torch_dtype default val set and logger.info * hqq.py isinstance fix * remove torch=None * torch_device test_hqq * rename test_hqq * MODEL_ID in test_hqq * quantizer_hqq setattr fix * quantizer_hqq typo fix * imports quantizer_hqq.py * isinstance quantizer_hqq * hqq_layer.bias reformat quantizer_hqq * Step 2 as comment in quantizer_hqq * prepare_for_hqq_linear() comment * keep_in_fp32_modules fix * HqqHfQuantizer reformat * quantization.md hqqconfig * quantization.md model example reformat * quantization.md # space * quantization.md space }) * quantization.md space }) * quantization_config fix doc Co-authored-by: amyeroberts <[email protected]> * axis value check in quantization_config * format * dynamic config explanation * quant config method in quantization.md * remove shard-level progress * .cuda fix modeling_utils * test_hqq fixes * make fix-copies --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 5995299 - Browse repository at this point
Copy the full SHA 5995299View commit details -
Fix CI after huggingface#30410 (huggingface#30612)
* Fix CI after huggingface#30410 * [run-slow] blenderbot
Configuration menu - View commit details
-
Copy full SHA for a0e77a1 - Browse repository at this point
Copy the full SHA a0e77a1View commit details
Commits on May 3, 2024
-
add mlp bias for llama models (huggingface#30031)
* add bias * fix quality
Configuration menu - View commit details
-
Copy full SHA for 425e1a0 - Browse repository at this point
Copy the full SHA 425e1a0View commit details -
Fix W&B run name (huggingface#30462)
* Remove comparison to output_dir * Update docs for `run_name` * Add warning
Configuration menu - View commit details
-
Copy full SHA for 66f675e - Browse repository at this point
Copy the full SHA 66f675eView commit details -
HQQ: PEFT support for HQQ (huggingface#30632)
Update quantizer_hqq.py
Configuration menu - View commit details
-
Copy full SHA for d0c72c1 - Browse repository at this point
Copy the full SHA d0c72c1View commit details -
Prevent
TextGenerationPipeline._sanitize_parameters
from overriding…… previously provided parameters (huggingface#30362) * Fixed TextGenerationPipeline._sanitize_parameters default params * removed empty spaces --------- Co-authored-by: Ng, Yen Ting <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for deb7605 - Browse repository at this point
Copy the full SHA deb7605View commit details -
Avoid duplication in PR slow CI model list (huggingface#30634)
update Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 91d155e - Browse repository at this point
Copy the full SHA 91d155eView commit details
Commits on May 6, 2024
-
[
CI update
] Try to use dockers and no cache (huggingface#29202)* change cis * nits * update * minor updates * [push-ci-image] * nit [push-ci-image] * nitsssss * [build-ci-image] * [push-ci-image] * [push-ci-image] * both * [push-ci-image] * this? * [push-ci-image] * pypi-kenlm needs g++ * [push-ci-image] * nit * more nits [push-ci-image] * nits [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * add vision * [push-ci-image] * [push-ci-image] * add new dummy file but will need to update them [push-ci-image] * [push-ci-image] * show package size as well * [push-ci-image] * potentially ignore failures * workflow updates * nits [push-ci-image] * [push-ci-image] * fix consistency * clean nciida triton * also show big packages [push-ci-image] * nit * update * another one * line escape? * add accelerate [push-ci-image] * updates [push-ci-image] * nits to run tests, no push-ci * try to parse skip reason to make sure nothing is skipped that should no be skippped * nit? * always show skipped reasons * nits * better parsing of the test outputs * action="store_true", * failure on failed * show matched * debug * update short summary with skipped, failed and errors * nits * nits * coolu pdates * remove docbuilder * fix * always run checks * oups * nits * don't error out on library printing * non zero exi codes * no warning * nit * WAT? * format nit * [push-ci-image] * fail if fail is needed * [push-ci-image] * sound file for torch light? * [push-ci-image] * order is important [push-ci-image] * [push-ci-image] reduce even further * [push-ci-image] * use pytest rich ! * yes [push-ci-image] * oupsy * bring back the full traceback, but pytest rich should help * nit * [push-ci-image] * re run * nit * [push-ci-image] * [push-ci-image] * [push-ci-image] * empty push to trigger * [push-ci-image] * nit? [push-ci-image] * empty * try to install timm with no deps * [push-ci-image] * oups [push-ci-image] * [push-ci-image] * [push-ci-image] ? * [push-ci-image] open ssh client for git checkout fast * empty for torch light * updates [push-ci-image] * nit * @v4 for checkout * [push-ci-image] * [push-ci-image] * fix fetch tests with parallelism * [push-ci-image] * more parallelism * nit * more nits * empty to re-trigger * empty to re-trigger * split by timing * did not work with previous commit * junit.xml * no path? * mmm this? * junitxml format * split by timing * nit * fix junit family * now we can test if the xunit1 is compatible! * this? * fully list tests * update * update * oups * finally * use classname * remove working directory to make sure the path does not interfere * okay no juni should have the correct path * name split? * sort by classname is what make most sense * some testing * naem * oups * test something fun * autodetect * 18? * nit * file size? * uip * 4 is best * update to see versions * better print * [push-ci-image] * [push-ci-image] * please install the correct keras version * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * uv is fucking me up * [push-ci-image] * [push-ci-image] * [push-ci-image] * nits * [push-ci-image] * [push-ci-image] * install issues an pins * tapas as well * nits * more paralellism * short tb * soundfile * soundfile * [push-ci-image] * [push-ci-image] * [push-ci-image] * oups * [push-ci-image] * fix some things * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * use torch-light for hub * small git lfs for hub job * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * fix tf tapas * [push-ci-image] * nits * [push-ci-image] * don't update the test * [push-ci-image] * [push-ci-image] * [push-ci-image] * no use them * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * update tf proba * [push-ci-image] * [push-ci-image] * woops * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * test with built dockers * [push-ci-image] * skip annoying tests * revert fix copy * update test values * update * last skip and fixup * nit * ALL GOOOD * quality * Update tests/models/layoutlmv2/test_image_processing_layoutlmv2.py * Update docker/quality.dockerfile Co-authored-by: Lysandre Debut <[email protected]> * Update src/transformers/models/tapas/modeling_tf_tapas.py Co-authored-by: Lysandre Debut <[email protected]> * Apply suggestions from code review Co-authored-by: Lysandre Debut <[email protected]> * use torch-speed * updates * [push-ci-image] * [push-ci-image] * [push-ci-image] * [push-ci-image] * fuck ken-lm [push-ci-image] * [push-ci-image] * [push-ci-image] --------- Co-authored-by: Lysandre Debut <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 307f632 - Browse repository at this point
Copy the full SHA 307f632View commit details -
Configuration menu - View commit details
-
Copy full SHA for 09edd77 - Browse repository at this point
Copy the full SHA 09edd77View commit details -
Hotfix-change-ci (huggingface#30669)
* dmmy change * fiux * revert change
Configuration menu - View commit details
-
Copy full SHA for a45c514 - Browse repository at this point
Copy the full SHA a45c514View commit details -
Quantization / HQQ: Fix HQQ tests on our runner (huggingface#30668)
Update test_hqq.py
Configuration menu - View commit details
-
Copy full SHA for 9c772ac - Browse repository at this point
Copy the full SHA 9c772acView commit details -
Fix llava next tie_word_embeddings config (huggingface#30640)
* fix llava next embedding * add docstring * Update src/transformers/models/llava_next/configuration_llava_next.py Co-authored-by: NielsRogge <[email protected]> --------- Co-authored-by: NielsRogge <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for aa64f08 - Browse repository at this point
Copy the full SHA aa64f08View commit details -
Trainer._load_from_checkpoint - support loading multiple Peft adapters (
huggingface#30505) * Trainer: load checkpoint model with multiple adapters * Trainer._load_from_checkpoint support multiple active adapters * PeftModel.set_adapter does not support multiple adapters yet * Trainer._load_from_checkpoint test multiple adapters --------- Co-authored-by: Clara Luise Pohland <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for e076953 - Browse repository at this point
Copy the full SHA e076953View commit details -
Trainer - add cache clearing and the option for batched eval metrics …
…computation (huggingface#28769) * Added cache clearing for GPU efficiency. * Added cache clearing for GPU efficiency. * Added batch_eval_metrics capability * Ran make fixup * Fixed bug * Fixed whitespace issue * Fixed outdated condition * Updated docstrings with instructions for batch_eval_metrics. Updated end of dataloader logic * Added first version of batch_eval_metrics Trainer test * Fixed batch_eval_metrics Trainer tests for both eval and predict * Fixed batch_eval_metrics behavior for new Trainer variables * Fixed batch_eval_metrics Trainer tests * Ran fixup
Configuration menu - View commit details
-
Copy full SHA for df475bf - Browse repository at this point
Copy the full SHA df475bfView commit details -
Configuration menu - View commit details
-
Copy full SHA for 277db23 - Browse repository at this point
Copy the full SHA 277db23View commit details -
Respect
resume_download
deprecation (huggingface#30620)* Deprecate resume_download * remove default resume_download value --------- Co-authored-by: Lysandre Debut <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 835de4c - Browse repository at this point
Copy the full SHA 835de4cView commit details
Commits on May 7, 2024
-
top-k instead of top-p in MixtralConfig docstring (huggingface#30687)
top-k instead of top-p in docstring
Configuration menu - View commit details
-
Copy full SHA for 4980d62 - Browse repository at this point
Copy the full SHA 4980d62View commit details -
Bump jinja2 from 3.1.3 to 3.1.4 in /examples/research_projects/decisi…
…on_transformer (huggingface#30680) Bump jinja2 in /examples/research_projects/decision_transformer Bumps [jinja2](https://github.com/pallets/jinja) from 3.1.3 to 3.1.4. - [Release notes](https://github.com/pallets/jinja/releases) - [Changelog](https://github.com/pallets/jinja/blob/main/CHANGES.rst) - [Commits](pallets/jinja@3.1.3...3.1.4) --- updated-dependencies: - dependency-name: jinja2 dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for a898fb9 - Browse repository at this point
Copy the full SHA a898fb9View commit details -
Bump werkzeug from 3.0.1 to 3.0.3 in /examples/research_projects/deci…
…sion_transformer (huggingface#30679) Bump werkzeug in /examples/research_projects/decision_transformer Bumps [werkzeug](https://github.com/pallets/werkzeug) from 3.0.1 to 3.0.3. - [Release notes](https://github.com/pallets/werkzeug/releases) - [Changelog](https://github.com/pallets/werkzeug/blob/main/CHANGES.rst) - [Commits](pallets/werkzeug@3.0.1...3.0.3) --- updated-dependencies: - dependency-name: werkzeug dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for ce47582 - Browse repository at this point
Copy the full SHA ce47582View commit details -
Adding _tie_weights() to prediction heads to support low_cpu_mem_usag…
…e=True (huggingface#29024) * Adding _tie_weights() to prediction heads to support low_cpu_mem_usage=True * Testing for the non-safe-tensors case, since the default is safe-tensors already * Running fixup/fix-copies * Adding accelerate annotations to tests
Configuration menu - View commit details
-
Copy full SHA for 54a2361 - Browse repository at this point
Copy the full SHA 54a2361View commit details -
Fix
cache_position
initialisation for generation with `use_cache=Fa……lse` (huggingface#30485) * Fix cache_position init for generation * Update src/transformers/generation/utils.py Co-authored-by: Arthur <[email protected]> * Fix cache position update --------- Co-authored-by: Arthur <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 4fda78c - Browse repository at this point
Copy the full SHA 4fda78cView commit details -
Word-level timestamps broken for short-form audio (huggingface#30325)
* force chunk_length_s in AutomaticSpeechRecognitionPipeline * compute num_frames even when stride is None * add slow tests * fix test * Update src/transformers/pipelines/automatic_speech_recognition.py Co-authored-by: amyeroberts <[email protected]> * Update tests/pipelines/test_pipelines_automatic_speech_recognition.py Co-authored-by: amyeroberts <[email protected]> * add input validation * fixup * small fix --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9c8979e - Browse repository at this point
Copy the full SHA 9c8979eView commit details -
Updated docs of
forward
inIdefics2ForConditionalGeneration
with ……correct `ignore_index` value (huggingface#30678) updated docs of `forward` in `Idefics2ForConditionalGeneration` with correct `ignore_index` value
Configuration menu - View commit details
-
Copy full SHA for e5f71ec - Browse repository at this point
Copy the full SHA e5f71ecView commit details -
Bump tqdm from 4.63.0 to 4.66.3 in /examples/research_projects/decisi…
…on_transformer (huggingface#30646) Bump tqdm in /examples/research_projects/decision_transformer Bumps [tqdm](https://github.com/tqdm/tqdm) from 4.63.0 to 4.66.3. - [Release notes](https://github.com/tqdm/tqdm/releases) - [Commits](tqdm/tqdm@v4.63.0...v4.66.3) --- updated-dependencies: - dependency-name: tqdm dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 4051d36 - Browse repository at this point
Copy the full SHA 4051d36View commit details -
Bump tqdm from 4.48.2 to 4.66.3 in /examples/research_projects/visual…
…_bert (huggingface#30645) Bump tqdm in /examples/research_projects/visual_bert Bumps [tqdm](https://github.com/tqdm/tqdm) from 4.48.2 to 4.66.3. - [Release notes](https://github.com/tqdm/tqdm/releases) - [Commits](tqdm/tqdm@v4.48.2...v4.66.3) --- updated-dependencies: - dependency-name: tqdm dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 3733391 - Browse repository at this point
Copy the full SHA 3733391View commit details -
Reboot Agents (huggingface#30387)
* Create CodeAgent and ReactAgent * Fix formatting errors * Update documentation for agents * Add custom errors, improve logging * Support variable usage in ReactAgent * add messages * Add message passing format * Create React Code Agent * Update * Refactoring * Fix errors * Improve python interpreter * Only non-tensor inputs should be sent to device * Calculator tool slight refactor * Improve docstrings * Refactor * Fix tests * Fix more tests * Fix even more tests * Fix tests by replacing output and input types * Fix operand type issue * two small fixes * EM TTS * Fix agent running type errors * Change text to speech tests to allow changed outputs * Update doc with new agent types * Improve code interpreter * If max iterations reached, provide a real answer instead of an error * Add edge case in interpreter * Add safe imports to the interpreter * Interpreter tweaks: tuples and listcomp * Make style * Make quality * Add dictcomp to interpreter * Rename ReactJSONAgent to ReactJsonAgent * Misc changes * ToolCollection * Rename agent's logger to self.logger * Add while loops to interpreter * Update doc with new tools. still need to mention collections * Add collections to the doc * Small fixes on logs and interpretor * Fix toolbox return type * Docs + fixup * Skip doctests * Correct prompts with improved examples and formatting * Update prompt * Remove outdated docs * Change agent to accept Toolbox object for tools * Remove calculator tool * Propagate removal of calculator in doc * Fix 2 failing workflows * Simplify additional argument passing * AgentType audio * Minor changes: function name, types * Remove calculator tests * Fix test * Fix torch requirement * Fix final answer tests * Style fixes * Fix tests * Update docstrings with calculator removal * Small type hint fixes * Update tests/agents/test_translation.py Co-authored-by: Arthur <[email protected]> * Update tests/agents/test_python_interpreter.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/agents/default_tools.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/agents/tools.py Co-authored-by: Arthur <[email protected]> * Update tests/agents/test_agents.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/models/bert/configuration_bert.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/agents/tools.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/agents/speech_to_text.py Co-authored-by: Arthur <[email protected]> * Update tests/agents/test_speech_to_text.py Co-authored-by: Arthur <[email protected]> * Update tests/agents/test_tools_common.py Co-authored-by: Arthur <[email protected]> * pygments * Answer comments * Cleaning up * Simplifying init for all agents * Improving prompts and making code nicer * Style fixes * Add multiple comparator test in interpreter * Style fixes * Improve BERT example in documentation * Add examples to doc * Fix python interpreter quality * Logging improvements * Change test flag to agents * Quality fix * Add example for HfEngine * Improve conversation example for HfEngine * typo fix * Verify doc * Update docs/source/en/agents.md Co-authored-by: Arthur <[email protected]> * Update src/transformers/agents/agents.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/agents/prompts.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/agents/python_interpreter.py Co-authored-by: Arthur <[email protected]> * Update docs/source/en/agents.md Co-authored-by: Arthur <[email protected]> * Fix style issues * local s2t tool --------- Co-authored-by: Cyril Kondratenko <[email protected]> Co-authored-by: Lysandre <[email protected]> Co-authored-by: Lysandre <[email protected]> Co-authored-by: Arthur <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 0ba15ce - Browse repository at this point
Copy the full SHA 0ba15ceView commit details -
Bump tqdm from 4.48.2 to 4.66.3 in /examples/research_projects/lxmert (…
…huggingface#30644) Bumps [tqdm](https://github.com/tqdm/tqdm) from 4.48.2 to 4.66.3. - [Release notes](https://github.com/tqdm/tqdm/releases) - [Commits](tqdm/tqdm@v4.48.2...v4.66.3) --- updated-dependencies: - dependency-name: tqdm dependency-type: direct:production ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 4a17200 - Browse repository at this point
Copy the full SHA 4a17200View commit details -
Separate tokenizer tests (huggingface#30675)
* nit * better filter * pipeline tests should only be models/xxx not anything else * nit to better see filtering of the files that are passed to test torch * oups
Configuration menu - View commit details
-
Copy full SHA for 4208c42 - Browse repository at this point
Copy the full SHA 4208c42View commit details -
Update
workflow_id
inutils/get_previous_daily_ci.py
(huggingface……#30695) fix Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 05ec950 - Browse repository at this point
Copy the full SHA 05ec950View commit details -
Rename artifact name
prev_ci_results
toci_results
(huggingface#3……0697) * rename * update --------- Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 884e3b1 - Browse repository at this point
Copy the full SHA 884e3b1View commit details -
Add safetensors to model not found error msg for default use_safetens…
…ors value (huggingface#30602) * add safetensors to model not found error for default use_safetensors=None case * format code w/ ruff * fix assert true typo
Configuration menu - View commit details
-
Copy full SHA for cf7bed9 - Browse repository at this point
Copy the full SHA cf7bed9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5b7a225 - Browse repository at this point
Copy the full SHA 5b7a225View commit details
Commits on May 8, 2024
-
Patch CLIP image preprocessor (huggingface#30698)
* patch clip preprocessor * Update image_processing_clip.py * Update src/transformers/models/clip/image_processing_clip.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 508c0bf - Browse repository at this point
Copy the full SHA 508c0bfView commit details -
Add examples for detection models finetuning (huggingface#30422)
* Training script for object detection * Evaluation script for object detection * Training script for object detection with eval loop outside trainer * Trainer DETR finetuning * No trainer DETR finetuning * Eval script * Refine object detection example with trainer * Remove commented code and enable telemetry * No trainer example * Add requirements for object detection examples * Add test for trainer example * Readme draft * Fix uploading to HUB * Readme improvements * Update eval script * Adding tests for object-detection examples * Add object-detection example * Add object-detection resources to docs * Update README with custom dataset instructions * Update year * Replace valid with validation * Update instructions for custom dataset * Remove eval script * Remove use_auth_token * Add copied from and telemetry * Fixup * Update readme * Fix id2label * Fix links in docs * Update examples/pytorch/object-detection/run_object_detection.py Co-authored-by: NielsRogge <[email protected]> * Update examples/pytorch/object-detection/run_object_detection.py Co-authored-by: NielsRogge <[email protected]> * Move description to the top * Fix Trainer example * Update no trainer example * Update albumentations version --------- Co-authored-by: NielsRogge <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 998dbe0 - Browse repository at this point
Copy the full SHA 998dbe0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1872bde - Browse repository at this point
Copy the full SHA 1872bdeView commit details -
Llava: remove dummy labels (huggingface#30706)
remove labels from llavas
Configuration menu - View commit details
-
Copy full SHA for 467164e - Browse repository at this point
Copy the full SHA 467164eView commit details -
Add installation of examples requirements in CI (huggingface#30708)
* Add installation of examples requirements in CI * Update .circleci/create_circleci_config.py Co-authored-by: Arthur <[email protected]> --------- Co-authored-by: Arthur <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for e7a5f45 - Browse repository at this point
Copy the full SHA e7a5f45View commit details -
Update object detection guide (huggingface#30683)
* Object detection guide * Minor update * Minor updates, links * Fix typo * Wording, add albu space * Add missing part * Update docs/source/en/tasks/object_detection.md Co-authored-by: amyeroberts <[email protected]> * Update docs/source/en/tasks/object_detection.md Co-authored-by: NielsRogge <[email protected]> * Update docs/source/en/tasks/object_detection.md Co-authored-by: NielsRogge <[email protected]> * Fix device, add imports for inference --------- Co-authored-by: amyeroberts <[email protected]> Co-authored-by: NielsRogge <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 5962d62 - Browse repository at this point
Copy the full SHA 5962d62View commit details -
Immutability for data collators (huggingface#30603)
* immutability fix for seq2seq as well as immutability tests for the collators * ensure we don't act on none labels and formatting * remove tf/pt in respective tests as they are not required * more type error fixes tf/np * remove todo * apply suggestions from code review * formatting / style
Configuration menu - View commit details
-
Copy full SHA for 71c1985 - Browse repository at this point
Copy the full SHA 71c1985View commit details -
Configuration menu - View commit details
-
Copy full SHA for f26e407 - Browse repository at this point
Copy the full SHA f26e407View commit details
Commits on May 9, 2024
-
Add dynamic resolution input/interpolate position embedding to SigLIP (…
…huggingface#30719) * Add interpolate positional encoding to siglip * Change # of patches for siglip interpolation test * fix formatting * Apply nit suggestions from code review Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 60293bd - Browse repository at this point
Copy the full SHA 60293bdView commit details -
Enable dynamic resolution for vivit (huggingface#30630)
* feat: enable dynamic resolution for vivit * fix: formatting * remove: print statement for testing * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update tests/models/vivit/test_modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update tests/models/vivit/test_modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update tests/models/vivit/test_modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * Update src/transformers/models/vivit/modeling_vivit.py Co-authored-by: amyeroberts <[email protected]> * fix: style check --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 8c5b3c1 - Browse repository at this point
Copy the full SHA 8c5b3c1View commit details -
Removal of deprecated maps (huggingface#30576)
* [test_all] Remove all imports Remove remaining ARCHIVE MAPS Remove remaining PRETRAINED maps * review comments * [test_all] empty commit to trigger tests
Configuration menu - View commit details
-
Copy full SHA for 297b732 - Browse repository at this point
Copy the full SHA 297b732View commit details -
Generate: add
min_p
sampling (huggingface#30639)* min_p * more relaxed test to avoid numerical issues * Update src/transformers/generation/logits_process.py Co-authored-by: menhguin <[email protected]> * Update src/transformers/generation/configuration_utils.py Co-authored-by: menhguin <[email protected]> * docstring clarifications * PR comments * Update tests/generation/test_logits_process.py Co-authored-by: amyeroberts <[email protected]> * make fixup --------- Co-authored-by: menhguin <[email protected]> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for df53c6e - Browse repository at this point
Copy the full SHA df53c6eView commit details -
Fix image post-processing for OWLv2 (huggingface#30686)
* feat: add note about owlv2 * fix: post processing coordinates * remove: workaround document * fix: extra quotes * update: owlv2 docstrings * fix: copies check * feat: add unit test for resize * Update tests/models/owlv2/test_image_processor_owlv2.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 218f441 - Browse repository at this point
Copy the full SHA 218f441View commit details -
KV cache is no longer a model attribute (huggingface#30730)
kv_cache is no longer a model attribute
Configuration menu - View commit details
-
Copy full SHA for 5413b89 - Browse repository at this point
Copy the full SHA 5413b89View commit details -
Generate: consistently handle special tokens as tensors (huggingface#…
…30624) * tmp commit * [test_all] mvp * missing not * [test_all] final test fixes * fix musicgen_melody and rag * [test_all] empty commit * PR comments * Update src/transformers/generation/utils.py Co-authored-by: Arthur <[email protected]> --------- Co-authored-by: Arthur <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 7130a22 - Browse repository at this point
Copy the full SHA 7130a22View commit details -
Update CodeLlama references (huggingface#30218)
* Update CodeLlama references * Update slow_documentation_tests.txt * Update slow_documentation_tests.txt
Configuration menu - View commit details
-
Copy full SHA for c99d88e - Browse repository at this point
Copy the full SHA c99d88eView commit details -
[docs] Update es/pipeline_tutorial.md (huggingface#30684)
* copy en/ contect to es/ * translate first section * translate the doc * fix typos * run make style
Configuration menu - View commit details
-
Copy full SHA for 47735f5 - Browse repository at this point
Copy the full SHA 47735f5View commit details
Commits on May 10, 2024
-
Update llama3.md, fix typo (huggingface#30739)
Update llama3.md fix typo again
Configuration menu - View commit details
-
Copy full SHA for 73fcfb2 - Browse repository at this point
Copy the full SHA 73fcfb2View commit details -
mlp_only_layers is more flexible than decoder_sparse_step (huggingfac…
…e#30552) * force back to commit ba40a21 and fix workflow errors * match the review suggestions * fix ci errors * fix CI * fix ci, format code * fix ci, ruff format * fix ci, ruff format again * Update src/transformers/models/qwen2_moe/configuration_qwen2_moe.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/models/qwen2_moe/configuration_qwen2_moe.py Co-authored-by: Arthur <[email protected]> * Update src/transformers/models/qwen2_moe/configuration_qwen2_moe.py Co-authored-by: Arthur <[email protected]> * solve this warning: Default Argument Value is mutable --------- Co-authored-by: Arthur <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 1c52cb7 - Browse repository at this point
Copy the full SHA 1c52cb7View commit details -
PEFT / Trainer: Make use of
model.active_adapters()
instead of depr……ecated `model.active_adapter` whenever possible (huggingface#30738) * Update trainer.py * Update src/transformers/trainer.py * Update src/transformers/trainer.py * Update src/transformers/trainer.py * style * Update src/transformers/trainer.py * Update src/transformers/trainer.py
Configuration menu - View commit details
-
Copy full SHA for 2d1602a - Browse repository at this point
Copy the full SHA 2d1602aView commit details -
[docs] Update link in es/pipeline_webserver.md (huggingface#30745)
* update link * run make style
Configuration menu - View commit details
-
Copy full SHA for 8ce4fef - Browse repository at this point
Copy the full SHA 8ce4fefView commit details -
hqq - fix weight check in check_quantized_param (huggingface#30748)
* hqq - fix weight check in check_quantized_param * ruff format
Configuration menu - View commit details
-
Copy full SHA for e0c3cee - Browse repository at this point
Copy the full SHA e0c3ceeView commit details
Commits on May 13, 2024
-
[awq] replace scale when we have GELU (huggingface#30074)
* fix awq test * style * add log * new fix * style * only modifying impacted model in the end * rename function
Configuration menu - View commit details
-
Copy full SHA for de6e0db - Browse repository at this point
Copy the full SHA de6e0dbView commit details -
Workflow: Replace
actions/post-slack
with centrally defined workflow (huggingface#30737) * Remove commit details * remove old workflow
Configuration menu - View commit details
-
Copy full SHA for a4e530e - Browse repository at this point
Copy the full SHA a4e530eView commit details -
Blip dynamic input resolution (huggingface#30722)
* blip with interpolated pos encoding * feat: Add interpolate_pos_encoding option to other models from `BLIP` family. * include check for textual generated content in tests
Configuration menu - View commit details
-
Copy full SHA for f63d822 - Browse repository at this point
Copy the full SHA f63d822View commit details -
Support for Falcon2-11B (huggingface#30771)
* remove unrelated changes * remove unrelated changes on phi and stable LM * add: Test for Falcon 10B * fix: formatting * fix: loading the falcon 10B in 8 bit precision using bitsanbytes. * fix: device placement * fix: broken tests. * fix: backwards compatibility for falcon 1B architecture. * chore: updated test. * chore: test_modeling_falcon.py to use the 11B model. * chore: minor edit * chore: formating. --------- Co-authored-by: Pablo Montalvo <[email protected]> Co-authored-by: ArthurZucker <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for e52741f - Browse repository at this point
Copy the full SHA e52741fView commit details -
[GroundingDino] Adding ms_deform_attn kernels (huggingface#30768)
* Adding ms_deform_attn kernels to GroundingDino * Pointing to deformable detr kernels
Configuration menu - View commit details
-
Copy full SHA for 453893e - Browse repository at this point
Copy the full SHA 453893eView commit details -
Llama: fix custom 4D masks, v2 (huggingface#30348)
* 4d mask fixes * Update custom 4D mask logic * test moved to mixin * extra tests 4d mask * upd 4d mask and StaticCache handling * added Mask4DTestHard to mistral tests * post-rebase fixes * test fixes for StaticCache * make fix-copies * upd 1 after huggingface#30476 * fix common tests * rm elif attention_mask.dim() == 4: * tests combined, fixed, mixtral supported * bigbird style chg reverted * rm if attention_mask.dim() == 2 * modeling_llama formatting chg --------- Co-authored-by: Joao Gante <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for a0779b9 - Browse repository at this point
Copy the full SHA a0779b9View commit details -
Generation / FIX: Fix multi-device generation (huggingface#30746)
* attempt to fix multi-device generation * fix * final fix * final fix * fix * fix * fix * fix * add joao suggestion * fix
Configuration menu - View commit details
-
Copy full SHA for f823fec - Browse repository at this point
Copy the full SHA f823fecView commit details -
Qwen: incorrect setup flag (huggingface#30776)
qwen does not support the new cache classes
Configuration menu - View commit details
-
Copy full SHA for f4dc26d - Browse repository at this point
Copy the full SHA f4dc26dView commit details -
enable Pipeline to get device from model (huggingface#30534)
* check model.device * fix * style fix * move model device * remove print * add comment * fix * add unit test * optimize * change test names and add more cases * Update tests/pipelines/test_pipelines_common.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 69d9bca - Browse repository at this point
Copy the full SHA 69d9bcaView commit details -
[Object detection pipeline] Lower threshold (huggingface#30710)
* Lower threshold * Address comment
Configuration menu - View commit details
-
Copy full SHA for ce87dca - Browse repository at this point
Copy the full SHA ce87dcaView commit details -
Configuration menu - View commit details
-
Copy full SHA for de2f722 - Browse repository at this point
Copy the full SHA de2f722View commit details -
Port IDEFICS to tensorflow (huggingface#26870)
* Initial commit * Just a copy of modeling_idefics.py that will be ported to TF * - Prepend TF to the name of all classes - Convert pytorch ops to TF (not all operations are converted yet) * Add TF imports * Add autotranslated files * Add TF classes to model_tf_auto.py * Add the TF classes in model_doc * include auto-translated code * Adopted from auto-translated version * Add a forgotten super().build * Add test code for TF version. * Fix indentation and load pytorch weights for now * Some fixes. Many tests are still failing but some are passing now. - I have added TODO's for some of the hacks I made to unblock me and I will address them soon - I have the processing_idefics.py hacked in my view to support TF temporarily * Add ALL_LAYERNORM_LAYERS to match pytorch * Revert "Add ALL_LAYERNORM_LAYERS to match pytorch" This reverts commit 7e0a351 as it is not needed in the tf implementation. * Fix freeze_relevant_params() * Some more fixes * Fix test_attention_outputs * Add tf stuff to processing_idefics.py processing_idefics.py supports both pytorch and tf now. test_processor_idefics.py for pytorch is passing, so i didn't break anything but still some issues with tf. I also need to add tf tests in test_processor_idefics.py. * Pass return_tensors to image processing code and fix test * Pass return_tensors to the image processor __init__ * Fix several test cases - Make input to some of the forward pass of type `TFModelInputType` - Decorate main layer forward pass with `@unpack_inputs` - Decorate main layer with `@keras_serializable` - Pass `inputs` to TFIdeficsModel * Some more fixes forgotten in last commit * Fix processing code and vision_tf.py * Fix perceiver bug * Import from * Auto-add build() methods + style pass * Fix build() errors due to `None` being passed as shape to some layers * Change name in TFIdeficsForVisionText2Text to attribute in IdeficsForVisionText2Text * Fix pytorch weights load for tf2 There were a lot of `name=` missing in weight initialization code. * Attempt to fix CI * Add back accidently removed line * Remove torch-specific stuff from the TF test file * make fix-copies, make style, remove autotranslated files * Fixes to imports/docstrings * Let's try the from future import in desperation * Fix the core random_attention_mask fn to match the torch/flax behaviour * Clean random_attention_mask up correctly * Remove torch-only test * Fix loss shape, couple of nits * make style * Don't test for OOB embeddings because IDEFICS uses those deliberately * Fix loss computation to handle masking * Fix test failures when flattening * Fix some test failures - Add cross attention gate which was missing and wasn't being passed arround - Fix overwriting of image_attention_mask due to hack I had for dummy inputs * Add a proper stateless scaled_dot_product_attention * make style * Adding missing attribute from the PyTorch version * Small cleanups to decoupledlinearlayer in case that helps * Pass epsilon to LayerNormalization * Attemp to fix pytorch weight cross-loading for TFIdeficsEmbedding * Fix a bug in TFIdeficsGatedCrossAttentionLayer * Patching up build() methods * Constant self.inv_freq * Constant self.inv_freq * First working version The TF implementation works now, there was a bug in the TFIdeficsDecoupledLinear where the weights were mis-intialized (in_features,out_features) when it should be: (out_features, in_features) I have tested this so far with tiny-random and idefics-9b-instruct and gives correct output. I also dumped the final outputs for both pytorch and TF and they are identical. * Fix some test failures * remove print statement * Fix return_tensors * Fix CI test failure check_code_quality * Attempt to fix CI failures by running `make fixup` The hardcoded IDs in test_modeling_tf_idefics.py are for the integration test and makes that file unreadable and should probably be moved to a seperate file. * Attempt to fix tests_pr_documentation_tests * Fix a test failure in test_image_processing_idefics.py * Fix test test_pt_tf_model_equivalence * Fix a few failures * Tiny fix * Some minor fixes * Remove a duplicate test * Override a few test failures for IDEFICS - `test_keras_save_load` is passing now - `test_compile_tf_model` is still failing * Fix processing_idefics.py after rebase * Guard import keras with is_tf_available * fix check code quality * fix check code quality * Minor fixes * Skip test_save_load temporarily This test passed on my local box but fails on the CI, skipping for now to see if there are other remaining failures on the CI. * Run `ruff format tests src utils` * Fix last failing test, `test_compile_tf_model` * Add fixes for vision_tf.py I forgot to add this file in last commit. * Minor fixes * Replace "<<<" with "<<" for doc tests IDEFICS-9B is too big for doctest runner, so don't run it there * Make code more readable * Fix bug after code review I added a layer_norm_eps to IdeficsConfig but I don't even need it since the vision config has a layer_norm_eps. * Fix after code review Use original code tokenizer.convert_tokens_to_ids * Keep PyTorch as the default return_tensors * Fixes to modeling_tf after code review * Fixes from code review - Remove all references of `TF_IDEFICS_PRETRAINED_MODEL_ARCHIVE_LIST` - Pass 1e-5 to LayerNormalization in perceiver * Run ruff * Undo a change * Refactor processing code after Matt's suggestion * Remove TODO's that aren't needed anymore * For pytorch, Use original pytorch processing code from main Since this PR is a TF port it shouldn't make any modifications to pytorch IDEFICS code. This changes undo's the pytorch processing modifications I made and uses original code from main. * Update tests/models/idefics/test_modeling_idefics.py * Update tests/models/idefics/test_modeling_tf_idefics.py * Add missing imports for is_pt_tf_cross_test * [DO NOT MERGE]: This is a commit for debugging and will be reverted The cross test `test_pt_tf_model_equivalence` passes locally but fails when running on the CI. This commit is to help debug that and will be reverted. * Revert "[DO NOT MERGE]: This is a commit for debugging and will be reverted" This reverts commit 8f0d709. * [DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted * [DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted * Revert "[DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted" This reverts commit 998cc38. * Revert "[DO NOT MERGE]: This commit is for debugging a CI failure and will be reverted" This reverts commit 1c695ac. * Don't skip test_save_load IIRC test_save_load was also failing on the CI but not on my local box, it might be easier to debug that on the CI first than the cross tests * Debugging commit, will be reverted * Revert "Debugging commit, will be reverted" This reverts commit 8eafc8e. * Override `test_save_load` and push model to save Maybe this will help me repro this weird bug * pass my repo_id * add endpoint * Pass a temp (write) token just for this CI * Undo last few commits, still pushing to hub for model debugging The issue seems to be with save_pretrained(), when I looked at the model saved from the CI test failure it is basically empty and has no weights. `self.save_weights(..)` seems to be failing in save_pretrained but needs more debugging * Add logging to modeling tf utils, will be reverted just for debugging * Debugging, will revert * Revert "Debugging, will revert" This reverts commit 9d0d307. * Revert "Add logging to modeling tf utils, will be reverted just for debugging" This reverts commit 774b6b7. * Remove `test_save_load` The CI failures are gone after my latest rebase, no idea why but I was still saving the model to my hub on HF and the tf_model.h5 file now has everything. * Run make fix-copies * Run ruff format tests src utils * Debugging commit, will be reverted * Run ruff, also trigger CI run * Run ruff again * Undo debugging commit --------- Co-authored-by: Matt <[email protected]> Co-authored-by: Matt <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9430635 - Browse repository at this point
Copy the full SHA 9430635View commit details -
Generate: assistant should be greedy in assisted decoding (huggingfac…
…e#30778) * assistant should be greedy * better comment * Update src/transformers/generation/candidate_generator.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 2e27291 - Browse repository at this point
Copy the full SHA 2e27291View commit details -
Save other CI jobs' result (torch/tf pipeline, example, deepspeed etc) (
huggingface#30699) * update * update * update * update * update * update * update * update * Update utils/notification_service.py Co-authored-by: amyeroberts <[email protected]> --------- Co-authored-by: ydshieh <[email protected]> Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 82c1625 - Browse repository at this point
Copy the full SHA 82c1625View commit details -
Deprecate models script (huggingface#30184)
* Add utility for finding candidate models for deprecation * Update model init * Make into configurable script * Fix path * Add sorting of base object alphabetically * Tidy * Refactor __init__ alpha ordering * Update script with logging * fix import * Fix logger * Fix logger * Get config file before moving files * Take models from CLI * Split models into lines to make easier to feed to deprecate_models script * Update * Use posix path * Print instead * Add example in module docstring * Fix up * Add clarifying comments; add models to DEPRECATE_MODELS * Address PR comments * Don't update relative paths on the same level
Configuration menu - View commit details
-
Copy full SHA for 0f8fefd - Browse repository at this point
Copy the full SHA 0f8fefdView commit details -
Configuration menu - View commit details
-
Copy full SHA for 539ed75 - Browse repository at this point
Copy the full SHA 539ed75View commit details -
CI: update to ROCm 6.0.2 and test MI300 (huggingface#30266)
* update to ROCm 6.0.2 and test MI300 * add callers for mi300 * update dockerfile * fix trainer tests * remove apex * style * Update tests/trainer/test_trainer_seq2seq.py * Update tests/trainer/test_trainer_seq2seq.py * Update tests/trainer/test_trainer_seq2seq.py * Update tests/trainer/test_trainer_seq2seq.py * update to torch 2.3 * add workflow dispatch target * we may need branches: mi300-ci after all * nit * fix docker build * nit * add check runner * remove docker-gpu * fix issues * fix --------- Co-authored-by: Yih-Dar <[email protected]> Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 37bba2a - Browse repository at this point
Copy the full SHA 37bba2aView commit details
Commits on May 14, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 449894d - Browse repository at this point
Copy the full SHA 449894dView commit details -
Fix cache type in Idefics2 (huggingface#30729)
standardize cache in idefics2
Configuration menu - View commit details
-
Copy full SHA for c02d302 - Browse repository at this point
Copy the full SHA c02d302View commit details -
Configuration menu - View commit details
-
Copy full SHA for 65ea190 - Browse repository at this point
Copy the full SHA 65ea190View commit details -
Add Watermarking LogitsProcessor and WatermarkDetector (huggingface#2…
…9676) * add watermarking processor * remove the other hashing (context width=1 always) * make style * Update src/transformers/generation/logits_process.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/logits_process.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/logits_process.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/configuration_utils.py Co-authored-by: Joao Gante <[email protected]> * update watermarking process * add detector * update tests to use detector * fix failing tests * rename `input_seq` * make style * doc for processor * minor fixes * docs * make quality * Update src/transformers/generation/configuration_utils.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/logits_process.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/watermarking.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/watermarking.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/watermarking.py Co-authored-by: Joao Gante <[email protected]> * add PR suggestions * let's use lru_cache's default max size (128) * import processor if torch available * maybe like this * lets move the config to torch independet file * add docs * tiny docs fix to make the test happy * Update src/transformers/generation/configuration_utils.py Co-authored-by: Joao Gante <[email protected]> * Update src/transformers/generation/watermarking.py Co-authored-by: Joao Gante <[email protected]> * PR suggestions * add docs * fix test * fix docs * address pr comments * style * Revert "style" This reverts commit 7f33cc3. * correct style * make doctest green --------- Co-authored-by: Joao Gante <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 5ad960f - Browse repository at this point
Copy the full SHA 5ad960fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 3ed384f - Browse repository at this point
Copy the full SHA 3ed384fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 55eb860 - Browse repository at this point
Copy the full SHA 55eb860View commit details -
Configuration menu - View commit details
-
Copy full SHA for f754335 - Browse repository at this point
Copy the full SHA f754335View commit details -
Configuration menu - View commit details
-
Copy full SHA for d8f8a9c - Browse repository at this point
Copy the full SHA d8f8a9cView commit details -
Configuration menu - View commit details
-
Copy full SHA for a449078 - Browse repository at this point
Copy the full SHA a449078View commit details
Commits on May 15, 2024
-
Configuration menu - View commit details
-
Copy full SHA for 3bdbb2e - Browse repository at this point
Copy the full SHA 3bdbb2eView commit details -
Update src/transformers/modeling_utils.py
Co-authored-by: amyeroberts <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 0ab79f6 - Browse repository at this point
Copy the full SHA 0ab79f6View commit details -
Configuration menu - View commit details
-
Copy full SHA for 65433c4 - Browse repository at this point
Copy the full SHA 65433c4View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1b5ae54 - Browse repository at this point
Copy the full SHA 1b5ae54View commit details -
Configuration menu - View commit details
-
Copy full SHA for d6b67c6 - Browse repository at this point
Copy the full SHA d6b67c6View commit details