Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add musicgen melody #3

Merged
merged 127 commits into from
Mar 22, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
127 commits
Select commit Hold shift + click to select a range
ed74d97
DeformableDETR support bfloat16 (#29232)
DonggeunYu Mar 4, 2024
836921f
Add UDOP (#22940)
NielsRogge Mar 4, 2024
e947683
[Docs] Spanish Translation -Torchscript md & Trainer md (#29310)
njackman-2344 Mar 4, 2024
81c8191
FIX [`Generation`] Fix some issues when running the MaxLength criteri…
younesbelkada Mar 5, 2024
4fc708f
Exllama kernels support for AWQ models (#28634)
IlyasMoutawwakil Mar 5, 2024
bd891ae
Fix max length for BLIP generation (#29296)
zucchini-nlp Mar 5, 2024
ebccb09
[docs] Update starcoder2 paper link (#29418)
xenova Mar 5, 2024
fa7f3cf
[tests] enable test_pipeline_accelerate_top_p on XPU (#29309)
faaany Mar 5, 2024
1328522
[`UdopTokenizer`] Fix post merge imports (#29451)
ArthurZucker Mar 5, 2024
b434f8a
fix md bug and add md to not_tested
ylacombe Mar 5, 2024
0d52f9f
more fix
ArthurZucker Mar 5, 2024
57d007b
Revert-commit 0d52f9f582efb82a12e8d9162b43a01b1aa0200f (#29455)
ArthurZucker Mar 5, 2024
4d892b7
[`Udop imports`] Processor tests were not run. (#29456)
ArthurZucker Mar 5, 2024
87a0783
Generate: inner decoding methods are no longer public (#29437)
gante Mar 5, 2024
fb1c62e
[`Add Mamba`] Adds support for the `Mamba` models (#28094)
ArthurZucker Mar 5, 2024
8f3f8e6
Fix bug with passing capture_* args to neptune callback (#29041)
AleksanderWWW Mar 5, 2024
9c5e560
Update pytest `import_path` location (#29154)
loadams Mar 5, 2024
a69cbf4
Automatic safetensors conversion when lacking these files (#29390)
LysandreJik Mar 5, 2024
638c423
[i18n-zh] Translate add_new_pipeline.md into Chinese (#29432)
windsonsea Mar 5, 2024
7b01579
🌐 [i18n-KO] Translated generation_strategies.md to Korean (#29086)
AI4Harmony Mar 5, 2024
00bf442
[FIX] `offload_weight()` takes from 3 to 4 positional arguments but 5…
faaany Mar 6, 2024
2a002d0
[`Docs` / `Awq`] Add docs on exllamav2 + AWQ (#29474)
younesbelkada Mar 6, 2024
b27aa20
[`docs`] Add starcoder2 docs (#29454)
younesbelkada Mar 6, 2024
2890116
Fix TrainingArguments regression with torch <2.0.0 for dataloader_pre…
ringohoffman Mar 6, 2024
41f7b7a
Generate: add tests for caches with `pad_to_multiple_of` (#29462)
gante Mar 6, 2024
700d48f
Generate: get generation mode from the generation config instance 🧼 (…
gante Mar 6, 2024
0a5b051
Avoid dummy token in PLD to optimize performance (#29445)
ofirzaf Mar 6, 2024
9322576
Fix test failure on DeepSpeed (#29444)
muellerzr Mar 6, 2024
ddb4fda
Generate: torch.compile-ready generation config preparation (#29443)
gante Mar 6, 2024
19fb1e2
added the max_matching_ngram_size to GenerationConfig (#29131)
mosheber Mar 6, 2024
965cf67
Fix `TextGenerationPipeline.__call__` docstring (#29491)
alvarobartt Mar 6, 2024
2a939f2
Substantially reduce memory usage in _update_causal_mask for large ba…
nqgl Mar 6, 2024
d45f47a
Fix: Disable torch.autocast in RotaryEmbedding of Gemma and LLaMa for…
currybab Mar 6, 2024
979fccc
Enable BLIP for auto VQA (#29499)
regisss Mar 7, 2024
ffe60fd
v4.39 deprecations 🧼 (#29492)
gante Mar 7, 2024
f6133d7
Revert "Automatic safetensors conversion when lacking these files (#2…
LysandreJik Mar 7, 2024
9288e75
fix: Avoid error when fsdp_config is missing xla_fsdp_v2 (#29480)
ashokponkumar Mar 7, 2024
923733c
Flava multimodal add attention mask (#29446)
zucchini-nlp Mar 7, 2024
45c0651
Add support for metadata format MLX (#29335)
alexweberk Mar 7, 2024
4ed9ae6
test_generation_config_is_loaded_with_model - fall back to pytorch m…
amyeroberts Mar 7, 2024
ddf177e
Set `inputs` as kwarg in `TextClassificationPipeline` (#29495)
alvarobartt Mar 7, 2024
b338a6c
Fix `VisionEncoderDecoder` Positional Arg (#29497)
nickthegroot Mar 7, 2024
631fa7b
Typo in mlx tensor support (#29509)
pcuenca Mar 8, 2024
bc764f4
Generate: left-padding test, revisited (#29515)
gante Mar 8, 2024
8e589c8
[tests] add the missing `require_sacremoses` decorator (#29504)
faaany Mar 8, 2024
8ee1d47
fix image-to-text batch incorrect output issue (#29342)
sywangyi Mar 8, 2024
14536c3
Typo fix in error message (#29535)
clefourrier Mar 8, 2024
1ea3ad1
[tests] use `torch_device` instead of `auto` for model testing (#29531)
faaany Mar 8, 2024
f386c51
StableLM: Fix dropout argument type error (#29236)
liangjs Mar 8, 2024
608fa54
Make sliding window size inclusive in eager attention (#29519)
jonatanklosko Mar 8, 2024
697f05b
fix typos in FSDP config parsing logic in `TrainingArguments` (#29189)
yundai424 Mar 8, 2024
1ba89dc
Fix WhisperNoSpeechDetection when input is full silence (#29065)
ylacombe Mar 8, 2024
3f6973d
[tests] use the correct `n_gpu` in `TrainerIntegrationTest::test_trai…
faaany Mar 8, 2024
469c132
Fix eval thread fork bomb (#29538)
muellerzr Mar 8, 2024
0290ec1
feat: use `warning_advice` for tensorflow warning (#29540)
winstxnhdw Mar 8, 2024
4f27ee9
[`Mamba doc`] Post merge updates (#29472)
ArthurZucker Mar 11, 2024
d80c9a3
[`Docs`] fixed minor typo (#29555)
j-gc Mar 11, 2024
6d67837
Add Fill-in-the-middle training objective example - PyTorch (#27464)
tanaymeh Mar 11, 2024
9a3f4d4
Bark model Flash Attention 2 Enabling to pass on check_device_map par…
damithsenanayake Mar 11, 2024
873d9bb
Make torch xla available on GPU (#29334)
yitongh Mar 11, 2024
dd1c905
[Docs] Fix FastSpeech2Conformer model doc links (#29574)
khipp Mar 11, 2024
e5eb55b
Don't use a subset in test fetcher if on `main` branch (#28816)
ydshieh Mar 11, 2024
47c9570
fix error: TypeError: Object of type Tensor is not JSON serializable …
yuanzhoulvpi2017 Mar 11, 2024
c1e478a
Add missing localized READMEs to the copies check (#29575)
khipp Mar 11, 2024
b45c0f5
Fixed broken link (#29558)
amritgupta98 Mar 11, 2024
73a2734
Tiny improvement for doc (#29581)
fzyzcjy Mar 11, 2024
b382a09
Experimental loading of MLX files (#29511)
pcuenca Mar 11, 2024
6cc5411
Fix Fuyu doc typos (#29601)
zucchini-nlp Mar 12, 2024
73efe89
Fix minor typo: softare => software (#29602)
DriesVerachtert Mar 12, 2024
81ec802
Stop passing None to compile() in TF examples (#29597)
Rocketknight1 Mar 12, 2024
50ec493
Fix typo (determine) (#29606)
koayon Mar 12, 2024
f1a565a
Implemented add_pooling_layer arg to TFBertModel (#29603)
tomigee Mar 12, 2024
b640486
Update legacy Repository usage in various example files (#29085)
Hvanderwilk Mar 12, 2024
df15425
Set env var to hold Keras at Keras 2 (#29598)
Rocketknight1 Mar 12, 2024
a15bd3a
Update flava tests (#29611)
ydshieh Mar 12, 2024
11163ff
Fix typo ; Update quantization.md (#29615)
furkanakkurt1335 Mar 12, 2024
8e64ba2
Add tests for batching support (#29297)
zucchini-nlp Mar 12, 2024
6b660d5
Fix: handle logging of scalars in Weights & Biases summary (#29612)
parambharat Mar 12, 2024
d479665
Examples: check `max_position_embeddings` in the translation example …
gante Mar 12, 2024
d522afe
[`Gemma`] Supports converting directly in half-precision (#29529)
younesbelkada Mar 12, 2024
be3fd8a
[Flash Attention 2] Add flash attention 2 for GPT-J (#28295)
bytebarde Mar 13, 2024
9acce7d
Core: Fix copies on main (#29624)
younesbelkada Mar 13, 2024
4afead8
[Whisper] Deprecate forced ids for v4.39 (#29485)
sanchit-gandhi Mar 13, 2024
38bff8c
Warn about tool use (#29628)
LysandreJik Mar 13, 2024
11bbb50
Adds pretrained IDs directly in the tests (#29534)
LysandreJik Mar 13, 2024
fafe909
[generate] deprecate forced ids processor (#29487)
sanchit-gandhi Mar 13, 2024
6247885
Fix minor typo: infenrece => inference (#29621)
DriesVerachtert Mar 13, 2024
88a4f68
[`MaskFormer`, `Mask2Former`] Use einsum where possible (#29544)
amyeroberts Mar 13, 2024
1e21c4f
Llama: allow custom 4d masks (#29618)
gante Mar 13, 2024
b340d90
[PyTorch/XLA] Fix extra TPU compilations introduced by recent changes…
alanwaketan Mar 13, 2024
d3801aa
[docs] Spanish translate chat_templating.md & yml addition (#29559)
njackman-2344 Mar 13, 2024
350c5d1
Add support for FSDP+QLoRA and DeepSpeed ZeRO3+QLoRA (#29587)
pacman100 Mar 13, 2024
3b6e95e
[`Mask2Former`] Move normalization for numerical stability (#29542)
amyeroberts Mar 13, 2024
a7e5e15
[tests] make `test_trainer_log_level_replica` to run on accelerators …
faaany Mar 13, 2024
31d0115
Refactor TFP call to just sigmoid() (#29641)
Rocketknight1 Mar 13, 2024
5ac264d
Fix batching tests for new models (Mamba and SegGPT) (#29633)
zucchini-nlp Mar 13, 2024
fe08556
Fix `multi_gpu_data_parallel_forward` for `MusicgenTest` (#29632)
ydshieh Mar 13, 2024
1fc505b
Add PvT-v2 Model (#26812)
FoamoftheSea Mar 13, 2024
f738ab3
[docs] Remove broken ChatML format link from chat_templating.md (#29643)
aaronjimv Mar 13, 2024
b4b9625
Add newly added PVTv2 model to all README files. (#29647)
robinverduijn Mar 14, 2024
c9e3c0b
[`PEFT`] Fix `save_pretrained` to make sure adapters weights are also…
shub-kris Mar 14, 2024
956f44f
Fix TPU checkpointing inside Trainer (#29657)
shub-kris Mar 14, 2024
2cc3cc8
Add `dataset_revision` argument to `RagConfig` (#29610)
ydshieh Mar 14, 2024
7b87ecb
Fix PVT v2 tests (#29660)
ydshieh Mar 14, 2024
23db187
Generate: handle `cache_position` update in `generate` (#29467)
gante Mar 14, 2024
48fbab7
Allow apply_chat_template to pass kwargs to the template and support …
Rocketknight1 Mar 14, 2024
56b64bf
Inaccurate code example within inline code-documentation (#29661)
MysteryManav Mar 14, 2024
f62407f
Extend import utils to cover "editable" torch versions (#29000)
bhack Mar 15, 2024
c47fcd0
Trainer: fail early in the presence of an unsavable `generation_confi…
gante Mar 15, 2024
53d8912
Pipeline: use tokenizer pad token at generation time if the model pad…
gante Mar 15, 2024
0e4a1c3
Cohere Model Release (#29622)
saurabhdash2512 Mar 15, 2024
c1993e6
[tests] remove deprecated tests for model loading (#29450)
faaany Mar 15, 2024
8a3cfaa
Fix AutoformerForPrediction example code (#29639)
m-torhan Mar 15, 2024
272f48e
[tests] ensure device-required software is available in the testing e…
faaany Mar 15, 2024
174aecd
Fix wrong condition used in `filter_models` (#29673)
ydshieh Mar 15, 2024
03847ef
fix: typos (#29653)
testwill Mar 15, 2024
f02aea2
Rename `glue` to `nyu-mll/glue` (#29679)
lhoestq Mar 15, 2024
28de2f4
[Quantization] Quanto quantizer (#29023)
SunMarc Mar 15, 2024
9e4df7c
Generate: replace breaks by a loop condition (#29662)
gante Mar 15, 2024
4e98d59
[FIX] Fix speech2test modeling tests (#29672)
ylacombe Mar 15, 2024
5011908
Revert "Fix wrong condition used in `filter_models`" (#29682)
ydshieh Mar 15, 2024
00c1d87
[docs] Spanish translation of attention.md (#29681)
aaronjimv Mar 15, 2024
ebeca43
Merge branch 'main' into add-musicgen-melody
ylacombe Mar 18, 2024
604a4c8
make fix-copies
ylacombe Mar 18, 2024
bf3dfd1
CI / generate: batch size computation compatible with all models (#29…
gante Mar 18, 2024
7bda3c3
Merge branch 'huggingface:main' into add-musicgen-melody
ylacombe Mar 18, 2024
5863cf9
fix hidden states test and batching
ylacombe Mar 18, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
2 changes: 2 additions & 0 deletions .circleci/create_circleci_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -475,6 +475,7 @@ def job_name(self):
"pip install -U --upgrade-strategy eager 'git+https://github.com/facebookresearch/detectron2.git'",
"sudo apt install tesseract-ocr",
"pip install -U --upgrade-strategy eager pytesseract",
"pip install --upgrade-strategy eager sentencepiece",
"pip install -U --upgrade-strategy eager natten==0.15.1+torch210cpu -f https://shi-labs.com/natten/wheels",
"pip install -U --upgrade-strategy eager python-Levenshtein",
"pip install -U --upgrade-strategy eager opencv-python",
Expand All @@ -485,6 +486,7 @@ def job_name(self):
"tests/models/*layoutlmv*",
"tests/models/*nat",
"tests/models/deta",
"tests/models/udop",
"tests/models/nougat",
],
pytest_num_workers=1,
Expand Down
8 changes: 6 additions & 2 deletions README.md

Large diffs are not rendered by default.

13 changes: 11 additions & 2 deletions README_de.md

Large diffs are not rendered by default.

8 changes: 6 additions & 2 deletions README_es.md

Large diffs are not rendered by default.

8 changes: 6 additions & 2 deletions README_fr.md

Large diffs are not rendered by default.

8 changes: 6 additions & 2 deletions README_hd.md

Large diffs are not rendered by default.

8 changes: 6 additions & 2 deletions README_ja.md

Large diffs are not rendered by default.

8 changes: 6 additions & 2 deletions README_ko.md

Large diffs are not rendered by default.

31 changes: 29 additions & 2 deletions README_pt-br.md

Large diffs are not rendered by default.

34 changes: 31 additions & 3 deletions README_ru.md

Large diffs are not rendered by default.

29 changes: 27 additions & 2 deletions README_te.md

Large diffs are not rendered by default.

9 changes: 8 additions & 1 deletion README_vi.md

Large diffs are not rendered by default.

8 changes: 6 additions & 2 deletions README_zh-hans.md

Large diffs are not rendered by default.

8 changes: 6 additions & 2 deletions README_zh-hant.md

Large diffs are not rendered by default.

5 changes: 4 additions & 1 deletion docker/transformers-quantization-latest-gpu/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,10 @@ RUN python3 -m pip install --no-cache-dir git+https://github.com/huggingface/opt
RUN python3 -m pip install --no-cache-dir aqlm[gpu]==1.0.2

# Add autoawq for quantization testing
RUN python3 -m pip install --no-cache-dir https://github.com/casper-hansen/AutoAWQ/releases/download/v0.1.8/autoawq-0.1.8+cu118-cp38-cp38-linux_x86_64.whl
RUN python3 -m pip install --no-cache-dir https://github.com/casper-hansen/AutoAWQ/releases/download/v0.2.0/autoawq-0.2.0+cu118-cp38-cp38-linux_x86_64.whl

# Add quanto for quantization testing
RUN python3 -m pip install --no-cache-dir quanto

# When installing in editable mode, `transformers` is not recognized as a package.
# this line must be added in order for python to be aware of transformers.
Expand Down
2 changes: 1 addition & 1 deletion docs/source/de/testing.md
Original file line number Diff line number Diff line change
Expand Up @@ -452,7 +452,7 @@ Dekorateure werden verwendet, um die Anforderungen von Tests in Bezug auf CPU/GP
- `require_torch_multi_gpu` - wie `require_torch` und zusätzlich mindestens 2 GPUs erforderlich
- `require_torch_non_multi_gpu` - wie `require_torch` plus benötigt 0 oder 1 GPUs
- `require_torch_up_to_2_gpus` - wie `require_torch` plus erfordert 0 oder 1 oder 2 GPUs
- `require_torch_tpu` - wie `require_torch` plus erfordert mindestens 1 TPU
- `require_torch_xla` - wie `require_torch` plus erfordert mindestens 1 TPU

Lassen Sie uns die GPU-Anforderungen in der folgenden Tabelle darstellen:

Expand Down
8 changes: 8 additions & 0 deletions docs/source/en/_toctree.yml
Original file line number Diff line number Diff line change
Expand Up @@ -310,6 +310,8 @@
title: CodeGen
- local: model_doc/code_llama
title: CodeLlama
- local: model_doc/cohere
title: Cohere
- local: model_doc/convbert
title: ConvBERT
- local: model_doc/cpm
Expand Down Expand Up @@ -398,6 +400,8 @@
title: M2M100
- local: model_doc/madlad-400
title: MADLAD-400
- local: model_doc/mamba
title: Mamba
- local: model_doc/marian
title: MarianMT
- local: model_doc/markuplm
Expand Down Expand Up @@ -581,6 +585,8 @@
title: PoolFormer
- local: model_doc/pvt
title: Pyramid Vision Transformer (PVT)
- local: model_doc/pvt_v2
title: Pyramid Vision Transformer v2 (PVTv2)
- local: model_doc/regnet
title: RegNet
- local: model_doc/resnet
Expand Down Expand Up @@ -772,6 +778,8 @@
title: TVLT
- local: model_doc/tvp
title: TVP
- local: model_doc/udop
title: UDOP
- local: model_doc/vilt
title: ViLT
- local: model_doc/vipllava
Expand Down
4 changes: 2 additions & 2 deletions docs/source/en/add_new_model.md
Original file line number Diff line number Diff line change
Expand Up @@ -89,8 +89,8 @@ model.config # model has access to its config
Similar to the model, the configuration inherits basic serialization and deserialization functionalities from
[`PretrainedConfig`]. Note that the configuration and the model are always serialized into two
different formats - the model to a *pytorch_model.bin* file and the configuration to a *config.json* file. Calling
[`~PreTrainedModel.save_pretrained`] will automatically call
[`~PretrainedConfig.save_pretrained`], so that both model and configuration are saved.
the model's [`~PreTrainedModel.save_pretrained`] will automatically call
the config's [`~PretrainedConfig.save_pretrained`], so that both model and configuration are saved.


### Code style
Expand Down
2 changes: 1 addition & 1 deletion docs/source/en/chat_templating.md
Original file line number Diff line number Diff line change
Expand Up @@ -375,7 +375,7 @@ best performance for inference or fine-tuning when you precisely match the token
If you're training a model from scratch, or fine-tuning a base language model for chat, on the other hand,
you have a lot of freedom to choose an appropriate template! LLMs are smart enough to learn to handle lots of different
input formats. Our default template for models that don't have a class-specific template follows the
[ChatML format](https://github.com/openai/openai-python/blob/main/chatml.md), and this is a good, flexible choice for many use-cases. It looks like this:
`ChatML` format, and this is a good, flexible choice for many use-cases. It looks like this:

```
{% for message in messages %}
Expand Down
9 changes: 9 additions & 0 deletions docs/source/en/custom_tools.md
Original file line number Diff line number Diff line change
Expand Up @@ -427,6 +427,15 @@ To upload your custom prompt on a repo on the Hub and share it with the communit

## Using custom tools

<Tip warning={true}>

Using custom tools in your local runtime means that you'll download code to run on your machine.

ALWAYS inspect the tool you're downloading before loading it within your runtime, as you would do when
installing a package using pip/npm/apt.

</Tip>

In this section, we'll be leveraging two existing custom tools that are specific to image generation:

- We replace [huggingface-tools/image-transformation](https://huggingface.co/spaces/huggingface-tools/image-transformation),
Expand Down
3 changes: 3 additions & 0 deletions docs/source/en/generation_strategies.md
Original file line number Diff line number Diff line change
Expand Up @@ -389,3 +389,6 @@ just like in multinomial sampling. However, in assisted decoding, reducing the t
>>> tokenizer.batch_decode(outputs, skip_special_tokens=True)
['Alice and Bob are going to the same party. It is a small party, in a small']
```

Alternativelly, you can also set the `prompt_lookup_num_tokens` to trigger n-gram based assisted decoding, as opposed
to model based assisted decoding. You can read more about it [here](https://twitter.com/joao_gante/status/1747322413006643259).
4 changes: 4 additions & 0 deletions docs/source/en/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -95,6 +95,7 @@ Flax), PyTorch, and/or TensorFlow.
| [CLVP](model_doc/clvp) | ✅ | ❌ | ❌ |
| [CodeGen](model_doc/codegen) | ✅ | ❌ | ❌ |
| [CodeLlama](model_doc/code_llama) | ✅ | ❌ | ✅ |
| [Cohere](model_doc/cohere) | ✅ | ❌ | ❌ |
| [Conditional DETR](model_doc/conditional_detr) | ✅ | ❌ | ❌ |
| [ConvBERT](model_doc/convbert) | ✅ | ✅ | ❌ |
| [ConvNeXT](model_doc/convnext) | ✅ | ✅ | ❌ |
Expand Down Expand Up @@ -180,6 +181,7 @@ Flax), PyTorch, and/or TensorFlow.
| [M-CTC-T](model_doc/mctct) | ✅ | ❌ | ❌ |
| [M2M100](model_doc/m2m_100) | ✅ | ❌ | ❌ |
| [MADLAD-400](model_doc/madlad-400) | ✅ | ✅ | ✅ |
| [Mamba](model_doc/mamba) | ✅ | ❌ | ❌ |
| [Marian](model_doc/marian) | ✅ | ✅ | ✅ |
| [MarkupLM](model_doc/markuplm) | ✅ | ❌ | ❌ |
| [Mask2Former](model_doc/mask2former) | ✅ | ❌ | ❌ |
Expand Down Expand Up @@ -234,6 +236,7 @@ Flax), PyTorch, and/or TensorFlow.
| [Pop2Piano](model_doc/pop2piano) | ✅ | ❌ | ❌ |
| [ProphetNet](model_doc/prophetnet) | ✅ | ❌ | ❌ |
| [PVT](model_doc/pvt) | ✅ | ❌ | ❌ |
| [PVTv2](model_doc/pvt_v2) | ✅ | ❌ | ❌ |
| [QDQBert](model_doc/qdqbert) | ✅ | ❌ | ❌ |
| [Qwen2](model_doc/qwen2) | ✅ | ❌ | ❌ |
| [RAG](model_doc/rag) | ✅ | ✅ | ❌ |
Expand Down Expand Up @@ -280,6 +283,7 @@ Flax), PyTorch, and/or TensorFlow.
| [TrOCR](model_doc/trocr) | ✅ | ❌ | ❌ |
| [TVLT](model_doc/tvlt) | ✅ | ❌ | ❌ |
| [TVP](model_doc/tvp) | ✅ | ❌ | ❌ |
| [UDOP](model_doc/udop) | ✅ | ❌ | ❌ |
| [UL2](model_doc/ul2) | ✅ | ✅ | ✅ |
| [UMT5](model_doc/umt5) | ✅ | ❌ | ❌ |
| [UniSpeech](model_doc/unispeech) | ✅ | ❌ | ❌ |
Expand Down
19 changes: 2 additions & 17 deletions docs/source/en/internal/generation_utils.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,16 +16,7 @@ rendered properly in your Markdown viewer.

# Utilities for Generation

This page lists all the utility functions used by [`~generation.GenerationMixin.generate`],
[`~generation.GenerationMixin.greedy_search`],
[`~generation.GenerationMixin.contrastive_search`],
[`~generation.GenerationMixin.sample`],
[`~generation.GenerationMixin.beam_search`],
[`~generation.GenerationMixin.beam_sample`],
[`~generation.GenerationMixin.group_beam_search`], and
[`~generation.GenerationMixin.constrained_beam_search`].

Most of those are only useful if you are studying the code of the generate methods in the library.
This page lists all the utility functions used by [`~generation.GenerationMixin.generate`].

## Generate Outputs

Expand Down Expand Up @@ -345,12 +336,6 @@ A [`Constraint`] can be used to force the generation to include specific tokens
- process
- finalize

## Utilities

[[autodoc]] top_k_top_p_filtering

[[autodoc]] tf_top_k_top_p_filtering

## Streamers

[[autodoc]] TextStreamer
Expand All @@ -376,4 +361,4 @@ A [`Constraint`] can be used to force the generation to include specific tokens

[[autodoc]] StaticCache
- update
- get_seq_length
- get_seq_length
6 changes: 5 additions & 1 deletion docs/source/en/main_classes/quantization.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ rendered properly in your Markdown viewer.

# Quantization

Quantization techniques reduces memory and computational costs by representing weights and activations with lower-precision data types like 8-bit integers (int8). This enables loading larger models you normally wouldn't be able to fit into memory, and speeding up inference. Transformers supports the AWQ and GPTQ quantization algorithms and it supports 8-bit and 4-bit quantization with bitsandbytes.
Quantization techniques reduce memory and computational costs by representing weights and activations with lower-precision data types like 8-bit integers (int8). This enables loading larger models you normally wouldn't be able to fit into memory, and speeding up inference. Transformers supports the AWQ and GPTQ quantization algorithms and it supports 8-bit and 4-bit quantization with bitsandbytes.

Quantization techniques that aren't supported in Transformers can be added with the [`HfQuantizer`] class.

Expand All @@ -26,6 +26,10 @@ Learn how to quantize models in the [Quantization](../quantization) guide.

</Tip>

## QuantoConfig

[[autodoc]] QuantoConfig

## AqlmConfig

[[autodoc]] AqlmConfig
Expand Down
10 changes: 3 additions & 7 deletions docs/source/en/main_classes/text_generation.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,19 +37,15 @@ like token streaming.
- from_pretrained
- from_model_config
- save_pretrained
- update
- validate
- get_generation_mode

## GenerationMixin

[[autodoc]] generation.GenerationMixin
- generate
- compute_transition_scores
- greedy_search
- sample
- beam_search
- beam_sample
- contrastive_search
- group_beam_search
- constrained_beam_search

## TFGenerationMixin

Expand Down
2 changes: 1 addition & 1 deletion docs/source/en/model_doc/bert.md
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ A list of official Hugging Face and community (indicated by 🌎) resources to h
<PipelineTag pipeline="token-classification"/>

- A blog post on how to use [Hugging Face Transformers with Keras: Fine-tune a non-English BERT for Named Entity Recognition](https://www.philschmid.de/huggingface-transformers-keras-tf).
- A notebook for [Finetuning BERT for named-entity recognition](https://colab.research.google.com/github/NielsRogge/Transformers-Tutorials/blob/master/Custom_Named_Entity_Recognition_with_BERT_only_first_wordpiece.ipynb) using only the first wordpiece of each word in the word label during tokenization. To propagate the label of the word to all wordpieces, see this [version](https://github.com/NielsRogge/Transformers-Tutorials/blob/master/BERT/Custom_Named_Entity_Recognition_with_BERT.ipynb) of the notebook instead.
- A notebook for [Finetuning BERT for named-entity recognition](https://colab.research.google.com/github/NielsRogge/Transformers-Tutorials/blob/master/BERT/Custom_Named_Entity_Recognition_with_BERT_only_first_wordpiece.ipynb) using only the first wordpiece of each word in the word label during tokenization. To propagate the label of the word to all wordpieces, see this [version](https://github.com/NielsRogge/Transformers-Tutorials/blob/master/BERT/Custom_Named_Entity_Recognition_with_BERT.ipynb) of the notebook instead.
- [`BertForTokenClassification`] is supported by this [example script](https://github.com/huggingface/transformers/tree/main/examples/pytorch/token-classification) and [notebook](https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/token_classification.ipynb).
- [`TFBertForTokenClassification`] is supported by this [example script](https://github.com/huggingface/transformers/tree/main/examples/tensorflow/token-classification) and [notebook](https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/token_classification-tf.ipynb).
- [`FlaxBertForTokenClassification`] is supported by this [example script](https://github.com/huggingface/transformers/tree/main/examples/flax/token-classification).
Expand Down
Loading
Loading