Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add EnCodec model #23655

Merged
merged 139 commits into from
Jun 14, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
139 commits
Select commit Hold shift + click to select a range
e18fd4d
boilerplate stuff
hollance May 22, 2023
1c98495
messing around with the feature extractor
hollance May 22, 2023
4b88774
fix feature extractor
hollance May 31, 2023
d075dfe
unit tests for feature extractor
hollance May 31, 2023
1d6a752
rename speech to audio
hollance May 31, 2023
56d29e6
quick-and-dirty import of Meta's code
hollance Jun 1, 2023
7229a84
import weights (sort of)
hollance Jun 1, 2023
c16b225
cleaning up
hollance Jun 1, 2023
b044acc
more cleaning up
hollance Jun 1, 2023
66978bd
move encoder/decoder args into config
hollance Jun 1, 2023
3e1dea4
cleanup model
hollance Jun 8, 2023
32d54d5
rename EnCodec -> Encodec
hollance Jun 8, 2023
d919579
RVQ parameters in config
hollance Jun 8, 2023
027ee65
add slow test
patrickvonplaten Jun 8, 2023
6d8319c
Merge branch 'encodec' of https://github.com/hollance/transformers in…
patrickvonplaten Jun 8, 2023
b12741d
add lstm init and test_init
ArthurZucker Jun 8, 2023
9fe5d98
Add save & load
patrickvonplaten Jun 8, 2023
d169637
finish EncodecModel
hollance Jun 8, 2023
b744397
remove decoder_input_values as they are ont used anywhere (not remove…
ArthurZucker Jun 8, 2023
548a5eb
fix test feature extraction model name
ArthurZucker Jun 8, 2023
ab24b2b
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 8, 2023
ebe61e3
Add better slow test
patrickvonplaten Jun 8, 2023
5ef5116
Merge branch 'encodec' of https://github.com/hollance/transformers in…
patrickvonplaten Jun 8, 2023
c7740ff
Fix tests
patrickvonplaten Jun 8, 2023
787ea90
some fixup and cleaning
ArthurZucker Jun 8, 2023
d353269
Improve further
patrickvonplaten Jun 8, 2023
b4e95b1
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 8, 2023
df01864
cleaning up quantizer
hollance Jun 8, 2023
623fedf
fix up conversion script
hollance Jun 8, 2023
84b6c67
test don't pass, _encode_fram does not work
ArthurZucker Jun 8, 2023
3d7d610
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 8, 2023
ab7fe31
update tests with output per encode and decode
ArthurZucker Jun 8, 2023
921aaa3
more cleanup
hollance Jun 8, 2023
7b77f2a
rename _codebook
hollance Jun 8, 2023
9b4f126
remove old config cruft
hollance Jun 8, 2023
b312836
ratios & hop_length
hollance Jun 8, 2023
16ec33c
use ModuleList instead of Sequential
hollance Jun 8, 2023
9961fb3
clean up resnet block
hollance Jun 8, 2023
71d1697
update types
ArthurZucker Jun 8, 2023
aa189bf
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 8, 2023
455f73f
update tests
ArthurZucker Jun 8, 2023
c3e6271
fixup
ArthurZucker Jun 8, 2023
3e7827f
quick cleanup
hollance Jun 8, 2023
006f8c1
fix padding
ArthurZucker Jun 8, 2023
9f2ff53
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 8, 2023
3e44d6b
more styl,ing
ArthurZucker Jun 8, 2023
0ebbeb7
add patrick feedback
hollance Jun 8, 2023
742fccf
fix copies
ArthurZucker Jun 8, 2023
7a865b7
fixup
ArthurZucker Jun 8, 2023
6bca401
fix lstm
hollance Jun 8, 2023
b8754df
fix shape issues
ArthurZucker Jun 8, 2023
eba5b15
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 8, 2023
c8075a8
fixup
ArthurZucker Jun 8, 2023
2d73c63
rename conv layers
hollance Jun 8, 2023
1d94c56
fixup
ArthurZucker Jun 8, 2023
66299b6
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 8, 2023
143a027
fix decoding
ArthurZucker Jun 8, 2023
eca836d
small conv refactoring
hollance Jun 8, 2023
161caad
remove norm_params
hollance Jun 8, 2023
7d7f8b5
simplify conv layers
hollance Jun 8, 2023
f235445
rename conv layers
hollance Jun 8, 2023
78e524c
stuff
hollance Jun 8, 2023
b90aca8
Clean up
patrickvonplaten Jun 10, 2023
531e235
Add padding logic
ArthurZucker Jun 8, 2023
fa6fc45
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 10, 2023
93ac09f
clean up more
patrickvonplaten Jun 10, 2023
df38828
clean up more
patrickvonplaten Jun 10, 2023
cb67a6f
clean up more
patrickvonplaten Jun 10, 2023
af09b86
More clean ups
patrickvonplaten Jun 10, 2023
9872b30
cleanup convolutions
ArthurZucker Jun 11, 2023
451eedf
typo
ArthurZucker Jun 11, 2023
f224377
fix typos
ArthurZucker Jun 11, 2023
abc427f
fixup
ArthurZucker Jun 11, 2023
48c7eb7
build PR doc?
ArthurZucker Jun 11, 2023
a24e759
start refactoring docstring
ArthurZucker Jun 12, 2023
9e1f613
fix don't pad when no strid and chunk
ArthurZucker Jun 12, 2023
3f72c10
update docstring
ArthurZucker Jun 12, 2023
db6d7c3
update docstring
ArthurZucker Jun 12, 2023
c2aaed1
nits
ArthurZucker Jun 12, 2023
ff155f6
update going to lunch
ArthurZucker Jun 12, 2023
ace6a9b
update config and model
ArthurZucker Jun 12, 2023
328b1d4
fix broken testse (becaue of the config changes)
ArthurZucker Jun 12, 2023
6a47642
fix scale computation
ArthurZucker Jun 12, 2023
f7b7e07
fixu[
ArthurZucker Jun 12, 2023
6605f24
only return dict if speciefied or if config returns it
ArthurZucker Jun 12, 2023
66ae559
remove todos
ArthurZucker Jun 12, 2023
c89bb4f
update defaults in config
ArthurZucker Jun 12, 2023
3d8233f
update conversion script
ArthurZucker Jun 12, 2023
c0a91ca
fix doctest
ArthurZucker Jun 12, 2023
57a6e7c
more docstring + fixup
ArthurZucker Jun 12, 2023
f5ba688
nits on batched_tests
ArthurZucker Jun 12, 2023
f30273b
more nits
ArthurZucker Jun 12, 2023
434dd0f
Apply suggestions from code review
ArthurZucker Jun 12, 2023
8fe483c
update basxed on review
ArthurZucker Jun 12, 2023
3a34633
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 12, 2023
836ec56
fix update
ArthurZucker Jun 12, 2023
f8025b8
updaet tests
ArthurZucker Jun 12, 2023
a77faed
Apply suggestions from code review
ArthurZucker Jun 12, 2023
4cb0310
fixup
ArthurZucker Jun 12, 2023
ef9151e
add overlap and chunl_length_s
ArthurZucker Jun 13, 2023
9c9ca46
cleanup feature extraction
ArthurZucker Jun 13, 2023
1a53fe8
teste edge cases truncation and padding
ArthurZucker Jun 13, 2023
b673e96
correct processor values
ArthurZucker Jun 13, 2023
f6bae2f
update config encodec, nits
ArthurZucker Jun 13, 2023
7d63e10
fix tests
ArthurZucker Jun 13, 2023
5a0f715
fixup
ArthurZucker Jun 13, 2023
e139702
fix 24Hz test
ArthurZucker Jun 13, 2023
2d17116
elle tests are green
ArthurZucker Jun 13, 2023
3cbc1c4
fix fixup
ArthurZucker Jun 13, 2023
799d0e7
Apply suggestions from code review
ArthurZucker Jun 13, 2023
0d9edea
revert readme changes
ArthurZucker Jun 13, 2023
7ab8927
fixup
ArthurZucker Jun 13, 2023
0a4a034
add example
ArthurZucker Jun 13, 2023
c7c4e04
use facebook checkpoints
ArthurZucker Jun 13, 2023
0631459
fix typo
ArthurZucker Jun 13, 2023
9767ad8
no pipeline tests
ArthurZucker Jun 13, 2023
5c05ae9
use slef.pad everywhere we can
ArthurZucker Jun 13, 2023
23b4308
Apply suggestions from code review
ArthurZucker Jun 13, 2023
a1b3723
update based on review
ArthurZucker Jun 13, 2023
eb3427f
update
ArthurZucker Jun 14, 2023
c833e52
update mdx
ArthurZucker Jun 14, 2023
01da412
fix bug and tests
ArthurZucker Jun 14, 2023
31a7e7e
fixup
ArthurZucker Jun 14, 2023
5073411
fix doctest
ArthurZucker Jun 14, 2023
768d149
remove comment
ArthurZucker Jun 14, 2023
e57aae1
more nits
ArthurZucker Jun 14, 2023
4e69fee
add more coverage for `test_truncation_and_padding`
ArthurZucker Jun 14, 2023
5ffa9af
fixup
ArthurZucker Jun 14, 2023
b25ce23
add last test
ArthurZucker Jun 14, 2023
84f8c5c
fix text
ArthurZucker Jun 14, 2023
c0162b4
nits
ArthurZucker Jun 14, 2023
14d5e3e
Update tests/models/encodec/test_modeling_encodec.py
ArthurZucker Jun 14, 2023
edff1a6
take care of the last comments
ArthurZucker Jun 14, 2023
f5332dc
typo
ArthurZucker Jun 14, 2023
7d9fcc0
fix test
ArthurZucker Jun 14, 2023
9d00eb2
Merge branch 'encodec' of https://github.com/hollance/transformers in…
ArthurZucker Jun 14, 2023
54f9d4c
nits
ArthurZucker Jun 14, 2023
e6b8333
fixup
ArthurZucker Jun 14, 2023
b187648
Update src/transformers/models/encodec/feature_extraction_encodec.py
ArthurZucker Jun 14, 2023
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -346,6 +346,7 @@ Current number of checkpoints: ![](https://img.shields.io/endpoint?url=https://h
1. **[EfficientFormer](https://huggingface.co/docs/transformers/model_doc/efficientformer)** (from Snap Research) released with the paper [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191) by Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren.
1. **[EfficientNet](https://huggingface.co/docs/transformers/model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](https://huggingface.co/docs/transformers/model_doc/electra)** (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang Luong, Quoc V. Le, Christopher D. Manning.
1. **[EnCodec](https://huggingface.co/docs/transformers/main/model_doc/encodec)** (from Meta AI) released with the paper [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438) by Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi.
1. **[EncoderDecoder](https://huggingface.co/docs/transformers/model_doc/encoder-decoder)** (from Google Research) released with the paper [Leveraging Pre-trained Checkpoints for Sequence Generation Tasks](https://arxiv.org/abs/1907.12461) by Sascha Rothe, Shashi Narayan, Aliaksei Severyn.
1. **[ERNIE](https://huggingface.co/docs/transformers/model_doc/ernie)** (from Baidu) released with the paper [ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/abs/1904.09223) by Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu.
1. **[ErnieM](https://huggingface.co/docs/transformers/model_doc/ernie_m)** (from Baidu) released with the paper [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674) by Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang.
Expand Down
1 change: 1 addition & 0 deletions README_es.md
Original file line number Diff line number Diff line change
Expand Up @@ -321,6 +321,7 @@ Número actual de puntos de control: ![](https://img.shields.io/endpoint?url=htt
1. **[EfficientFormer](https://huggingface.co/docs/transformers/model_doc/efficientformer)** (from Snap Research) released with the paper [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191) by Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren.
1. **[EfficientNet](https://huggingface.co/docs/transformers/model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](https://huggingface.co/docs/transformers/model_doc/electra)** (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang Luong, Quoc V. Le, Christopher D. Manning.
1. **[EnCodec](https://huggingface.co/docs/transformers/main/model_doc/encodec)** (from Meta AI) released with the paper [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438) by Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi.
1. **[EncoderDecoder](https://huggingface.co/docs/transformers/model_doc/encoder-decoder)** (from Google Research) released with the paper [Leveraging Pre-trained Checkpoints for Sequence Generation Tasks](https://arxiv.org/abs/1907.12461) by Sascha Rothe, Shashi Narayan, Aliaksei Severyn.
1. **[ERNIE](https://huggingface.co/docs/transformers/model_doc/ernie)** (from Baidu) released with the paper [ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/abs/1904.09223) by Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu.
1. **[ErnieM](https://huggingface.co/docs/transformers/model_doc/ernie_m)** (from Baidu) released with the paper [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674) by Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang.
Expand Down
1 change: 1 addition & 0 deletions README_hd.md
Original file line number Diff line number Diff line change
Expand Up @@ -293,6 +293,7 @@ conda install -c huggingface transformers
1. **[EfficientFormer](https://huggingface.co/docs/transformers/model_doc/efficientformer)** (from Snap Research) released with the paper [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191) by Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren.
1. **[EfficientNet](https://huggingface.co/docs/transformers/model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](https://huggingface.co/docs/transformers/model_doc/electra)** (Google रिसर्च/स्टैनफोर्ड यूनिवर्सिटी से) साथ में दिया गया पेपर [इलेक्ट्रा: जेनरेटर के बजाय भेदभाव करने वाले के रूप में टेक्स्ट एन्कोडर्स का पूर्व-प्रशिक्षण] (https://arxiv.org/abs/2003.10555) केविन क्लार्क, मिन्ह-थांग लुओंग, क्वोक वी. ले, क्रिस्टोफर डी. मैनिंग द्वारा पोस्ट किया गया।
1. **[EnCodec](https://huggingface.co/docs/transformers/main/model_doc/encodec)** (Meta AI से) Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi. द्वाराअनुसंधान पत्र [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438) के साथ जारी किया गया
1. **[EncoderDecoder](https://huggingface.co/docs/transformers/model_doc/encoder-decoder)** (Google रिसर्च से) साथ में दिया गया पेपर [सीक्वेंस जेनरेशन टास्क के लिए प्री-ट्रेंड चेकपॉइंट का इस्तेमाल करना](https:/ /arxiv.org/abs/1907.12461) साशा रोठे, शशि नारायण, अलियाक्सि सेवेरिन द्वारा।
1. **[ERNIE](https://huggingface.co/docs/transformers/model_doc/ernie)**(Baidu से) साथ देने वाला पेपर [ERNIE: एन्हांस्ड रिप्रेजेंटेशन थ्रू नॉलेज इंटीग्रेशन](https://arxiv.org/abs/1904.09223) यू सन, शुओहुआन वांग, युकुन ली, शिकुन फेंग, ज़ुई चेन, हान झांग, शिन तियान, डैनक्सियांग झू, हाओ तियान, हुआ वू द्वारा पोस्ट किया गया।
1. **[ErnieM](https://huggingface.co/docs/transformers/model_doc/ernie_m)** (Baidu से) Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang. द्वाराअनुसंधान पत्र [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674) के साथ जारी किया गया
Expand Down
1 change: 1 addition & 0 deletions README_ja.md
Original file line number Diff line number Diff line change
Expand Up @@ -355,6 +355,7 @@ Flax、PyTorch、TensorFlowをcondaでインストールする方法は、それ
1. **[EfficientFormer](https://huggingface.co/docs/transformers/model_doc/efficientformer)** (Snap Research から) Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren. から公開された研究論文 [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191)
1. **[EfficientNet](https://huggingface.co/docs/transformers/model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](https://huggingface.co/docs/transformers/model_doc/electra)** (Google Research/Stanford University から) Kevin Clark, Minh-Thang Luong, Quoc V. Le, Christopher D. Manning から公開された研究論文: [ELECTRA: Pre-training text encoders as discriminators rather than generators](https://arxiv.org/abs/2003.10555)
1. **[EnCodec](https://huggingface.co/docs/transformers/main/model_doc/encodec)** (Meta AI から) Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi. から公開された研究論文 [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438)
1. **[EncoderDecoder](https://huggingface.co/docs/transformers/model_doc/encoder-decoder)** (Google Research から) Sascha Rothe, Shashi Narayan, Aliaksei Severyn から公開された研究論文: [Leveraging Pre-trained Checkpoints for Sequence Generation Tasks](https://arxiv.org/abs/1907.12461)
1. **[ERNIE](https://huggingface.co/docs/transformers/model_doc/ernie)** (Baidu から) Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu から公開された研究論文: [ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/abs/1904.09223)
1. **[ErnieM](https://huggingface.co/docs/transformers/model_doc/ernie_m)** (Baidu から) Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang. から公開された研究論文 [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674)
Expand Down
1 change: 1 addition & 0 deletions README_ko.md
Original file line number Diff line number Diff line change
Expand Up @@ -270,6 +270,7 @@ Flax, PyTorch, TensorFlow 설치 페이지에서 이들을 conda로 설치하는
1. **[EfficientFormer](https://huggingface.co/docs/transformers/model_doc/efficientformer)** (from Snap Research) released with the paper [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191) by Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren.
1. **[EfficientNet](https://huggingface.co/docs/transformers/model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](https://huggingface.co/docs/transformers/model_doc/electra)** (Google Research/Stanford University 에서) Kevin Clark, Minh-Thang Luong, Quoc V. Le, Christopher D. Manning 의 [ELECTRA: Pre-training text encoders as discriminators rather than generators](https://arxiv.org/abs/2003.10555) 논문과 함께 발표했습니다.
1. **[EnCodec](https://huggingface.co/docs/transformers/main/model_doc/encodec)** (Meta AI 에서 제공)은 Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi.의 [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438)논문과 함께 발표했습니다.
1. **[EncoderDecoder](https://huggingface.co/docs/transformers/model_doc/encoder-decoder)** (Google Research 에서) Sascha Rothe, Shashi Narayan, Aliaksei Severyn 의 [Leveraging Pre-trained Checkpoints for Sequence Generation Tasks](https://arxiv.org/abs/1907.12461) 논문과 함께 발표했습니다.
1. **[ERNIE](https://huggingface.co/docs/transformers/model_doc/ernie)** (Baidu 에서) Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu 의 [ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/abs/1904.09223) 논문과 함께 발표했습니다.
1. **[ErnieM](https://huggingface.co/docs/transformers/model_doc/ernie_m)** (Baidu 에서 제공)은 Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang.의 [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674)논문과 함께 발표했습니다.
Expand Down
1 change: 1 addition & 0 deletions README_zh-hans.md
Original file line number Diff line number Diff line change
Expand Up @@ -294,6 +294,7 @@ conda install -c huggingface transformers
1. **[EfficientFormer](https://huggingface.co/docs/transformers/model_doc/efficientformer)** (来自 Snap Research) 伴随论文 [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191) 由 Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren 发布。
1. **[EfficientNet](https://huggingface.co/docs/transformers/model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](https://huggingface.co/docs/transformers/model_doc/electra)** (来自 Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le, Christopher D. Manning 发布。
1. **[EnCodec](https://huggingface.co/docs/transformers/main/model_doc/encodec)** (来自 Meta AI) 伴随论文 [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438) 由 Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi 发布。
1. **[EncoderDecoder](https://huggingface.co/docs/transformers/model_doc/encoder-decoder)** (来自 Google Research) 伴随论文 [Leveraging Pre-trained Checkpoints for Sequence Generation Tasks](https://arxiv.org/abs/1907.12461) 由 Sascha Rothe, Shashi Narayan, Aliaksei Severyn 发布。
1. **[ERNIE](https://huggingface.co/docs/transformers/model_doc/ernie)** (来自 Baidu) 伴随论文 [ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/abs/1904.09223) by Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu 发布。
1. **[ErnieM](https://huggingface.co/docs/transformers/model_doc/ernie_m)** (来自 Baidu) 伴随论文 [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674) 由 Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang 发布。
Expand Down
1 change: 1 addition & 0 deletions README_zh-hant.md
Original file line number Diff line number Diff line change
Expand Up @@ -306,6 +306,7 @@ conda install -c huggingface transformers
1. **[EfficientFormer](https://huggingface.co/docs/transformers/model_doc/efficientformer)** (from Snap Research) released with the paper [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191) by Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren.
1. **[EfficientNet](https://huggingface.co/docs/transformers/model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](https://huggingface.co/docs/transformers/model_doc/electra)** (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang Luong, Quoc V. Le, Christopher D. Manning.
1. **[EnCodec](https://huggingface.co/docs/transformers/main/model_doc/encodec)** (from Meta AI) released with the paper [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438) by Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi.
1. **[EncoderDecoder](https://huggingface.co/docs/transformers/model_doc/encoder-decoder)** (from Google Research) released with the paper [Leveraging Pre-trained Checkpoints for Sequence Generation Tasks](https://arxiv.org/abs/1907.12461) by Sascha Rothe, Shashi Narayan, Aliaksei Severyn.
1. **[ERNIE](https://huggingface.co/docs/transformers/model_doc/ernie)** (from Baidu) released with the paper [ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/abs/1904.09223) by Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu.
1. **[ErnieM](https://huggingface.co/docs/transformers/model_doc/ernie_m)** (from Baidu) released with the paper [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674) by Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang.
Expand Down
2 changes: 2 additions & 0 deletions docs/source/en/_toctree.yml
Original file line number Diff line number Diff line change
Expand Up @@ -541,6 +541,8 @@
title: Audio Spectrogram Transformer
- local: model_doc/clap
title: CLAP
- local: model_doc/encodec
title: EnCodec
- local: model_doc/hubert
title: Hubert
- local: model_doc/mctct
Expand Down
2 changes: 2 additions & 0 deletions docs/source/en/index.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -107,6 +107,7 @@ The documentation is organized into five sections:
1. **[EfficientFormer](model_doc/efficientformer)** (from Snap Research) released with the paper [EfficientFormer: Vision Transformers at MobileNetSpeed](https://arxiv.org/abs/2206.01191) by Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, Jian Ren.
1. **[EfficientNet](model_doc/efficientnet)** (from Google Brain) released with the paper [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) by Mingxing Tan, Quoc V. Le.
1. **[ELECTRA](model_doc/electra)** (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang Luong, Quoc V. Le, Christopher D. Manning.
1. **[EnCodec](model_doc/encodec)** (from Meta AI) released with the paper [High Fidelity Neural Audio Compression](https://arxiv.org/abs/2210.13438) by Alexandre Défossez, Jade Copet, Gabriel Synnaeve, Yossi Adi.
1. **[EncoderDecoder](model_doc/encoder-decoder)** (from Google Research) released with the paper [Leveraging Pre-trained Checkpoints for Sequence Generation Tasks](https://arxiv.org/abs/1907.12461) by Sascha Rothe, Shashi Narayan, Aliaksei Severyn.
1. **[ERNIE](model_doc/ernie)** (from Baidu) released with the paper [ERNIE: Enhanced Representation through Knowledge Integration](https://arxiv.org/abs/1904.09223) by Yu Sun, Shuohuan Wang, Yukun Li, Shikun Feng, Xuyi Chen, Han Zhang, Xin Tian, Danxiang Zhu, Hao Tian, Hua Wu.
1. **[ErnieM](model_doc/ernie_m)** (from Baidu) released with the paper [ERNIE-M: Enhanced Multilingual Representation by Aligning Cross-lingual Semantics with Monolingual Corpora](https://arxiv.org/abs/2012.15674) by Xuan Ouyang, Shuohuan Wang, Chao Pang, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang.
Expand Down Expand Up @@ -318,6 +319,7 @@ Flax), PyTorch, and/or TensorFlow.
| EfficientFormer | ❌ | ❌ | ✅ | ✅ | ❌ |
| EfficientNet | ❌ | ❌ | ✅ | ❌ | ❌ |
| ELECTRA | ✅ | ✅ | ✅ | ✅ | ✅ |
| EnCodec | ❌ | ❌ | ✅ | ❌ | ❌ |
| Encoder decoder | ❌ | ❌ | ✅ | ✅ | ✅ |
| ERNIE | ❌ | ❌ | ✅ | ❌ | ❌ |
| ErnieM | ✅ | ❌ | ✅ | ❌ | ❌ |
Expand Down
Loading