mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00

* add model like clip * update * text model ok * clap text works * some refactor - `CLAPVision` to `CLAPAudio` - refactor kwargs of audio modules * more refactor * more refactor * more refactor * correct fusion * more refactor * new modules * add basic processor * fixup * remove whisper copioed from * audio logits match * add doc * correct filters mel and add maxlength * style * few fixes * forward passes * fixup * fixup * some clean up * remove mels form the dictionnary * pad after the repeat * update padding when dsmaller * fix padding * style * use swin patch merging * use copied from swin * processor with any tokenizer * more copied from * some clean up * more refactor * fix mel when rand_trunc * style * remove unused imports * update processing * remove image processing tests * add testing fiel * fixmodeling issues * replace with `is_longer` * clap in serialization * more refactor * `make fixup` * make fixup * fix feature extractor * update test feature extractor * `make fixup` * clean up config * more clean up * more cleanup * update tests * refactor tests and inits * removeCLAP vision config * remove CLAP from image procssing auto and dummy vision objects * update inits * style * re order classes in modeling clap * Use roberta tokenizer as the other weights are not open sourced * small cleaup * remove tokenization CLAP * processor tokenizr is roberta * update feature extraction doc * remove vclap from model zero shot * update f_min and f_max to frequency_xx * some changes - fix modeling keys - add `is_longer` in the forward pass - make fixup * make fixup * consistent behavior ebtween rand_crop and fusion * add numpy resize and bilinear and documentation * move resizing to image utils * clean feature extraction * import resize from correct file * resize in image transforms * update * style * style * nit * remove unused arguments form the feature extractor * style * few fixes + make fixup * oops * fix more tests * add zero shot audio classification pipeline * update zeroshot classification pipeline * fixup * fix copies * all CI tests pass * make fixup + fix docs * fix docs * fix docs * update tests pip;eline * update zero shot pipeline * update feature extraction clap * update tokenization auto * use nested simplify * update pipeline tests * Apply suggestions from code review Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * split in two lines * fixes * refactor * clean up * add integration tests * update config docstring * style * update processor * fix processor test * fix feat extractor tests * update docs * Apply suggestions from code review Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * fix readmes * fix tips * Update src/transformers/models/auto/configuration_auto.py * update doc and remove todo -> properly explained * fix idx and typo * typoe * cleanup config * cleanup tests, styles and doc * ignore docstyle on image transform * add conversion script * remove the `clap` indx in favor of `CLAP` * update __init * nits * Update src/transformers/pipelines/__init__.py * fix bug * clarifiy config * fix copy * fix init * Apply suggestions from code review Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * fix model output * fix comment * make fixup * make fixup * rename to `Clap` * replace to `Clap` * replace to `Clap` * repo consistency * again repo-consistency * make fixup * Apply suggestions from code review Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * add config * changes * update conversion * Apply suggestions from code review Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * remove unused function * update based on code reviews * style * more comments * cleanup * clean up * style * apply suggestions * Empty commit * pipeline will be added in a different PR * update calls to audio utils functions * update pipeline init * style * style * styling again * use pad * fix repo-consistency * update utils and add doc for audio utils * clean up resize by using torch. update inits accordingly * style * CLap's tokenizer is RobertA * add audio utils to internal toctreee * update totctree * style * update documentation and normalize naming accross audio utils and feature extraction clap * style * clean up * update doc and typos * fix doctest * update modelin code, got rid of a lot of reshaping * style on added doc audio utils * update modeling clap * style * Apply suggestions from code review Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * docstringvariables with CLAP * rename key * update modeling CLAP * update audio utils docstring * update processing clap * fix readmes * fix toctree * udpate configuration clap * fix init * make fixup * fix * fix * update naming * update * update checkpoint path * Apply suggestions from code review * Major refactoring * Update src/transformers/models/clap/configuration_clap.py * merge --------- Co-authored-by: younesbelkada <younesbelkada@gmail.com> Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
231 lines
12 KiB
Plaintext
231 lines
12 KiB
Plaintext
docs/source/en/quicktour.mdx
|
|
docs/source/es/quicktour.mdx
|
|
docs/source/en/pipeline_tutorial.mdx
|
|
docs/source/en/autoclass_tutorial.mdx
|
|
docs/source/en/task_summary.mdx
|
|
docs/source/en/model_doc/markuplm.mdx
|
|
docs/source/en/model_doc/speech_to_text.mdx
|
|
docs/source/en/model_doc/switch_transformers.mdx
|
|
docs/source/en/model_doc/t5.mdx
|
|
docs/source/en/model_doc/t5v1.1.mdx
|
|
docs/source/en/model_doc/byt5.mdx
|
|
docs/source/en/model_doc/tapex.mdx
|
|
docs/source/en/model_doc/donut.mdx
|
|
docs/source/en/model_doc/encoder-decoder.mdx
|
|
src/transformers/generation/configuration_utils.py
|
|
src/transformers/generation/tf_utils.py
|
|
src/transformers/generation/utils.py
|
|
src/transformers/models/albert/configuration_albert.py
|
|
src/transformers/models/albert/modeling_albert.py
|
|
src/transformers/models/albert/modeling_tf_albert.py
|
|
src/transformers/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py
|
|
src/transformers/models/bart/configuration_bart.py
|
|
src/transformers/models/bart/modeling_bart.py
|
|
src/transformers/models/beit/configuration_beit.py
|
|
src/transformers/models/beit/modeling_beit.py
|
|
src/transformers/models/bert/configuration_bert.py
|
|
src/transformers/models/bert/modeling_bert.py
|
|
src/transformers/models/bert/modeling_tf_bert.py
|
|
src/transformers/models/bert_generation/configuration_bert_generation.py
|
|
src/transformers/models/bigbird_pegasus/configuration_bigbird_pegasus.py
|
|
src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py
|
|
src/transformers/models/big_bird/configuration_big_bird.py
|
|
src/transformers/models/big_bird/modeling_big_bird.py
|
|
src/transformers/models/blenderbot/configuration_blenderbot.py
|
|
src/transformers/models/blenderbot/modeling_blenderbot.py
|
|
src/transformers/models/blenderbot_small/configuration_blenderbot_small.py
|
|
src/transformers/models/blenderbot_small/modeling_blenderbot_small.py
|
|
src/transformers/models/blip/modeling_blip.py
|
|
src/transformers/models/bloom/configuration_bloom.py
|
|
src/transformers/models/camembert/configuration_camembert.py
|
|
src/transformers/models/canine/configuration_canine.py
|
|
src/transformers/models/canine/modeling_canine.py
|
|
src/transformers/models/clap/configuration_clap.py
|
|
src/transformers/models/clap/modeling_clap.py
|
|
src/transformers/models/clip/configuration_clip.py
|
|
src/transformers/models/clipseg/modeling_clipseg.py
|
|
src/transformers/models/codegen/configuration_codegen.py
|
|
src/transformers/models/conditional_detr/configuration_conditional_detr.py
|
|
src/transformers/models/conditional_detr/modeling_conditional_detr.py
|
|
src/transformers/models/convbert/configuration_convbert.py
|
|
src/transformers/models/convnext/configuration_convnext.py
|
|
src/transformers/models/convnext/modeling_convnext.py
|
|
src/transformers/models/ctrl/configuration_ctrl.py
|
|
src/transformers/models/ctrl/modeling_ctrl.py
|
|
src/transformers/models/cvt/configuration_cvt.py
|
|
src/transformers/models/cvt/modeling_cvt.py
|
|
src/transformers/models/data2vec/configuration_data2vec_audio.py
|
|
src/transformers/models/data2vec/configuration_data2vec_text.py
|
|
src/transformers/models/data2vec/configuration_data2vec_vision.py
|
|
src/transformers/models/data2vec/modeling_data2vec_audio.py
|
|
src/transformers/models/data2vec/modeling_data2vec_vision.py
|
|
src/transformers/models/deberta/configuration_deberta.py
|
|
src/transformers/models/deberta/modeling_deberta.py
|
|
src/transformers/models/deberta_v2/configuration_deberta_v2.py
|
|
src/transformers/models/deberta_v2/modeling_deberta_v2.py
|
|
src/transformers/models/decision_transformer/configuration_decision_transformer.py
|
|
src/transformers/models/deformable_detr/configuration_deformable_detr.py
|
|
src/transformers/models/deformable_detr/modeling_deformable_detr.py
|
|
src/transformers/models/deit/configuration_deit.py
|
|
src/transformers/models/deit/modeling_deit.py
|
|
src/transformers/models/deit/modeling_tf_deit.py
|
|
src/transformers/models/deta/configuration_deta.py
|
|
src/transformers/models/deta/modeling_deta.py
|
|
src/transformers/models/detr/configuration_detr.py
|
|
src/transformers/models/detr/modeling_detr.py
|
|
src/transformers/models/dinat/configuration_dinat.py
|
|
src/transformers/models/dinat/modeling_dinat.py
|
|
src/transformers/models/distilbert/configuration_distilbert.py
|
|
src/transformers/models/dpr/configuration_dpr.py
|
|
src/transformers/models/dpt/modeling_dpt.py
|
|
src/transformers/models/electra/configuration_electra.py
|
|
src/transformers/models/electra/modeling_electra.py
|
|
src/transformers/models/electra/modeling_tf_electra.py
|
|
src/transformers/models/ernie/configuration_ernie.py
|
|
src/transformers/models/ernie_m/configuration_ernie_m.py
|
|
src/transformers/models/ernie_m/modeling_ernie_m.py
|
|
src/transformers/models/flava/configuration_flava.py
|
|
src/transformers/models/fnet/configuration_fnet.py
|
|
src/transformers/models/fsmt/configuration_fsmt.py
|
|
src/transformers/models/git/modeling_git.py
|
|
src/transformers/models/glpn/modeling_glpn.py
|
|
src/transformers/models/gpt2/configuration_gpt2.py
|
|
src/transformers/models/gpt2/modeling_gpt2.py
|
|
src/transformers/models/gptj/modeling_gptj.py
|
|
src/transformers/models/gpt_neo/configuration_gpt_neo.py
|
|
src/transformers/models/gpt_neox/configuration_gpt_neox.py
|
|
src/transformers/models/gpt_neox_japanese/configuration_gpt_neox_japanese.py
|
|
src/transformers/models/groupvit/modeling_groupvit.py
|
|
src/transformers/models/groupvit/modeling_tf_groupvit.py
|
|
src/transformers/models/hubert/modeling_hubert.py
|
|
src/transformers/models/imagegpt/configuration_imagegpt.py
|
|
src/transformers/models/layoutlm/configuration_layoutlm.py
|
|
src/transformers/models/layoutlm/modeling_layoutlm.py
|
|
src/transformers/models/layoutlm/modeling_tf_layoutlm.py
|
|
src/transformers/models/layoutlmv2/configuration_layoutlmv2.py
|
|
src/transformers/models/layoutlmv2/modeling_layoutlmv2.py
|
|
src/transformers/models/layoutlmv3/configuration_layoutlmv3.py
|
|
src/transformers/models/layoutlmv3/modeling_layoutlmv3.py
|
|
src/transformers/models/layoutlmv3/modeling_tf_layoutlmv3.py
|
|
src/transformers/models/levit/configuration_levit.py
|
|
src/transformers/models/lilt/modeling_lilt.py
|
|
src/transformers/models/longformer/modeling_longformer.py
|
|
src/transformers/models/longformer/modeling_tf_longformer.py
|
|
src/transformers/models/longt5/modeling_longt5.py
|
|
src/transformers/models/marian/modeling_marian.py
|
|
src/transformers/models/markuplm/modeling_markuplm.py
|
|
src/transformers/models/mask2former/configuration_mask2former.py
|
|
src/transformers/models/mask2former/modeling_mask2former.py
|
|
src/transformers/models/maskformer/configuration_maskformer.py
|
|
src/transformers/models/maskformer/modeling_maskformer.py
|
|
src/transformers/models/mbart/configuration_mbart.py
|
|
src/transformers/models/mbart/modeling_mbart.py
|
|
src/transformers/models/mctct/configuration_mctct.py
|
|
src/transformers/models/megatron_bert/configuration_megatron_bert.py
|
|
src/transformers/models/mobilebert/configuration_mobilebert.py
|
|
src/transformers/models/mobilebert/modeling_mobilebert.py
|
|
src/transformers/models/mobilebert/modeling_tf_mobilebert.py
|
|
src/transformers/models/mobilenet_v1/modeling_mobilenet_v1.py
|
|
src/transformers/models/mobilenet_v2/modeling_mobilenet_v2.py
|
|
src/transformers/models/mobilevit/modeling_mobilevit.py
|
|
src/transformers/models/mobilevit/modeling_tf_mobilevit.py
|
|
src/transformers/models/nat/configuration_nat.py
|
|
src/transformers/models/nat/modeling_nat.py
|
|
src/transformers/models/nezha/configuration_nezha.py
|
|
src/transformers/models/oneformer/configuration_oneformer.py
|
|
src/transformers/models/oneformer/modeling_oneformer.py
|
|
src/transformers/models/openai/configuration_openai.py
|
|
src/transformers/models/opt/configuration_opt.py
|
|
src/transformers/models/opt/modeling_opt.py
|
|
src/transformers/models/opt/modeling_tf_opt.py
|
|
src/transformers/models/owlvit/modeling_owlvit.py
|
|
src/transformers/models/pegasus/configuration_pegasus.py
|
|
src/transformers/models/pegasus/modeling_pegasus.py
|
|
src/transformers/models/pegasus_x/configuration_pegasus_x.py
|
|
src/transformers/models/perceiver/modeling_perceiver.py
|
|
src/transformers/models/plbart/configuration_plbart.py
|
|
src/transformers/models/plbart/modeling_plbart.py
|
|
src/transformers/models/poolformer/configuration_poolformer.py
|
|
src/transformers/models/poolformer/modeling_poolformer.py
|
|
src/transformers/models/realm/configuration_realm.py
|
|
src/transformers/models/reformer/configuration_reformer.py
|
|
src/transformers/models/reformer/modeling_reformer.py
|
|
src/transformers/models/regnet/modeling_regnet.py
|
|
src/transformers/models/regnet/modeling_tf_regnet.py
|
|
src/transformers/models/resnet/configuration_resnet.py
|
|
src/transformers/models/resnet/modeling_resnet.py
|
|
src/transformers/models/resnet/modeling_tf_resnet.py
|
|
src/transformers/models/roberta/configuration_roberta.py
|
|
src/transformers/models/roberta/modeling_roberta.py
|
|
src/transformers/models/roberta/modeling_tf_roberta.py
|
|
src/transformers/models/roberta_prelayernorm/configuration_roberta_prelayernorm.py
|
|
src/transformers/models/roberta_prelayernorm/modeling_roberta_prelayernorm.py
|
|
src/transformers/models/roberta_prelayernorm/modeling_tf_roberta_prelayernorm.py
|
|
src/transformers/models/roc_bert/modeling_roc_bert.py
|
|
src/transformers/models/roc_bert/tokenization_roc_bert.py
|
|
src/transformers/models/segformer/modeling_segformer.py
|
|
src/transformers/models/sew/configuration_sew.py
|
|
src/transformers/models/sew/modeling_sew.py
|
|
src/transformers/models/sew_d/configuration_sew_d.py
|
|
src/transformers/models/sew_d/modeling_sew_d.py
|
|
src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py
|
|
src/transformers/models/speech_to_text/configuration_speech_to_text.py
|
|
src/transformers/models/speech_to_text/modeling_speech_to_text.py
|
|
src/transformers/models/speech_to_text_2/configuration_speech_to_text_2.py
|
|
src/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py
|
|
src/transformers/models/speecht5/modeling_speecht5.py
|
|
src/transformers/models/speecht5/tokenization_speecht5.py
|
|
src/transformers/models/segformer/modeling_tf_segformer.py
|
|
src/transformers/models/squeezebert/configuration_squeezebert.py
|
|
src/transformers/models/swin/configuration_swin.py
|
|
src/transformers/models/swin/modeling_swin.py
|
|
src/transformers/models/swin2sr/modeling_swin2sr.py
|
|
src/transformers/models/swinv2/configuration_swinv2.py
|
|
src/transformers/models/table_transformer/modeling_table_transformer.py
|
|
src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
|
|
src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
|
|
src/transformers/models/trajectory_transformer/configuration_trajectory_transformer.py
|
|
src/transformers/models/transfo_xl/configuration_transfo_xl.py
|
|
src/transformers/models/trocr/configuration_trocr.py
|
|
src/transformers/models/trocr/modeling_trocr.py
|
|
src/transformers/models/unispeech/configuration_unispeech.py
|
|
src/transformers/models/unispeech/modeling_unispeech.py
|
|
src/transformers/models/unispeech_sat/modeling_unispeech_sat.py
|
|
src/transformers/models/upernet/modeling_upernet.py
|
|
src/transformers/models/van/modeling_van.py
|
|
src/transformers/models/videomae/modeling_videomae.py
|
|
src/transformers/models/vilt/modeling_vilt.py
|
|
src/transformers/models/vision_encoder_decoder/configuration_vision_encoder_decoder.py
|
|
src/transformers/models/vision_encoder_decoder/modeling_vision_encoder_decoder.py
|
|
src/transformers/models/vision_text_dual_encoder/configuration_vision_text_dual_encoder.py
|
|
src/transformers/models/vit/configuration_vit.py
|
|
src/transformers/models/vit/modeling_vit.py
|
|
src/transformers/models/vit/modeling_tf_vit.py
|
|
src/transformers/models/vit_mae/modeling_vit_mae.py
|
|
src/transformers/models/vit_mae/configuration_vit_mae.py
|
|
src/transformers/models/vit_msn/modeling_vit_msn.py
|
|
src/transformers/models/visual_bert/configuration_visual_bert.py
|
|
src/transformers/models/wav2vec2/configuration_wav2vec2.py
|
|
src/transformers/models/wav2vec2/modeling_wav2vec2.py
|
|
src/transformers/models/wav2vec2/tokenization_wav2vec2.py
|
|
src/transformers/models/wav2vec2_conformer/configuration_wav2vec2_conformer.py
|
|
src/transformers/models/wav2vec2_conformer/modeling_wav2vec2_conformer.py
|
|
src/transformers/models/wav2vec2_with_lm/processing_wav2vec2_with_lm.py
|
|
src/transformers/models/wavlm/configuration_wavlm.py
|
|
src/transformers/models/wavlm/modeling_wavlm.py
|
|
src/transformers/models/whisper/configuration_whisper.py
|
|
src/transformers/models/whisper/modeling_whisper.py
|
|
src/transformers/models/whisper/modeling_tf_whisper.py
|
|
src/transformers/models/xlm/configuration_xlm.py
|
|
src/transformers/models/xlm_roberta/configuration_xlm_roberta.py
|
|
src/transformers/models/xlm_roberta_xl/configuration_xlm_roberta_xl.py
|
|
src/transformers/models/xlnet/configuration_xlnet.py
|
|
src/transformers/models/xmod/configuration_xmod.py
|
|
src/transformers/models/xmod/modeling_xmod.py
|
|
src/transformers/models/yolos/configuration_yolos.py
|
|
src/transformers/models/yolos/modeling_yolos.py
|
|
src/transformers/models/x_clip/modeling_x_clip.py
|
|
src/transformers/models/yoso/configuration_yoso.py
|
|
src/transformers/models/timesformer/configuration_timesformer.py
|
|
src/transformers/pipelines/
|