This is a limited proof of concept to search for research data, not a production system.

Search the MIT Libraries

Title: Transformers: State-of-the-Art Natural Language Processing

Type Software Wolf, Thomas, Debut, Lysandre, Sanh, Victor, Chaumond, Julien, Delangue, Clement, Moi, Anthony, Cistac, Perric, Ma, Clara, Jernite, Yacine, Plu, Julien, Xu, Canwen, Le Scao, Teven, Gugger, Sylvain, Drame, Mariama, Lhoest, Quentin, Rush, Alexander M. (2020): Transformers: State-of-the-Art Natural Language Processing. Zenodo. Software. https://zenodo.org/record/7271740

Authors: Wolf, Thomas ; Debut, Lysandre ; Sanh, Victor ; Chaumond, Julien ; Delangue, Clement ; Moi, Anthony ; Cistac, Perric ; Ma, Clara ; Jernite, Yacine ; Plu, Julien ; Xu, Canwen ; Le Scao, Teven ; Gugger, Sylvain ; Drame, Mariama ; Lhoest, Quentin ; Rush, Alexander M. ;

Links

Summary

ESM-2/ESMFold

ESM-2 and ESMFold are new state-of-the-art Transformer protein language and folding models from Meta AI's Fundamental AI Research Team (FAIR). ESM-2 is trained with a masked language modeling objective, and it can be easily transferred to sequence and token classification tasks for proteins. Checkpoints exist in various sizes, from 8 million parameters up to a huge 15 billion parameter model.

ESMFold is a state-of-the-art single sequence protein folding model which produces high accuracy predictions significantly faster. Unlike previous protein folding tools like AlphaFold2 and openfold, ESMFold uses a pretrained protein language model to generate token embeddings that are used as input to the folding model, and so does not require a multiple sequence alignment (MSA) of related proteins as input. As a result, proteins can be folded in a single forward pass of the model without requiring any external databases or search/alignment tools to be present at inference time. This hugely reduces the time and compute requirements for folding.

Transformer protein language models were introduced in the paper Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences by Alexander Rives, Joshua Meier, Tom Sercu, Siddharth Goyal, Zeming Lin, Jason Liu, Demi Guo, Myle Ott, C. Lawrence Zitnick, Jerry Ma, and Rob Fergus.

ESMFold was introduced in the paper Language models of protein sequences at the scale of evolution enable accurate structure prediction by Zeming Lin, Halil Akin, Roshan Rao, Brian Hie, Zhongkai Zhu, Wenting Lu, Allan dos Santos Costa, Maryam Fazel-Zarandi, Tom Sercu, Sal Candido, and Alexander Rives.

Add ESMFold by @Rocketknight1 in #19977 TF port of ESM by @Rocketknight1 in #19587 LiLT

LiLT allows to combine any pre-trained RoBERTa text encoder with a lightweight Layout Transformer, to enable LayoutLM-like document understanding for many languages.

It was proposed in LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding by Jiapeng Wang, Lianwen Jin, Kai Ding.

Add LiLT by @NielsRogge in #19450 Flan-T5

FLAN-T5 is an enhanced version of T5 that has been finetuned on a mixture of tasks.

It was released in the paper Scaling Instruction-Finetuned Language Models by Hyung Won Chung, Le Hou, Shayne Longpre, Barret Zoph, Yi Tay, William Fedus, Eric Li, Xuezhi Wang, Mostafa Dehghani, Siddhartha Brahma, Albert Webson, Shixiang Shane Gu, Zhuyun Dai, Mirac Suzgun, Xinyun Chen, Aakanksha Chowdhery, Sharan Narang, Gaurav Mishra, Adams Yu, Vincent Zhao, Yanping Huang, Andrew Dai, Hongkun Yu, Slav Petrov, Ed H. Chi, Jeff Dean, Jacob Devlin, Adam Roberts, Denny Zhou, Quoc V. Le, and Jason Wei.

Add flan-t5 documentation page by @younesbelkada in #19892 Table Transformer

Table Transformer is a model that can perform table extraction and table structure recognition from unstructured documents based on the DETR architecture.

It was proposed in PubTables-1M: Towards comprehensive table extraction from unstructured documents by Brandon Smock, Rohith Pesala, Robin Abraham.

Add table transformer [v2] by @NielsRogge in #19614 Contrastive search decoding

Contrastive search decoding is a new state-of-the-art generation method which aims at reducing the repetitive patterns in which generation models often fall.

It was introduced in A Contrastive Framework for Neural Text Generation by Yixuan Su, Tian Lan, Yan Wang, Dani Yogatama, Lingpeng Kong, Nigel Collier.

Adding the state-of-the-art contrastive search decoding methods for the codebase of generation_utils.py by @gmftbyGMFTBY in #19477 Safety and security

We continue to explore the new serialization format not using Pickle via the safetensors library, this time by adding support for TensorFlow models. More checkpoints have been converted to this format. Support is still experimental.

Safetensors tf by @sgugger in #19900 🚨 Breaking changes

The following changes are bugfixes that we have chosen to fix even if it changes the resulting behavior. We mark them as breaking changes, so if you are using this part of the codebase, we recommend you take a look at the PRs to understand what changes were done exactly.

🚨🚨🚨 TF: Remove TFWrappedEmbeddings (breaking: TF embedding initialization updated for encoder-decoder models) by @gante in #19263 🚨🚨🚨 [Breaking change] Deformable DETR intermediate representations by @Narsil in #19678 Bugfixes and improvements Enabling custom TF signature draft by @dimitreOliveira in #19249 Fix whisper for pipeline by @ArthurZucker in #19482 Extend nested_XXX functions to mappings/dicts. by @Guillem96 in #19455 Syntax issues (lines 126, 203) by @kant in #19444 CLI: add import protection to datasets by @gante in #19470 Fix TFGroupViT CI by @ydshieh in #19461 Fix doctests for DeiT and TFGroupViT by @ydshieh in #19466 Update WhisperModelIntegrationTests.test_large_batched_generation by @ydshieh in #19472 [Swin] Replace hard-coded batch size to enable dynamic ONNX export by @lewtun in #19475 TF: TFBart embedding initialization by @gante in #19460 Make LayoutLM tokenizers independent from BertTokenizer by @arnaudstiegler in #19351 Make XLMRoberta model and config independent from Roberta by @asofiaoliveira in #19359 Fix get_embedding dtype at init. time by @ydshieh in #19473 Decouples XLMProphet model from Prophet by @srhrshr in #19406 Implement multiple span support for DocumentQuestionAnswering by @ankrgyl in #19204 Add warning in generate & device_map=auto & half precision models by @younesbelkada in #19468 Update TF whisper doc tests by @amyeroberts in #19484 Make bert_japanese and cpm independent of their inherited modules by @Davidy22 in #19431 Added tokenize keyword arguments to feature extraction pipeline by @quancore in #19382 Adding the README_es.md and reference to it in the others files readme by @Oussamaosman02 in #19427 [CvT] Tensorflow implementation by @mathieujouffroy in #18597 python3 instead of python in push CI setup job by @ydshieh in #19492 Update PT to TF CLI for audio models by @amyeroberts in #19465 New by @IMvision12 in #19481 Fix OPTForQuestionAnswering doctest by @ydshieh in #19479 Use a dynamic configuration for circleCI tests by @sgugger in #19325 Add multi-node conditions in trainer_qa.py and trainer_seq2seq.py by @regisss in #19502 update doc for perf_train_cpu_many by @sywangyi in #19506 Avoid Push CI failing to report due to many commits being merged by @ydshieh in #19496 [Doctest] Add configuration_bert.py to doctest by @ydshieh in #19485 Fix whisper doc by @ArthurZucker in #19518 Syntax issue (line 497, 526) Documentation by @kant in #19442 Fix pytorch seq2seq qa by @FilipposVentirozos in #19258 Add depth estimation pipeline by @nandwalritik in #18618 Adding links to pipelines parameters documentation by @AndreaSottana in #19227 fix MarkupLMProcessor option flag by @davanstrien in #19526 [Doctest] Bart configuration update by @imarekkus in #19524 Remove roberta dependency from longformer fast tokenizer by @sirmammingtonham in #19501 made tokenization_roformer independent of bert by @naveennamani in #19426 Remove bert fast dependency from electra by @Threepointone4 in #19520 [Examples] Fix typos in run speech recognition seq2seq by @sanchit-gandhi in #19514 [X-CLIP] Fix doc tests by @NielsRogge in #19523 Update Marian config default vocabulary size by @gante in #19464 Make MobileBert tokenizers independent from Bert by @501Good in #19531 [Whisper] Fix gradient checkpointing by @sanchit-gandhi in #19538 Syntax issues (paragraphs 122, 130, 147, 155) Documentation: @sgugger by @kant in #19437 using trunc_normal for weight init & cls_token by @mathieujouffroy in #19486 Remove MarkupLMForMaskedLM from MODEL_WITH_LM_HEAD_MAPPING_NAMES by @ydshieh in #19534 Image transforms library by @amyeroberts in #18520 Add a decorator for flaky tests by @sgugger in #19498 [Doctest] Add configuration_yolos.py by @daspartho in #19539 Albert config update by @imarekkus in #19541 [Doctest] Add configuration_whisper.py by @daspartho in #19540 Throw an error if getattribute_from_module can't find anything by @ydshieh in #19535 [Doctest] Beit Config for doctest by @daspartho in #19542 Create the arange tensor on device for enabling CUDA-Graph for Clip Encoder by @RezaYazdaniAminabadi in #19503 [Doctest] GPT2 Config for doctest by @daspartho in #19549 Build Push CI images also in a daily basis by @ydshieh in #19532 Fix checkpoint used in MarkupLMConfig by @ydshieh in #19547 add a note to whisper docs clarifying support of long-form decoding by @akashmjn in #19497 [Whisper] Freeze params of encoder by @sanchit-gandhi in #19527 [Doctest] Fixing the Doctest for imageGPT config by @RamitPahwa in #19556 [Doctest] Fixing mobile bert configuration doctest by @RamitPahwa in #19557 [Doctest] Fixing doctest bert_generation configuration by @Threepointone4 in #19558 [Doctest] DeiT Config for doctest by @daspartho in #19560 [Doctest] Reformer Config for doctest by @daspartho in #19562 [Doctest] RoBERTa Config for doctest by @daspartho in #19563 [Doctest] Add configuration_vit.py by @daspartho in #19561 [Doctest] bloom config update by @imarekkus in #19566 [Re-submit] Compute true loss Flax examples by @duongna21 in #19504 Fix fairseq wav2vec2-xls-r pretrained weights conversion scripts by @heatz123 in #19508 [Doctest] CTRL config by @imarekkus in #19574 [Doctest] Add configuration_canine.py by @IzicTemi in #19575 [Doctests] Config files for ViTMAE and YOSO by @grgkaran03 in #19567 Added type hints to DebertaV2ForMultipleChoice Pytorch by @IMvision12 in #19536 [WIP] Add type hints for Lxmert (TF) by @elusenji in #19441 [Doctests] add configuration_blenderbot.py by @grgkaran03 in #19577 [Doctest] adds trajectory_transformer config to Docs test by @SD-13 in #19586 [Doctests] add configuration_blenderbot_small.py by @grgkaran03 in #19589 [Doctest] Swin V2 Config for doctest by @daspartho in #19595 [Doctest] Swin Config for doctest by @daspartho in #19594 [Doctest] SEW Config for doctest by @daspartho in #19597 [Doctest] UniSpeech Config for doctest by @daspartho in #19596 [Doctest] SEW-D Config for doctest by @daspartho in #19598 [Doctest] fix doc test for megatron bert by @RamitPahwa in #19600 Adding type hints for TFXLnet by @thliang01 in #19344 [Doctest] Add configuration_bigbird_pegasus.py and configuration_big_bird.py by @Xabilahu in #19606 Cast masks to np.unit8 before converting to PIL.Image.Image by @amyeroberts in #19616 [Whisper] Don't return attention mask in feat extractor by @sanchit-gandhi in #19521 [Time Series Transformer] Add doc tests by @NielsRogge in #19607 fix BLOOM ONNX config by @NouamaneTazi in #19573 Fix test_tf_encode_plus_sent_to_model for TAPAS by @ydshieh in #19559 Allow usage of TF Text BertTokenizer on TFBertTokenizer to make it servable on TF Serving by @piEsposito in #19590 add gloo backend support for CPU DDP by @sywangyi in #19555 Fix ImageToTextPipelineTests.test_small_model_tf by @ydshieh in #19565 Fix FlaubertTokenizer by @ydshieh in #19552 Visual Bert config for doctest by @ztjhz in #19605 GPTTokenizer dependency removed from deberta class by @RamitPahwa in #19551 xlm roberta config for doctest by @ztjhz in #19609 Ernie config for doctest by @ztjhz in #19611 xlm roberta xl config for doctest by @ztjhz in #19610 fix: small error by @0xflotus in #19612 Improve error messaging for ASR pipeline. by @Narsil in #19570 [Doctest] LeViT Config for doctest by @daspartho in #19622 [Doctest] DistilBERT Config for doctest by @daspartho in #19621 [Whisper] Fix gradient checkpointing (again!) by @sanchit-gandhi in #19548 [Doctest] Add configuration_resnet.py by @daspartho in #19620 Fix whisper doc by @ArthurZucker in #19608 Sharding fails in TF when absolute scope was modified if . in layer name by @ArthurZucker in #19124 [Doctest] Add configuration_vision_text_dual_encoder.py by @SD-13 in #19580 [Doctest] Add configuration_vision_encoder_decoder.py by @SD-13 in #19583 [Doctest] Add configuration_time_series_transformer.py by @SD-13 in #19582 Tokenizer from_pretrained should not use local files named like tokenizer files by @sgugger in #19626 [Doctest] CodeGen config for doctest by @AymenBer99 in #19633 [Doctest] Add configuration_data2vec_text.py by @daspartho in #19636 [Doctest] Conditional DETR config for doctest by @AymenBer99 in #19641 [Doctest] XLNet config for doctest by @AymenBer99 in #19649 [Doctest] Add configuration_trocr.py by @thliang01 in #19658 Add doctest info in testingmdx by @ArthurZucker in #19623 Add pillow to layoutlmv3 example requirements.txt by @Spacefish in #19663 add return types for tf gptj, xlm, and xlnet by @sirmammingtonham in #19638 Fix pipeline predict transform methods by @s-udhaya in #19657 Type hints MCTCT by @rchan26 in #19618 added type hints for Yolos Pytorch model by @WhiteWolf47 in #19545 A few CI fixes for DocumentQuestionAnsweringPipeline by @ankrgyl in #19584 Removed Bert interdependency from Funnel transformer by @mukesh663 in #19655 fix warnings in deberta by @sanderland in #19458 word replacement line #231 by @shreem-123 in #19662 [Doctest] Add configuration_transfo_xl.py by @thliang01 in #19651 Update perf_train_gpu_one.mdx by @cakiki in #19676 object-detection instead of object_detection by @Spacefish in #19677 add return_tensor parameter for feature extraction by @ajsanjoaquin in #19257 Fix code examples of DETR and YOLOS by @NielsRogge in #19669 Revert "add return_tensor parameter for feature extraction by @sgugger in #19257)" Fixed the docstring and type hint for forced_decoder_ids option in Ge… by @koreyou in #19640 Add normalize to image transforms module by @amyeroberts in #19544 [Doctest] Data2VecAudio Config for doctest by @daspartho in #19635 Update ESM checkpoints to point to facebook/ by @Rocketknight1 in #19675 Removed XLMModel inheritance from FlaubertModel(torch+tf) by @D3xter1922 in #19432 [Examples] make default preprocessing_num_workers=1 by @Yang-YiFan in #19684 [Doctest] Add configuration_convbert.py by @AymenBer99 in #19643 [Doctest] Add configuration_realm.py by @ak04p in #19646 Update CONTRIBUTING.md by @shreem-123 in #19689 [Doctest] Add configuration_data2vec_vision.py by @daspartho in #19637 Fix some CI torch device issues for PyTorch 1.13 by @ydshieh in #19681 Fix checkpoint used in VisualBertConfig doc example by @ydshieh in #19692 Fix dtype in radnomly initialized head by @sgugger in #19690 fix tests by @ArthurZucker in #19670 fix test whisper with new max length by @ArthurZucker in #19668 check decoder_inputs_embeds is None before shifting labels by @ArthurZucker in #19671 Fix docs by @NielsRogge in #19687 update documentation by @ArthurZucker in #19706 Improve DETR models by @NielsRogge in #19644 Small fixes for TF-ESM1b and ESM-1b weight conversions by @Rocketknight1 in #19683 Fix typo in perf docs by @cakiki in #19705 Fix redundant normalization of OWL-ViT text embeddings by @alaradirik in #19712 Allow user-managed Pool in Wav2Vec2ProcessorWithLM.batch_decode by @falcaopetri in #18351 [Doctest] CVT config for doctest by @AymenBer99 in #19695 [Doctest] Add configuration_wav2vec2.py to documentation_tests.py by @juancopi81 in #19698 ]Fixed pegasus config doctest by @mukesh663 in #19722 fix seq2seqtrainer predict without labels by @IvanSedykh in #19721 add return_tensors parameter for feature_extraction 2 by @Narsil in #19707 Improving image-segmentation pipeline tests. by @Narsil in #19710 [Doctest] Adding config files for convnext by @soma2000-lang in #19717 [Doctest] Fixing doctest configuration_pegasus_x.py by @mukesh663 in #19725 Specify TF framework in TF-related pipeline tests by @ydshieh in #19719 Add docs by @NielsRogge in #19729 Fix activations being all the same module by @sgugger in #19728 add accelerate support for Whisper by @younesbelkada in #19697 Clean up deprecation warnings by @Davidy22 in #19654 Repo utils test by @sgugger in #19696 Add decorator to flaky test by @amyeroberts in #19674 [Doctest] Add doctest for FlavaConfig and FNetConfig by @ndrohith09 in #19724 Update contribution guide by @stevhliu in #19700 [Doctest] Add wav2vec2_conformer for doctest by @juancopi81 in #19734 [Doctest] XLM Config for doctest by @AymenBer99 in #19685 [Doctest] Add configuration_clip.py by @daspartho in #19647 [Doctest] GPTNeoConfig , GPTNeoXConfig , GPTNeoXJapaneseConfig by @ndrohith09 in #19741 Update modeling_markuplm.py by @IMvision12 in #19723 Fix issue #19300 by @raghavanone in #19483 [Doctest] Add configuration_wavlm.py by @juancopi81 in #19749 Specify TF framework explicitly in more pipeline tests by @ydshieh in #19748 Fix cache version file creation by @sgugger in #19750

More information

  • DOI: 10.5281/zenodo.7271740

Dates

  • Publication date: 2020
  • Issued: October 01, 2020

Notes

Other: If you use this software, please cite it using these metadata.

Rights

  • info:eu-repo/semantics/openAccess Open Access

Much of the data past this point we don't have good examples of yet. Please share in #rdi slack if you have good examples for anything that appears below. Thanks!

Format

electronic resource

Relateditems

DescriptionItem typeRelationshipUri
IsSupplementTohttps://github.com/huggingface/transformers/tree/v4.24.0
IsVersionOfhttps://doi.org/10.5281/zenodo.3385997
IsPartOfhttps://zenodo.org/communities/zenodo