Skip to content

Data2Vec

This model was released on 2022-02-07 and added to Hugging Face Transformers on 2022-03-01.

PyTorch FlashAttention SDPA

The Data2Vec model was proposed in data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language by Alexei Baevski, Wei-Ning Hsu, Qiantong Xu, Arun Babu, Jiatao Gu and Michael Auli. Data2Vec proposes a unified framework for self-supervised learning across different data modalities - text, audio and images. Importantly, predicted targets for pre-training are contextualized latent representations of the inputs, rather than modality-specific, context-independent targets.

The abstract from the paper is the following:

While the general idea of self-supervised learning is identical across modalities, the actual algorithms and objectives differ widely because they were developed with a single modality in mind. To get us closer to general self-supervised learning, we present data2vec, a framework that uses the same learning method for either speech, NLP or computer vision. The core idea is to predict latent representations of the full input data based on a masked view of the input in a selfdistillation setup using a standard Transformer architecture. Instead of predicting modality-specific targets such as words, visual tokens or units of human speech which are local in nature, data2vec predicts contextualized latent representations that contain information from the entire input. Experiments on the major benchmarks of speech recognition, image classification, and natural language understanding demonstrate a new state of the art or competitive performance to predominant approaches. Models and code are available at www.github.com/pytorch/fairseq/tree/master/examples/data2vec.

This model was contributed by edugp and patrickvonplaten.

The original code (for NLP and Speech) can be found here. The original code for vision can be found here.

  • Data2VecAudio, Data2VecText, and Data2VecVision have all been trained using the same self-supervised learning method.
  • For Data2VecAudio, preprocessing is identical to Wav2Vec2Model, including feature extraction
  • For Data2VecText, preprocessing is identical to RobertaModel, including tokenization.
  • For Data2VecVision, preprocessing is identical to BeitModel, including feature extraction.

PyTorch includes a native scaled dot-product attention (SDPA) operator as part of torch.nn.functional. This function encompasses several implementations that can be applied depending on the inputs and the hardware in use. See the official documentation or the GPU Inference page for more information.

SDPA is used by default for torch>=2.1.1 when an implementation is available, but you may also set attn_implementation="sdpa" in from_pretrained() to explicitly request SDPA to be used.

The SDPA implementation is currently available for the Data2VecAudio and Data2VecVision models.

from transformers import Data2VecVisionForImageClassification
model = Data2VecVisionForImageClassification.from_pretrained("facebook/data2vec-vision-base", attn_implementation="sdpa", dtype=torch.float16)
...

For the best speedups, we recommend loading the model in half-precision (e.g. torch.float16 or torch.bfloat16).

For the Data2VecVision model, on a local benchmark (NVIDIA GeForce RTX 2060-8GB, PyTorch 2.5.1, OS Ubuntu 20.04) with float16 and facebook/data2vec-vision-base model, we saw the following improvements during training and inference:

num_training_stepsbatch_sizeimage_sizeis_cudaTime per batch (eager - s)Time per batch (sdpa - s)Speedup (%)Eager peak mem (MB)SDPA peak mem (MB)Mem saving (%)
502(1048, 640)True0.9960.75432.1476722.1984264.65357.626
Image batch sizeEager (s/iter)Eager CI, %Eager memory (MB)SDPA (s/iter)SDPA CI, %SDPA memory (MB)SDPA speedupSDPA memory saved
10.011±0.3%3.76143e+080.01±0.3%3.74397e+081.1010.466
40.014±0.1%4.02756e+080.012±0.2%3.91373e+081.2192.909
160.046±0.3%4.96482e+080.035±0.2%4.51017e+081.31410.081
320.088±0.1%6.23903e+080.067±0.1%5.32974e+081.3317.061

A list of official Hugging Face and community (indicated by 🌎) resources to help you get started with Data2Vec.

Data2VecText documentation resources

Data2VecAudio documentation resources

Data2VecVision documentation resources

If you’re interested in submitting a resource to be included here, please feel free to open a Pull Request and we’ll review it! The resource should ideally demonstrate something new instead of duplicating an existing resource.

[[autodoc]] Data2VecTextConfig

[[autodoc]] Data2VecAudioConfig

[[autodoc]] Data2VecVisionConfig

[[autodoc]] Data2VecAudioModel - forward

[[autodoc]] Data2VecAudioForAudioFrameClassification - forward

[[autodoc]] Data2VecAudioForCTC - forward

[[autodoc]] Data2VecAudioForSequenceClassification - forward

[[autodoc]] Data2VecAudioForXVector - forward

[[autodoc]] Data2VecTextModel - forward

[[autodoc]] Data2VecTextForCausalLM - forward

[[autodoc]] Data2VecTextForMaskedLM - forward

[[autodoc]] Data2VecTextForSequenceClassification - forward

[[autodoc]] Data2VecTextForMultipleChoice - forward

[[autodoc]] Data2VecTextForTokenClassification - forward

[[autodoc]] Data2VecTextForQuestionAnswering - forward

[[autodoc]] Data2VecVisionModel - forward

[[autodoc]] Data2VecVisionForImageClassification - forward

[[autodoc]] Data2VecVisionForSemanticSegmentation - forward