-
Notifications
You must be signed in to change notification settings - Fork 229
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[WIP] Adding OBELICS DataLoader #663
base: main
Are you sure you want to change the base?
Conversation
Hi @TJ-Solergibert! Thank you for your pull request and welcome to our community. Action RequiredIn order to merge any pull request (code, docs, etc.), we require contributors to sign our Contributor License Agreement, and we don't seem to have one on file for you. ProcessIn order for us to review and merge your suggested changes, please sign at https://code.facebook.com/cla. If you are contributing on behalf of someone else (eg your employer), the individual CLA may not be sufficient and your employer may need to sign the corporate CLA. Once the CLA is signed, our tooling will perform checks and validations. Afterwards, the pull request will be tagged with If you have received this in error or have any questions, please contact us at [email protected]. Thanks! |
Thank you for signing our Contributor License Agreement. We can now accept your code for this (and any) Meta Open Source project. Thanks! |
|
||
import torch | ||
|
||
import torchvision |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is just a quick scan for this PR. I need more time to look into this PR. But we should add torchvision as the dependent of Torchtitan, if haven't.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes! Don't worry, it will take time 😅
BATCH_NUMBER = 4 | ||
|
||
|
||
def main(): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
maybe we can make this as a unit test? WDYT?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would add as a unit test some checks of shapes & types on the DP axis rather than this script that just checks the amount of padding in each batch
text, bos=True, eos=True, allowed_special=set(["<|image|>"]) | ||
) | ||
input_ids = tokens[:-1] | ||
labels = tokens[1:] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
what if the first token is null? And for HuggingFaceM4/OBELICS
specifically, "text" is a list of string or null. We don't need special treatment of it?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We take care of the null/None values in the format_obelics
function.
This function produces:
images
: List of decoded images in the sampletext
: str with the text of the sample ready to be tokenized including the image tokens (the null values you are referring)
So the text
is ready to be tokenized!
Hi,
In this PR I present a first draft of the Multimodal DataLoader. First I will describe how the batches are created and then I will explain the padding problem.
Let's begin checking the OBELICS dataset. For every sample on the dataset we have 4 keys, but we are just interested in 2 of them:
images
: A list either with URLs of images ORNone
s to specify the position of the text.texts
: A list either with text strings ORNone
s to specify the position of the images.It's important to highlight that
len(images)==len(texts)
and that for each index, one element and only one is notNone
.The
format_obelics
function will transform each sample to a format that can be later fed into the transform block that will prepare the samples to the target type. Each formatted sample will be a dictionary containing 2 keys:images
:List
of PIL Images with the loaded images.text
:str
with the text of the sample ready to be tokenized, including the image tokens.Once formatted, we will process each sample with the transform block. This transform block is composed of
CLIPPreprocess
,TikTokenizer
&VisionCrossAttentionMask
modules.CLIPPreprocess
This module will prepare the List of images to be fed into the CLIP model. The most relevant steps is resizing the image without distortion, dividing the image into tiles and padding if necessary. Highlight the fact that it will still produce a List of tensors and NOT a tensor as every image can have a different number of tiles. This will be addressed in the collator where we will pad the image tiles to the largest in the batch. Also, we keep the maximum number of tiles to 4 and the tile size to 448 for pretraining [1], [2].
TikTokenizer
I've included a new method in the tokenizer to encode the multimodal text. In short, it just encodes the text adding the special
image_id
token and returns both theinput_ids
&labels
masking thebos
,eos
&image_id
tokens.VisionCrossAttentionMask
This module will create the attention mask for the Fused layers. In short, for each TILE we will have 1025
image_tokens
and this mask will specify for eachtext_token
to whichimage_tokens
should attend to. We are returning again a List of tensors as the quantity ofimage_tokens
will depend on the number of tiles. Again, we will solve this in the collator.Padding & the collator
As we've previously seen, both the outputs of the
CLIPPreprocess
&VisionCrossAttentionMask
are list of tensors because of the different number of tiles. Within the same sample we should pad both artifacts to the maximum number of tiles, but the issue arises when we runbatch_size > 1
as we will also need to pad theinput_ids
(&labels
) which is relatively cheap BUT also the Number of images, as the input to the CLIP model will be a tensor of shape [Batch size, Number of images, Number of tiles, Channels, Tile size, Tile size]. Padding to the maximum number of tiles is bad, but in the worst case scenario you end up increasing the tensor x4 (from 1 tile to maximum number of tiles = 4). But for the number of images it can get really really big, as there are samples with +30 images.To check this phenomenon I've included
scripts/check_padding_mm.py
which computes the % of padding in a sample. Feel free to give it a try but it's very easy to get samples where the majority of the input is padding.That's why I proposed continue working on a DataLoader & Dataset than can pack multiple samples up to a given
input_ids
length OR number of images in a batch. Packing theinput_ids
is fairly easy while packing the cross attention masks will require a bit more effort. Let me know if you would be interested on supporting that feature or you just want to include in the repo an example of the multimodal pipeline despite the padding issue described. I also plan including some unit test, to check the generated samples & recovering from failures abilities.Other comments:
scripts/check_padding_mm.py
script.torchtune
cleaning the unnecessary parts like the code for the inference case. Also in theformat_obelics
function we could drop the last images in the case the sample end with images and not text as no token will attend to them and we dont compute the loss with the image tokens (So they are useless)input_ids
/tokens
across the repo.Toni