-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
time for pretraining #3
Comments
where u able to download the dataset? |
So for the dataset, I used the HowTo100M dataset and filtered out the cooking videos. The specific ids for only cooking videos are listed in VideoBERT/data/ids.txt. Here are the steps I took to download the dataset: @Tortoise17 Unfortunately I do not have the pretraining figures for this model. I do not recall which GPU I was using at the time since I was using a VM. However, this should run on a google colab equivalent GPU if the model size is decreased a bit. Hope that helps. |
yeah there are over 47k ids, what would be the size of all 47k videos? |
It was about 400-500 GB |
I have to ask that do you have any figures how much this takes for pretraining and on which GPU and which dataset_ ? please if you can guide.
The text was updated successfully, but these errors were encountered: