Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

time for pretraining #3

Open
Tortoise17 opened this issue May 7, 2021 · 4 comments
Open

time for pretraining #3

Tortoise17 opened this issue May 7, 2021 · 4 comments

Comments

@Tortoise17
Copy link

I have to ask that do you have any figures how much this takes for pretraining and on which GPU and which dataset_ ? please if you can guide.

@harshraj32
Copy link

where u able to download the dataset?

@ammesatyajit
Copy link
Owner

So for the dataset, I used the HowTo100M dataset and filtered out the cooking videos. The specific ids for only cooking videos are listed in VideoBERT/data/ids.txt. Here are the steps I took to download the dataset:
go to this link: https://www.di.ens.fr/willow/research/howto100m/
On the bottom of the page there should be a download section. To get access to the private server to download it easily, you will need to fill out the form that it links. The creator of the dataset should respond within 2-3 days giving you a login for the server. Using that, you can download the videos with the instructions given along with the ids.txt file.

@Tortoise17 Unfortunately I do not have the pretraining figures for this model. I do not recall which GPU I was using at the time since I was using a VM. However, this should run on a google colab equivalent GPU if the model size is decreased a bit. Hope that helps.

@harshraj32
Copy link

yeah there are over 47k ids, what would be the size of all 47k videos?

@ammesatyajit
Copy link
Owner

It was about 400-500 GB

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants