This repository was archived by the owner on Dec 14, 2023. It is now read-only.
Replies: 1 comment 2 replies
-
|
Hey! The more variance in data the better, but having a quality, balanced dataset (cleaned, cropped, etc.) will always win regardless of size. You can go as high or low as you wish for whatever domain you're trying to train on. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm prepping to train a general 448x256 model and have already acquired a dataset of over 7000 videos that are each 30s - 6 minutes long (before they are cut up and captioned). How much data do you think is too much to be useful for iteratively training and improving this model? I will be training on a 3090 and would like to see results within a day so I can tweak it as needed.
Beta Was this translation helpful? Give feedback.
All reactions