-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pose datasets start/end #8
Comments
This is something I have thought about a lot. The problem is that to make the conversion to frames (useful for some modalities) I need to know the fps of the input file. I would like to find a way to extract the fps directly from the file (be it mp4 or whatever). I am aware that for poses, pose-format allows to specify start and end in milliseconds. The problem is currently in video (which is not fully implemented at the moment, but will extend image2text). I know it's something to look at, but I'm focusing on more priority things at the moment. Any suggestions are welcome. |
for poses - specifying start and end time uses the for videos - depending on the use case - can perform the same thing by opening the video and asking for the fps - thing is, videos don't necessarily have constant fps, and then, it is way more important to work with times. I feel like working with |
Noted, upcoming commits will change default unit to seconds. |
milliseconds is better* |
It is most common to get data in subtitles format, or other time aligned formats
Therefore, when something like this:
multimodalhugs/multimodalhugs/data/datasets/pose2text.py
Line 89 in adcc3ea
multimodalhugs/multimodalhugs/processors/pose2text_preprocessor.py
Line 61 in adcc3ea
happen, it forces the user to do unnecessary calculation. (also, for videos for example)
Instead, please use
start_time
andend_time
which are in milliseconds.And either way, they are not floats, but ints
multimodalhugs/multimodalhugs/data/datasets/pose2text.py
Lines 35 to 36 in adcc3ea
The text was updated successfully, but these errors were encountered: