I am using the Affectiva’s SDK for Linux to conduct emotion processing on videos. Even though I disabled the visual display which shows the tracking of facial movements, my processing time still seems to be equivalent to that of the video itself. I have been processing videos from 30 seconds to 5 minutes, but I plan to go up to larger videos in the future.
Is the processing time required to detect emotion in a video equivalent to the length of the video itself?
I guess this would make sense if the software is still playing back the entire video to analyze it. Is there a way to speed things up? Perhaps I am missing something?
Thank you for your time, consideration and help!