You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The current synching methods (audio cross correlation and brightness contrast thresholding) both work in the majority of cases. But there are no options for a different method if a user's videos do not synchronize properly (although the brightness contrast method can be run with different thresholds to change the results). Adding methods will help ensure a method works for each set of videos, and different synching methods will be robust to different types of noise. I would like to have at least 2 audio and two brightness methods.
Audio fingerprinting:
As described in this academic poster by Mark Koh, you can use the Python package Dejavu to match audio files and use the output to synchronize videos.
Brightness Cross Correlation:
Since we already take the brightness over time of videos in the brightness contrast thresholding method, we can simply use the cross correlation methods from the audio cross correlation method and apply it to the brightness time series.
The text was updated successfully, but these errors were encountered:
I also suggest you try to avoid the "reinventing the wheel" issue by focusing your efforts on aspects (e.g. phase/polarity adjustments after synching) that other projects have not yet considered/implemented.
Thanks for the thoughtful comment, and including us in your resource.
audalign and audio-offset-finder definitely look interesting, and spiritually similar to this project. I'll check them out and see if there's anything that could help make our tool more robust.
As far as a phase polarity post-processing step goes, it's not top on our list, as we're focused on synchronizing videos for use in motion capture, where downstream audio isn't very important. I would gladly review a contribution of it if it's something you're motivated to implement though!
Also, for the Learning Robust Video Synchronization without Annotations, it seems to be a different kind of video synchronization than we're interested in. They perform spatial alignment (align images in space), while we're interested in temporal alignment (align frames by when they were shot)
The current synching methods (audio cross correlation and brightness contrast thresholding) both work in the majority of cases. But there are no options for a different method if a user's videos do not synchronize properly (although the brightness contrast method can be run with different thresholds to change the results). Adding methods will help ensure a method works for each set of videos, and different synching methods will be robust to different types of noise. I would like to have at least 2 audio and two brightness methods.
As described in this academic poster by Mark Koh, you can use the Python package Dejavu to match audio files and use the output to synchronize videos.
Since we already take the brightness over time of videos in the brightness contrast thresholding method, we can simply use the cross correlation methods from the audio cross correlation method and apply it to the brightness time series.
The text was updated successfully, but these errors were encountered: