Finer Beat Timings #7
Unanswered
ljendrusch
asked this question in
Q&A
Replies: 1 comment
-
Hi, I'm glad that you use this model 🤩 Sadly, the temporal resolution is fixed to 100 FPS (frames per second). In detail, it extracts 100 embeddings per second and then they are fed into DBNDownBeatTrackingProcessor, which finally decides the positions of beats and downbeats. But, I think 100 FPS should be enough for many cases so I'm curious why you need higher temporal resolution. Maybe we can think together and find another solution. Thanks! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Greetings,
First off, incredible tool, thanks for making it available. Awesome job representing KAIST.
I'm doing a project to calculate audio similarity of song segments over a database of songs.
Allin1 has proven invaluable and I've done some heavy testing to prove its boundary detection is very accurate.
Is the output from the Beats head originally 100ths of a second, or is it finer? If it is finer, how could I access the finer timings?
Finer timings would increase my accuracy by a good deal because I'm comparing raw sample arrays.
Example code:
I'm getting segment bounds
a1, a2, b1, b2
and downbeats from allin1 AnalysisResults.But since
seg_a
andseg_b
are 44.1khz I'm afraid aligning on 100ths of seconds will have a large effect on accuracy.Any help or advice you may provide would be greatly appreciated. Again, awesome model, thanks for your hard work.
Beta Was this translation helpful? Give feedback.
All reactions