Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Maintaining state across file audio source chunks #248

Open
Aduomas opened this issue Oct 28, 2024 · 1 comment
Open

Maintaining state across file audio source chunks #248

Aduomas opened this issue Oct 28, 2024 · 1 comment

Comments

@Aduomas
Copy link

Aduomas commented Oct 28, 2024

Hello,

I am looking for a way to do chunk-based inference instead of streaming inference using audio files.
The issue now is that each file audio will have new inference and thus new state (new speaker embeddings) which is unwanted behaviour for my program.

How should I do achieve wanted behaviour of doing inference on larger chunks of audio (such as 20 second) and keeping the pipeline state across ?

@juanmc2005
Copy link
Owner

Hi @Aduomas, given your description, do you actually require a streaming pipeline? It looks like pyannote.audio could achieve what you want

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants