Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how to inference with batch? #17

Open
dinoSpeech opened this issue Jan 13, 2024 · 1 comment
Open

how to inference with batch? #17

dinoSpeech opened this issue Jan 13, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@dinoSpeech
Copy link

Hello, thank you for sharing really nice code.

However, I cannot find batch-wise inference codes for transcribing .

(I referred quick start example code in ReadMe)

Is there any batch-wise codes for inference?

best regards

@YuanGongND YuanGongND added the enhancement New feature or request label Jan 13, 2024
@YuanGongND
Copy link
Owner

YuanGongND commented Jan 13, 2024

hi there,

Thanks for the question.

No, the current version does not support batch inference, the reason is that this repo is based on the official OpenAI whisper repo, which does not support batch inference.

However, in evaluating Whisper-AT on large datasets (e.g., AudioSet), we do use batch inference, the implementation is pre-extract and store Whisper encoder features on disk (which is done one by one), and feed a batch to TLTR module for training or inference (which supports batch input).

There are implementations of Whisper that support batch inference from the third party, as long as their encoder feature is same with the official Whisper, you can use them to extract features from Whisper in batch.

-Yuan

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants