Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Increase concurrency capability #241

Open
daiDai-study opened this issue Sep 27, 2024 · 1 comment
Open

Increase concurrency capability #241

daiDai-study opened this issue Sep 27, 2024 · 1 comment

Comments

@daiDai-study
Copy link

daiDai-study commented Sep 27, 2024

I want to increase concurrency capability, but i see source code like:

    with model_lock:
        segments = []
        text = ""
        segment_generator, info = model.transcribe(audio, beam_size=5, **options_dict)
        for segment in segment_generator:
            segments.append(segment)
            text = text + segment.text
        result = {"language": options_dict.get("language", info.language), "segments": segments, "text": text}

can i remove the lock?

@aidancrowther
Copy link
Contributor

From my work with the code, the lock prevents multiple accesses to the same model instance. If you wanted to run multiple in parallel, you would need to initialize multiple models. As the model, when running, will more or less saturate the processing capability of the GPU/CPU it is running on, this likely wouldn't be as useful as you hope. It is much easier/configurable to spawn multiple instances of the docker container instead. For example, using --gpus device=n where n is the id of a compatible GPU allows you to run the model on a specific GPU. I have used this to run testing on two Nvidia Tesla GPUs as well as the system CPU

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants