-
-
Notifications
You must be signed in to change notification settings - Fork 4
/
readme-vars.yml
55 lines (45 loc) · 2.97 KB
/
readme-vars.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
---
# project information
project_name: faster-whisper
project_url: "https://github.com/SYSTRAN/faster-whisper"
project_logo: "https://raw.githubusercontent.com/linuxserver/docker-templates/master/linuxserver.io/img/faster-whisper-logo.png"
project_blurb: "[{{ project_name|capitalize }}]({{ project_url }}) is a reimplementation of OpenAI's Whisper model using CTranslate2, which is a fast inference engine for Transformer models. This container provides a Wyoming protocol server for faster-whisper."
project_lsio_github_repo_url: "https://github.com/linuxserver/docker-{{ project_name }}"
# supported architectures
available_architectures:
- { arch: "{{ arch_x86_64 }}", tag: "amd64-latest"}
# development version
development_versions: true
development_versions_items:
- { tag: "latest", desc: "Stable releases" }
- { tag: "gpu", desc: "Releases with Nvidia GPU support" }
# container parameters
common_param_env_vars_enabled: true
param_container_name: "{{ project_name }}"
param_usage_include_env: true
param_env_vars:
- { env_var: "WHISPER_MODEL", env_value: "tiny-int8", desc: "Whisper model that will be used for transcription. From `tiny`, `base`, `small` and `medium`, all with `-int8` compressed variants", env_options: ["tiny-int8", "tiny", "base-int8", "base", "small-int8", "small", "medium-int8"] }
param_usage_include_vols: true
param_volumes:
- { vol_path: "/config", vol_host_path: "/path/to/data", desc: "Local path for Whisper config files." }
param_usage_include_ports: true
param_ports:
- { external_port: "10300", internal_port: "10300", port_desc: "Wyoming connection port." }
# optional container parameters
opt_param_usage_include_env: true
opt_param_env_vars:
- { env_var: "WHISPER_BEAM", env_value: "1", desc: "Number of candidates to consider simultaneously during transcription." }
- { env_var: "WHISPER_LANG", env_value: "en", desc: "Language that you will speak to the add-on." }
readonly_supported: true
# application setup block
app_setup_block_enabled: true
app_setup_block: |
For use with Home Assistant [Assist](https://www.home-assistant.io/voice_control/voice_remote_local_assistant/), add the Wyoming integration and supply the hostname/IP and port that Whisper is running add-on."
When using the `gpu` tag with Nvidia GPUs, make sure you set the container to use the `nvidia` runtime and that you have the [Nvidia Container Toolkit](https://github.com/NVIDIA/nvidia-container-toolkit) installed on the host and that you run the container with the correct GPU(s) exposed. See the [Nvidia Container Toolkit docs](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/sample-workload.html) for more details.
For more information see the [faster-whisper docs](https://github.com/SYSTRAN/faster-whisper),
# changelog
changelogs:
- { date: "18.07.24:", desc: "Rebase to Ubuntu Noble." }
- { date: "19.05.24:", desc: "Bump CUDA to 12 on GPU branch." }
- { date: "08.01.24:", desc: "Add GPU branch." }
- { date: "25.11.23:", desc: "Initial Release." }