-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Consider adding onVoiceActivity event on MediaStreamTrack for audio #145
Comments
A few thoughts.
Maybe this could be exposed as part of MediaStreamTrackAudioStats.
|
yes. VAD should be opted-in by web app before UA starts to emit the event. |
This issue had an associated resolution in WebRTC June 18 2024 meeting – 18 June 2024 (Issue #145 Consider adding onVoiceActivity event on MediaStreamTrack for audio):
|
Browsers may apply voice activity detection(VAD) for AudioTrack as a prelude step for subsequent audio processings(noise suppression, echo cancellation, etc).
Today without this capability exposed through MediaStreamTrack, if a video conferencing application wants to detect voice activity when user "mutes" their microphone from browser UI, application typically sets up a WebAudio worklet to perform ASR, and then provide hints to user to unmute in order to be heard by others.
This is a double VAD, and introduces un-necessary overhead since the audio worklet is invoked at a high frequency(typically every 10ms) and relatively high priority.
Could we consider add capability for user to query VAD capability as read-only , and listens to onVoiceActivity event if enabled?
This is similar to wire WebSpeech API to MediaStreamTrack, however the purpose here is not to perform speech recognition(STT), but just VAD.
The text was updated successfully, but these errors were encountered: