Skip to content

Commit 7e53452

Browse files
authored
update outdated answer (#392)
1 parent 0240622 commit 7e53452

File tree

1 file changed

+1
-5
lines changed

1 file changed

+1
-5
lines changed

fern/pages/faq/async/how-does-the-api-handle-files-that-contain-spoken-audio-in-multiple-languages.mdx

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2,8 +2,4 @@
22
title: "How does the API handle files that contain spoken audio in multiple languages?"
33
---
44

5-
Each of our language models are trained on that specific language. Generally speaking, those models will only transcribe audio spoken in that specific language. There can be some edge cases where the language model is able to transcribe a few words that are in a different language. This can happen when the model may have learned those words due to them being present in the training data.
6-
7-
For instance, the Spanish model is trained with Spanish audio, but it is not uncommon that some English words are used when speaking Spanish. So when English words appear in the training data used for the Spanish model, the Spanish model can recognize and transcribe those words.
8-
9-
It's important to note that a language model transcribing words from a different language would be an edge case and not behavior that should be reliably counted on. The expectation is that whatever language model the file is submitted to, will transcribe the spoken audio for that language and will ignore any other spoken languages in the audio.
5+
Our models are capable of transcribing audio files that contain spoken audio in more than one language. This is referred to as code switching. Code switching performance varies by language pair. Check out [this section](/docs/speech-to-text/pre-recorded-audio/code-switching) of our documentation to learn more.

0 commit comments

Comments
 (0)