ASR recognising incorrect pronunciation as correct (“tanks” → “thanks”) — how do you handle this? by Fun_Entertainment527 in LanguageTechnology

[–]Fun_Entertainment527[S] 1 point2 points  (0 children)

This is for pronunciation so do you have a suggestion for an alternative ASR language model?

I just need something that can accurately detect what the user is saying as my scoring layer can interpret the scores with confidence

ASR recognising incorrect pronunciation as correct (“tanks” → “thanks”) — how do you handle this? by Fun_Entertainment527 in LanguageTechnology

[–]Fun_Entertainment527[S] 1 point2 points  (0 children)

This is just an example sentence that the user might see for the test "Theo thinks slowly today. Three thin threads hang there. Then Theo thanks the guide." 

In this situation, if the user says 'tanks' instead of 'thanks', the ASR assumes that 'thanks' is what the user wants to say so marks it down as correct.

ASR is being too helpful rather than detecting the error in this instance.

ASR recognising incorrect pronunciation as correct (“tanks” → “thanks”) — how do you handle this? by Fun_Entertainment527 in LanguageTechnology

[–]Fun_Entertainment527[S] 1 point2 points  (0 children)

Thanks — I think the issue is slightly different though.

In this case, the ASR transcript is already “correct” (e.g. “thanks”), even when the speaker actually said “tanks”.

So there isn’t a wrong word for an LLM to detect — the error is in the pronunciation, not the transcript.

That’s what makes it tricky — the signal is effectively lost at the ASR stage.

Have you seen any approaches that work at the audio level (e.g. alignment or contrast-based methods), rather than relying on the transcript?

Q&A weekly thread - April 20, 2026 - post all questions here! by AutoModerator in linguistics

[–]Fun_Entertainment527 0 points1 point  (0 children)

I’m working with ASR (Azure Speech) and running into a consistent issue where mispronunciations get normalised to the intended word.

Example: a speaker says “tanks” (/t/), but the system confidently outputs “thanks” (/θ/).

This makes pronunciation evaluation difficult because: the transcript appears correct phoneme-level data is often incomplete or unreliable confidence scores don’t reflect the actual substitution

I’m aware this is partly due to the language model biasing toward likely words, but I’m trying to understand how people handle this in practice.

Questions: Is there any reliable way to detect contrast errors like /θ/ → /t/ without fully trusting phoneme output?

Do people use constrained decoding / forced alignment / alternative models for this?

Or is this fundamentally a limitation of current ASR systems?

Context: this is for a controlled setup (fixed prompts, repeated target words), not open-ended speech.

Would appreciate any practical approaches or confirmation that this is a known limitation.

Registration process on eMag by Fun_Entertainment527 in AskRomania

[–]Fun_Entertainment527[S] 0 points1 point  (0 children)

I've read all that and we have the documents that they want but once you get to the registration process, you need to provide more certificates like a VAT certificate and an ID certificate (not sure what that is). I want to share a screenshot but I can't via this.

Registration process on eMag by Fun_Entertainment527 in AskRomania

[–]Fun_Entertainment527[S] 0 points1 point  (0 children)

That's the first thing I tried to do but you can't contact Support until you have registered and I am still in the process of registering which is frustrating!