US 11,837,253 B2
Distinguishing user speech from background speech in speech-dense environments
David D. Hardek, Allison Park, PA (US)
Assigned to VOCOLLECT, INC., Pittsburgh, PA (US)
Filed by Vocollect, Inc., Pittsburgh, PA (US)
Filed on Sep. 28, 2021, as Appl. No. 17/449,213.
Application 17/449,213 is a continuation of application No. 16/695,555, filed on Nov. 26, 2019, granted, now 11,158,336.
Application 16/695,555 is a continuation of application No. 15/220,584, filed on Jul. 27, 2016, granted, now 10,714,121, issued on Jul. 14, 2020.
Prior Publication US 2022/0013137 A1, Jan. 13, 2022
Int. Cl. G10L 25/84 (2013.01); G10L 25/51 (2013.01); G10L 15/07 (2013.01); G10L 15/06 (2013.01); G10L 15/16 (2006.01); G10L 25/78 (2013.01)
CPC G10L 25/84 (2013.01) [G10L 15/063 (2013.01); G10L 15/07 (2013.01); G10L 15/16 (2013.01); G10L 25/51 (2013.01); G10L 2025/783 (2013.01)] 20 Claims
OG exemplary drawing
 
1. A method of speech recognition, the method comprising:
generating a normalized audio input based on an accessed audio input;
determining if the normalized audio input matches a standardized user voice, the standardized user voice based on a plurality of training samples, wherein a single training word is selected from the plurality of training samples and normalized to generate the standardized user voice;
categorizing the received audio input as a user speech originating from an operator of the SRD in an instance in which it is determined that the normalized audio input matches the standardized user voice;
comparing the received audio input against a template comprising a plurality of samples of user speech in an instance in which it is determined that the normalized audio input does not match the standardized user voice; and
categorizing the received audio input as a background sound in an instance in which it is determined that the received audio input does not match a sample of user speech.