System.Speech.Recognition (Espacio de nombres)
The Windows Desktop Speech Technology software offers a basic speech recognition infrastructure that digitizes acoustical signals, and recovers words and speech elements from audio input.
Applications use System.Speech.Recognition namespace to access extend this basic speech recognition technology, by defining algorithms for identifying and acting on specific phrases or word patterns, and by managing the run time behavior of this speech infrastructure.
Applications manage and obtain use grammars -- sets of rules defining how specific combinations of words and phrases are to be understood --through the general purpose System.Speech.Recognition.Grammar class, which hosts runtime, persisted, or dynamically constructed instances of System.Speech.Recognition.SrgsGrammar.SrgsDocument. SrgsDocument instances contain W3C Speech Recognition Grammar Specification (SRGS) compliant grammar documents.
A simplified means of specifying grammar documents is provided through the System.Speech.Recognition.GrammarBuilder and System.Speech.Recognition.Choices classes. While full support for generating SRGS compliant grammars is provided by the members of the System.Speech.Recognition.SrgsGrammar name space.
In addition, a special case grammar to support a conventional dictation model is available through System.Speech.Recognition.DictationGrammar objects .
Instances of System.Speech.Recognition.SpeechRecognizer and System.Speech.Recognition.SpeechRecognitionEngine objects supplied with appropriate Grammar objects provide the primary access to the Windows Desktop Speech Technology recognition engines.
The SpeechRecognizer class is used to create client applications making use of a system's current recognition technology, which is configured through the Audio Input member of the Control Panel, and a computers default audio input mechanism.
Building and application using SpeechRecognitionEngine allows for more control on the configuration and type of recognition engine, which runs in process. Using SpeechRecognitionEngine also provides for the dynamic selection of audio input, whether from devices or files.
|Choices||Represents a list of alternative items to make up an element in a grammar.|
|DictationGrammar||Provides access to the system provided grammar used for free text dictation.|
|EmulateRecognizeCompletedEventArgs||Class passed to handlers of events generated as upon completion of an asynchronous emulation of speech recognition.|
|Grammar||Provides run time support for obtaining and managing grammar information.|
|GrammarBuilder||Provides an easy to use mechanism for constructing complicated Grammar from simple inputs.|
|LoadGrammarCompletedEventArgs||Class passed to handlers of events generated when a new grammar is asynchronously loaded into a recognition engine.|
|RecognitionEventArgs||Base class for event arguments objects passed to handlers of the speech recognition events.|
|RecognitionResult||Represents the result of recognition engine on audio input as detailed information about the best candidate phrases matching phrase, and a list of all candidate matching phrases.|
|RecognizeCompletedEventArgs||Passes information to handlers of the p|
|RecognizedAudio||Represents audio input a recognition engine used to generate candidate phrases result.|
|RecognizedPhrase||Represents detailed information about a candidate phrase found by a recognition engine as matching audio input.|
|RecognizedWordUnit||Provides the atomic unit of recognized speech.|
|RecognizerInfo||Encapsulates identifying information about a recognition engine (SpeechRecognizer or SpeechRecognitionEngine).|
|RecognizerUpdateReachedEventArgs||Class passed to handlers of recognition engine update events.|
|ReplacementText||Contains originally recognized text replace using speech normalization by a recognition engine.|
|SemanticResultKey||Attaches key string to SemanticResultValue values to define the SemanticValue objects created in Grammar using GrammarBuilder instances|
|SemanticResultValue||Sets values to SemanticValue objects created in Grammar using GrammarBuilder instances.|
|SemanticValue||Represents the semantic organization of a recognized phrase.|
|SpeechDetectedEventArgs||Class passed to handlers of events generated by recognition engines upon detection of human speech in an audio input.|
|SpeechHypothesizedEventArgs||Class passed as an event argument to handlers generated as a speech recognition engine develops tentative recognition results for an input phrase.|
|SpeechRecognitionEngine||Provides access to run any properly installed speech recognition services found on a Windows Desktop system.|
|SpeechRecognitionRejectedEventArgs||Class passed to handlers of events generated as when a speech recognition engine fails to identify a phrase from input.|
|SpeechRecognizedEventArgs||Class passed to handlers of events generated as upon phrase recognition by a recognition engine.|
|SpeechRecognizer||Provides access to the default shared speech recognition service available on the Windows Desktop.|
|SpeechUI||Provides text and status information to the operating system in speech user interface to display to the user.|
|StateChangedEventArgs||Class passed to handlers of events generated when a recognition engine state changes.|
|AudioSignalProblem||Enumeration used to indicate the problem state of audio signals received by a recognition engine.|
|AudioState||Enumeration describing current state of the audio input a recognition engine has detected.|
|DisplayAttributes||Describes formatting for printed output of words of a recognized phrase.|
|RecognizeMode||Specifies if asynchronous SpeechRecognitionEngine recognition is to terminate or continue following a recognition operation.|
|RecognizerState||Describes the current state of a desktop speech recognition engine.|
|SubsetMatchingMode||Defines the method used to match audio input which may be a subset of the match string.|