Package ru.yandex.speechkit

Interfaces

AudioSource

Interface for an audio source.

AudioSourceListener

Interface of an audio source listener.

EventLogger

Interface for logging events that occur in RecognizerActivity.

IdentificationListener

The interface defines callback methods that are called by an object of the Request class.

LocationProvider

Interface for accessing geolocation.

OnlineRecognizer.GetPlatformLanguagesListener

The interface defines callback methods that are called when getting a list of languages supported by the platform.

PhraseSpotterListener

The interface defines callback methods to track voice activation events.

Recognizer

Interface of the object that recognizes speech.

RecognizerListener

The interface defines callback methods that are invoked during the recognition lifecycle.

RegistrationListener

The interface defines callback methods that are called by an object of the Request class.

Request

Generic interface that represents a request.

Vocalizer

The interface of the object that performs text-to-speech synthesis and conversion.

VocalizerListener

The interface defines callback methods for tracking speech synthesis and playback events.

VoiceRequest

Interface for requests that pass speech.

Classes

AudioProcessingMode

Processing mode of incoming audio.

AutoStartStopAudioSource

An audio source that automatically starts and stops audio recording using the standard device input.

AutoStartStopAudioSource.Builder

Class for creating and configuring the AutoStartStopAudioSource object.

BaseSpeechKit

Singleton class that is used to configure and control the SpeechKit operation.

BaseSpeechKit.LibraryInitializationException

Сlass that identifies the causes of problems with loading of the native part of the library.

Biometry

Estimated biometrics of the user (age, gender, age group, and other characteristics), which are returned in the Recognition class object during speech recognition.

Biometry.Emotion

The estimates of whether the user is experiencing one of the following emotions: happiness, annoyance, anger, calmness, or fun.

Biometry.Group

Estimated gender and age group characteristics.

Biometry.LanguageScore

Contains the language code and a score of what the user said in that language.

BiometryGroup

Estimated gender and age group characteristics.

BiometryGroup.Builder

Class for creating and configuring the BiometryGroup object.

DefaultEarconsBundle

Class used to get the sounds played by default while the library components are running.

Earcons

Class to store the sounds played while the library components are running.

Earcons.BaseBuilder
EchoCancellingAudioSource

Implementation of the AudioSource interface that applies the "acoustic echo cancellation" (AEC) algorithm to the incoming audio.

Emotion

Possible emotions of synthesized voice.

Error

Descriptions of errors that may occur in various library components.

Language

Available recognition languages.

ManualStartStopAudioSource

An audio source that lets you control the start and end of audio recording by using the standard device input.

ManualStartStopAudioSource.Builder

Class for creating and configuring the ManualStartStopAudioSource object.

MutingAudioSource

Implementation of the AudioSource interface that suspends an audio that comes through an audio source.

OnlineModel

Language models for online recognition.

OnlineRecognizer

Class for online speech recognition.

OnlineRecognizer.Builder

Class for creating and configuring the OnlineRecognizer object.

OnlineVocalizer

Class for online speech synthesis and text-to-speech.

OnlineVocalizer.Builder

Class for creating and configuring OnlineVocalizer.

PhraseSpotter

Class for voice activation.

PhraseSpotter.Builder

Class for creating and configuring an object of the PhraseSpotter class.

Quality

Quality of synthesized speech.

Recognition

The recognition result that returns an object of the class Recognizer.

RecognitionHypothesis

An object representing a single recognition hypothesis.

RecognitionWord

Word or phrase that is part of the RecognitionHypothesis hypothesis.

SoundBuffer

The class contains a data buffer and information on the audio.

SoundFormat

Formats of synthesized speech.

SoundInfo

Information on an audio.

SpeechKit

Singleton class for configuring and managing the SpeechKit operation.

Synthesis

The class contains the result of speech synthesis.

Track

Music recognition result.

Track.Artist

Performer recognition result.

UniProxySession

Creates a universal server connection.

UniProxySession.Builder

Class for creating and configuring UniProxySession.

Voice

The class specifies voice settings for synthesized speech.

Enums

BaseSpeechKit.LogLevel

Logging level.

Vocalizer.TextSynthesizingMode

Defines the behavior of an object of the Vocalizer class when speech synthesis starts, if the previously launched speech synthesis process was not completed.