PocketsphinxController + Rapidears Live Speech Recogn without using Vacubullar

Home Forums OpenEars PocketsphinxController + Rapidears Live Speech Recogn without using Vacubullar

Viewing 2 posts - 1 through 2 (of 2 total)

  • Author
    Posts
  • #1019877
    ManojNeelam
    Participant

    In my application i need live speech recognition. i dont want any predefined words.If persons speak any thing application should recognize.For Live Speech Recognize i should use Rapidears. I follwed your Tutorial and I downloaded Distibution and Rapidears Free Demo.
    I integrated the openears sample app with rapidears.
    in startListening Method i have written:
    [self.pocketsphinxController startRealtimeListeningWithLanguageModelAtPath:self.pathToGrammarToStartAppWith dictionaryAtPath:self.pathToDictionaryToStartAppWith acousticModelAtPath:[AcousticModel pathToModel:@”AcousticModelEnglish”]];

    Delegate methods are calling they are not recognizing other words except that are present in in the dictionary.
    (void) rapidEarsDidReceiveLiveSpeechHypothesis:(NSString *)hypothesis recognitionScore:(NSString *)recognitionScore {
    NSLog(@”rapidEarsDidReceiveLiveSpeechHypothesis: %@”,hypothesis);
    }

    – (void) rapidEarsDidReceiveFinishedSpeechHypothesis:(NSString *)hypothesis recognitionScore:(NSString *)recognitionScore {
    NSLog(@”rapidEarsDidReceiveFinishedSpeechHypothesis: %@”,hypothesis);

    I dont want to use any Vocabulary Dictionary.The user what he speak it should Recognize.for that what shall i write in this

    [self.pocketsphinxController startRealtimeListeningWithLanguageModelAtPath:<#(NSString *)#> dictionaryAtPath:<#(NSString *)#> acousticModelAtPath:<#(NSString *)#>];

    dictionaryAtPath = ?

    startRealtimeListeningWithLanguageModelAtPath = ?

    #1019879
    Halle Winkler
    Politepix

    Hello,

    “Live speech recognition” in this case doesn’t mean “the speech recognition engine can recognize words it is not aware of”, it means “the speech recognition is getting hypotheses for the user’s speech in real time as it is being spoken, rather than waiting for the user to pause and then analyzing all of the speech that happened before the pause”.

    There is no kind of speech recognition in which the engine can recognize words that a developer has not already given it as a vocabulary. The closest thing to that is cloud-based speech recognition (not offline like OpenEars, but online via a network API) where the API developer has already defined an extremely large vocabulary for the service, something that is possible because there is an extremely large amount of memory and CPU cycles available. The idea that some speech recognition engines can recognize anything the user might say is an illusion created by the fact that with a large data farm it is possible to give them access to such a large vocabulary. Because the vocabulary size has to be smaller in offline recognition, offline recognition cannot perform the illusion of “recognizing any word the user might say”. It has to be used with smaller vocabularies that you define for your particular app.

    What you need for your requirements is to find a network service which can do large vocabulary recognition. Nuance has this and I believe iSpeech also does. All of Politepix’s products are offline products that are designed not to send data over the network, so none of them are intended as tools for taking unstructured dictation from users using vocabularies that are the size of entire languages. I hope this is helpful.

Viewing 2 posts - 1 through 2 (of 2 total)
  • You must be logged in to reply to this topic.