How to structure NSArray of text passed to LanguageModelGenerator?

Home Forums OpenEars How to structure NSArray of text passed to LanguageModelGenerator?

Viewing 2 posts - 1 through 2 (of 2 total)

  • Author
    Posts
  • #1021787
    omgbobbyg
    Participant

    Hello,

    I am using OpenEars/RapidEars/Rejecto to match speech to text in a speech being displayed on screen. Right now when I create the NSArray of text to pass to the LanguageModelGenerator, I create an NSArray with each element corresponding to a line of text in the speech. However, I am noticing that the quality of the hypothesis being generated seems to degrade as the lengths of the lines in each element of the NSArray get shorter (which happens when the user increases the font size for the speech). Is that expected behavior?

    In terms of structuring the NSArray input to the LanguageModelGenerator, is there any difference in passing a single element NSArray with all text contained in the first element versus dividing it up so that each line of text is its own element?

    Thanks
    Bobby

    #1021788
    Halle Winkler
    Politepix

    Yes, there is a difference in the way probabilities are calculated for individual words versus phrases, as well as how phrases are handled between language models and grammars. You can take a quick look at the docs for LanguageModelGenerator to get the high-level on those options and their effects.

Viewing 2 posts - 1 through 2 (of 2 total)
  • You must be logged in to reply to this topic.