Increasing speech recognition accuracy using a custom languagemodel?

Home Forums OpenEars Increasing speech recognition accuracy using a custom languagemodel?

Tagged: 

Viewing 18 posts - 1 through 18 (of 18 total)

  • Author
    Posts
  • #1026537
    jepjep7
    Participant

    Hi Halle,

    I have two questions:

    1) I am trying to increase the speech recognition accuracy. Do you think it would help to make a custom languagemodel using the lmtool?

    http://www.speech.cs.cmu.edu/tools/lmtool-new.html

    2) I tried to make my own languagemodel using the lmtool instead of the built-in languagemodel generator. I ran into problems here:

    NSString *lmDirectory = [NSString stringWithFormat:@”%@/words.%@”,[NSSearchPathForDirectoriesInDomains(NSCachesDirectory, NSUserDomainMask, YES) objectAtIndex:0],@”languagemodel”];
    NSString *dicDirectory = [NSString stringWithFormat:@”%@/words.%@”,[NSSearchPathForDirectoriesInDomains(NSCachesDirectory, NSUserDomainMask, YES) objectAtIndex:0],@”dic”];

    self.pathToGrammarToStartAppWith = lmDirectory;
    self.pathToDictionaryToStartAppWith = dicDirectory;

    When I ran this, it said that the dictionary path was null. I also noticed that this may require a .DMP file instead of a .languagemodel file. How does one use a custom .dic and .languagemodel file with the updated version of Openears?

    Thank you.

    #1026539
    Halle Winkler
    Politepix

    Welcome,

    Do you think it would help to make a custom languagemodel using the lmtool?

    No, there is no accuracy difference between OpenEars custom models and the models from that tool, which are both generated using CMU language model generation. It would probably be more effective to share your goals and issues here and see if I can help you.

    Regarding your paths being null, this just needs the normal kind of troubleshooting you’d need to do to make sure that any app bundle file path had a real file at the end of it – check that the items are really added to the target and that you’re running the target you think you are, find out where to expect an added file to be found in the bundle and how that is referenced (in this case, you look like you are trying to get a file from a caches directory, but normally a file that you add to an app target is found in the app bundle rather than the sandbox user files so I would probably start there with investigating the issue, plus you must already know the full names of your files since you named them before adding them to the bundle so that would be the second thing to look into, since your name appears to be dynamically generated), and use NSFileManager to verify the existence of your file or return an error.

    OpenEars’ file loading is known to work fine using normal NSString representations of file paths, so you can just focus on figuring out why you have a null path, which is going to relate to app bundle structures/Cocoa APIs rather than OpenEars.

    #1026573
    jepjep7
    Participant

    Thank you for your fast response, Halle. I appreciate it!

    The reason I was trying to modify the languagemodel in my original post is that I thought I was having problems with number recognition, specifically “fives” and “nines”. After more testing, I narrowed down the problem. I am having problems recognizing words at the end of utterances. For example, when I try to recognize the phrase “three five five”, Openears often only recognizes “three five”. Why is that? It is almost like the last word gets cut off from being recognized.

    Settings:
    setVadThreshold:3.0
    setSecondsOfSilenceToDetect:0.3

    #1026574
    Halle Winkler
    Politepix

    It looks like too short of a setSecondsOfSilenceToDetect, but if you want you can create a replication case for me to run:

    https://www.politepix.com/forums/topic/how-to-create-a-minimal-case-for-replication/

    #1026575
    jepjep7
    Participant

    I ran Openears Demo and it seemed to be working better than mine at recognizing numbers 1-9. So I checked my code and I had an extra “nil” on my firstLanguageArray. Here is what it looked like:

    NSArray *firstLanguageArray = [[NSArray alloc] initWithArray:[NSArray arrayWithObjects:
    @”ONE”,
    @”TWO”,
    @”THREE”,
    @”FOUR”,
    @”FIVE”,
    @”SIX”,
    @”SEVEN”,
    @”EIGHT”,
    @”NINE”,
    nil]];

    I changed the array by deleting the nil to look like this:

    NSArray *firstLanguageArray = @[@”ONE”,
    @”TWO”,
    @”THREE”,
    @”FOUR”,
    @”FIVE”,
    @”SIX”,
    @”SEVEN”,
    @”EIGHT”,
    @”NINE”];

    It seems to be working better now! Do you think that was it!?

    Thanks.

    #1026581
    Halle Winkler
    Politepix

    I lean towards thinking it’s related to something else or maybe a small sample size, but it would be interesting to see the contents of the two generated .dic files from the Caches directory resulting from each of those arrays.

    #1026587
    jepjep7
    Participant

    I checked, and the DIC files were exactly the same. Openears Demo is only recognizing “three five” when I say “three five five”. I will create a replication case.

    #1026589
    jepjep7
    Participant

    I am having trouble getting saveThatWave to work.
    1) I added SaveThatWaveDemo.framework to the “Frameworks” folder in the OpenEarSample App.
    2) In ViewController.h, I added:
    #import <SaveThatWaveDemo/OEEventsObserver+SaveThatWave.h>
    #import <SaveThatWaveDemo/SaveThatWaveController.h>
    @property (strong, nonatomic) SaveThatWaveController *saveThatWaveController;
    3) In ViewController.m, I added:
    @synthesize saveThatWaveController;

    Everything complied perfectly until this point.
    Then when I added this line to ViewDidLoad:
    self.saveThatWaveController = [[SaveThatWaveController alloc] init];

    It threw a “linker command failed with exit code 1 (use -v to see invocation)” error.

    Do you have any idea why this is not working? I would like to make a test-case for the “three five five” example above. Thanks!

    #1026590
    Halle Winkler
    Politepix

    Probably there is no -ObjC linker flag added as per the instructions, but it’s necessary to copy/paste the full Xcode errors in order to troubleshoot them.

    #1026615
    jepjep7
    Participant

    I checked and the -ObjC linker flag was already added. Here are the Xcode errors:

    Ld /Users/Joe/Library/Developer/Xcode/DerivedData/OpenEarsSampleApp-anzfsmtoufwysiaqbnobltsaejum/Build/Products/Debug-iphonesimulator/OpenEarsSampleApp\ Tests.xctest/OpenEarsSampleApp\ Tests normal x86_64
    cd /Users/Joe/Downloads/OpenEarsDistribution/OpenEarsSampleApp
    export IPHONEOS_DEPLOYMENT_TARGET=8.0
    export PATH=”/Applications/Xcode.app/Contents/Developer/Platforms/iPhoneSimulator.platform/Developer/usr/bin:/Applications/Xcode.app/Contents/Developer/usr/bin:/usr/bin:/bin:/usr/sbin:/sbin”
    /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/clang -arch x86_64 -bundle -isysroot /Applications/Xcode.app/Contents/Developer/Platforms/iPhoneSimulator.platform/Developer/SDKs/iPhoneSimulator8.4.sdk -L/Users/Joe/Library/Developer/Xcode/DerivedData/OpenEarsSampleApp-anzfsmtoufwysiaqbnobltsaejum/Build/Products/Debug-iphonesimulator -F/Users/Joe/Library/Developer/Xcode/DerivedData/OpenEarsSampleApp-anzfsmtoufwysiaqbnobltsaejum/Build/Products/Debug-iphonesimulator -F/Applications/Xcode.app/Contents/Developer/Platforms/iPhoneSimulator.platform/Developer/SDKs/iPhoneSimulator8.4.sdk/Developer/Library/Frameworks -F/Users/Joe/Downloads/OpenEarsDistribution/OpenEarsSampleApp/../Framework -F/Applications/Xcode.app/Contents/Developer/Platforms/iPhoneSimulator.platform/Developer/Library/Frameworks -F/Applications/Xcode.app/Contents/Developer/Platforms/iPhoneSimulator.platform/Developer/SDKs/iPhoneSimulator8.4.sdk/Developer/Library/Frameworks -filelist /Users/Joe/Library/Developer/Xcode/DerivedData/OpenEarsSampleApp-anzfsmtoufwysiaqbnobltsaejum/Build/Intermediates/OpenEarsSampleApp.build/Debug-iphonesimulator/OpenEarsSampleApp\ Tests.build/Objects-normal/x86_64/OpenEarsSampleApp\ Tests.LinkFileList -Xlinker -rpath -Xlinker @executable_path/Frameworks -Xlinker -rpath -Xlinker @loader_path/Frameworks -Xlinker -objc_abi_version -Xlinker 2 -ObjC -fobjc-arc -fobjc-link-runtime -Xlinker -no_implicit_dylibs -mios-simulator-version-min=8.0 -framework Slt -framework OpenEars -framework AudioToolbox -framework AVFoundation -Xlinker -dependency_info -Xlinker /Users/Joe/Library/Developer/Xcode/DerivedData/OpenEarsSampleApp-anzfsmtoufwysiaqbnobltsaejum/Build/Intermediates/OpenEarsSampleApp.build/Debug-iphonesimulator/OpenEarsSampleApp\ Tests.build/Objects-normal/x86_64/OpenEarsSampleApp\ Tests_dependency_info.dat -o /Users/Joe/Library/Developer/Xcode/DerivedData/OpenEarsSampleApp-anzfsmtoufwysiaqbnobltsaejum/Build/Products/Debug-iphonesimulator/OpenEarsSampleApp\ Tests.xctest/OpenEarsSampleApp\ Tests

    Undefined symbols for architecture x86_64:
    “_OBJC_CLASS_$_SaveThatWaveController”, referenced from:
    objc-class-ref in ViewController.o
    ld: symbol(s) not found for architecture x86_64
    clang: error: linker command failed with exit code 1 (use -v to see invocation)

    #1026617
    Halle Winkler
    Politepix

    Does it work when you compile for a device? I checked the demo version of SaveThatWave and it definitely has that architecture in it, so the possibilities are that that -ObjC linker flag isn’t added to the target that is being run, or the entry for the target in the Framework Search Paths build setting that was automatically created at the time of adding the framework is incorrect or has been mangled by Xcode (this is an occasional Xcode bug), or the framework wasn’t added to the target that is being run. Otherwise there’s nothing that should be preventing it from working. Since it fails at runtime rather than compile time, I’d be suspicious of the linker flag.

    #1026622
    jepjep7
    Participant

    Thanks- I keep trying, but it does not compile for my device.
    “Other Linker Flags” is set to -ObjC in the Build Settings of the OpenEarsSampleApp target.
    I checked “Framework Search Path”, and it had the “/Users/Joe/Downloads/OpenEarsDistribution/Framework” path listed.
    The SaveThatWaveDemo.framework is listed in the Framework folder and is also in the filesystem.
    I am running Xcode 6.4.

    #1026623
    Halle Winkler
    Politepix

    But is the plugin framework added to your target? It will show a checkmark next to the app target in the right-hand info pane.

    #1026624
    jepjep7
    Participant

    It now complies and runs on my device, but it has a warning:

    ld: warning: directory not found for option ‘-F”/Users/Joe/Downloads/OpenEarsDistribution/OpenEarsSampleApp/../Framework/”‘

    #1026637
    jepjep7
    Participant

    Xcode quit unexpectedly today, and now there are no more linker warnings when I run SaveThatWave on my device! I suspect I should have deleted the Openears Demo from my iPhone prior to compiling and cleaned and restarted Xcode. Anyway, I am on to the next step!

    I used the following line and then tested SaveThatWave on my iPhone:
    [self.saveThatWaveController startSessionDebugRecord];

    Where is the resultant WAV file stored? I cannot find the “caches” directory on my Mac. Thanks.

    #1026639
    Halle Winkler
    Politepix

    It’s stored on the device, in the app’s caches directory. Stack Overflow should have lots of help in getting at your app files via Xcode.

    #1026836
    jepjep7
    Participant

    Here is my replication case. I am using an iPhone 6 running OS9 and Xcode 7.

    https://www.dropbox.com/s/bnsk2i4j1q0r31r/test3.wav?dl=0

    Log:
    2015-09-20 23:15:45.655 OpenEarsSampleApp[4345:1045344] Starting OpenEars logging for OpenEars version 2.041 on 64-bit device (or build): iPhone running iOS version: 9.000000
    2015-09-20 23:15:45.656 OpenEarsSampleApp[4345:1045344] Creating shared instance of OEPocketsphinxController
    2015-09-20 23:15:45.705 OpenEarsSampleApp[4345:1045344] Starting dynamic language model generation

    INFO: cmd_ln.c(703): Parsing command line:
    sphinx_lm_convert \
    -i /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.arpa \
    -o /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.DMP

    Current configuration:
    [NAME] [DEFLT] [VALUE]
    -case
    -debug 0
    -help no no
    -i /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.arpa
    -ifmt
    -logbase 1.0001 1.000100e+00
    -mmap no no
    -o /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.DMP
    -ofmt

    INFO: ngram_model_arpa.c(503): ngrams 1=11, 2=18, 3=9
    INFO: ngram_model_arpa.c(135): Reading unigrams
    INFO: ngram_model_arpa.c(542): 11 = #unigrams created
    INFO: ngram_model_arpa.c(195): Reading bigrams
    INFO: ngram_model_arpa.c(560): 18 = #bigrams created
    INFO: ngram_model_arpa.c(561): 3 = #prob2 entries
    INFO: ngram_model_arpa.c(569): 3 = #bo_wt2 entries
    INFO: ngram_model_arpa.c(292): Reading trigrams
    INFO: ngram_model_arpa.c(582): 9 = #trigrams created
    INFO: ngram_model_arpa.c(583): 2 = #prob3 entries
    INFO: ngram_model_dmp.c(518): Building DMP model…
    INFO: ngram_model_dmp.c(548): 11 = #unigrams created
    INFO: ngram_model_dmp.c(649): 18 = #bigrams created
    INFO: ngram_model_dmp.c(650): 3 = #prob2 entries
    INFO: ngram_model_dmp.c(657): 3 = #bo_wt2 entries
    INFO: ngram_model_dmp.c(661): 9 = #trigrams created
    INFO: ngram_model_dmp.c(662): 2 = #prob3 entries
    2015-09-20 23:15:45.750 OpenEarsSampleApp[4345:1045344] Done creating language model with CMUCLMTK in 0.045052 seconds.
    2015-09-20 23:15:45.798 OpenEarsSampleApp[4345:1045344] I’m done running performDictionaryLookup and it took 0.033995 seconds
    2015-09-20 23:15:45.806 OpenEarsSampleApp[4345:1045344] I’m done running dynamic language model generation and it took 0.144854 seconds
    2015-09-20 23:15:45.814 OpenEarsSampleApp[4345:1045344] Starting dynamic language model generation

    INFO: cmd_ln.c(703): Parsing command line:
    sphinx_lm_convert \
    -i /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/SecondOpenEarsDynamicLanguageModel.arpa \
    -o /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/SecondOpenEarsDynamicLanguageModel.DMP

    Current configuration:
    [NAME] [DEFLT] [VALUE]
    -case
    -debug 0
    -help no no
    -i /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/SecondOpenEarsDynamicLanguageModel.arpa
    -ifmt
    -logbase 1.0001 1.000100e+00
    -mmap no no
    -o /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/SecondOpenEarsDynamicLanguageModel.DMP
    -ofmt

    INFO: ngram_model_arpa.c(503): ngrams 1=12, 2=19, 3=10
    INFO: ngram_model_arpa.c(135): Reading unigrams
    INFO: ngram_model_arpa.c(542): 12 = #unigrams created
    INFO: ngram_model_arpa.c(195): Reading bigrams
    INFO: ngram_model_arpa.c(560): 19 = #bigrams created
    INFO: ngram_model_arpa.c(561): 3 = #prob2 entries
    INFO: ngram_model_arpa.c(569): 3 = #bo_wt2 entries
    INFO: ngram_model_arpa.c(292): Reading trigrams
    INFO: ngram_model_arpa.c(582): 10 = #trigrams created
    INFO: ngram_model_arpa.c(583): 2 = #prob3 entries
    INFO: ngram_model_dmp.c(518): Building DMP model…
    INFO: ngram_model_dmp.c(548): 12 = #unigrams created
    INFO: ngram_model_dmp.c(649): 19 = #bigrams created
    INFO: ngram_model_dmp.c(650): 3 = #prob2 entries
    INFO: ngram_model_dmp.c(657): 3 = #bo_wt2 entries
    INFO: ngram_model_dmp.c(661): 10 = #trigrams created
    INFO: ngram_model_dmp.c(662): 2 = #prob3 entries
    2015-09-20 23:15:45.880 OpenEarsSampleApp[4345:1045344] Done creating language model with CMUCLMTK in 0.065760 seconds.
    2015-09-20 23:15:45.915 OpenEarsSampleApp[4345:1045344] The word QUIDNUNC was not found in the dictionary /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/LanguageModelGeneratorLookupList.text/LanguageModelGeneratorLookupList.text.
    2015-09-20 23:15:45.915 OpenEarsSampleApp[4345:1045344] Now using the fallback method to look up the word QUIDNUNC
    2015-09-20 23:15:45.916 OpenEarsSampleApp[4345:1045344] If this is happening more frequently than you would expect, the most likely cause for it is since you are using the English phonetic lookup dictionary is that your words are not in English or aren’t dictionary words.
    2015-09-20 23:15:45.916 OpenEarsSampleApp[4345:1045344] Using convertGraphemes for the word or phrase QUIDNUNC which doesn’t appear in the dictionary
    2015-09-20 23:15:45.938 OpenEarsSampleApp[4345:1045344] I’m done running performDictionaryLookup and it took 0.053308 seconds
    2015-09-20 23:15:45.945 OpenEarsSampleApp[4345:1045344] I’m done running dynamic language model generation and it took 0.138591 seconds
    2015-09-20 23:15:45.945 OpenEarsSampleApp[4345:1045344]

    Welcome to the OpenEars sample project. This project understands the words:
    BACKWARD,
    CHANGE,
    FORWARD,
    GO,
    LEFT,
    MODEL,
    RIGHT,
    TURN,
    and if you say “CHANGE MODEL” it will switch to its dynamically-generated model which understands the words:
    CHANGE,
    MODEL,
    MONDAY,
    TUESDAY,
    WEDNESDAY,
    THURSDAY,
    FRIDAY,
    SATURDAY,
    SUNDAY,
    QUIDNUNC
    2015-09-20 23:15:45.946 OpenEarsSampleApp[4345:1045344] Attempting to start listening session from startListeningWithLanguageModelAtPath:
    2015-09-20 23:15:45.950 OpenEarsSampleApp[4345:1045344] User gave mic permission for this app.
    2015-09-20 23:15:45.950 OpenEarsSampleApp[4345:1045344] setSecondsOfSilence wasn’t set, using default of 0.700000.
    2015-09-20 23:15:45.951 OpenEarsSampleApp[4345:1045344] Successfully started listening session from startListeningWithLanguageModelAtPath:
    2015-09-20 23:15:45.951 OpenEarsSampleApp[4345:1045401] Starting listening.
    2015-09-20 23:15:45.951 OpenEarsSampleApp[4345:1045401] about to set up audio session
    2015-09-20 23:15:45.952 OpenEarsSampleApp[4345:1045401] Creating audio session with default settings.
    2015-09-20 23:15:46.323 OpenEarsSampleApp[4345:1045419] Audio route has changed for the following reason:
    2015-09-20 23:15:46.331 OpenEarsSampleApp[4345:1045419] There was a category change. The new category is AVAudioSessionCategoryPlayAndRecord
    2015-09-20 23:15:46.335 OpenEarsSampleApp[4345:1045419] This is not a case in which OpenEars notifies of a route change. At the close of this function, the new audio route is —SpeakerMicrophoneBuiltIn—. The previous route before changing to this route was <AVAudioSessionRouteDescription: 0x146d47340,
    inputs = (
    “<AVAudioSessionPortDescription: 0x146d5f180, type = MicrophoneBuiltIn; name = iPhone Microphone; UID = Built-In Microphone; selectedDataSource = Front>”
    );
    outputs = (
    “<AVAudioSessionPortDescription: 0x146d6ec50, type = Speaker; name = Speaker; UID = Speaker; selectedDataSource = (null)>”
    )>.
    2015-09-20 23:15:46.532 OpenEarsSampleApp[4345:1045401] done starting audio unit
    INFO: cmd_ln.c(703): Parsing command line:
    \
    -lm /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.DMP \
    -vad_prespeech 10 \
    -vad_postspeech 69 \
    -vad_threshold 3.000000 \
    -remove_noise yes \
    -remove_silence yes \
    -bestpath yes \
    -lw 6.500000 \
    -dict /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.dic \
    -hmm /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle

    Current configuration:
    [NAME] [DEFLT] [VALUE]
    -agc none none
    -agcthresh 2.0 2.000000e+00
    -allphone
    -allphone_ci no no
    -alpha 0.97 9.700000e-01
    -argfile
    -ascale 20.0 2.000000e+01
    -aw 1 1
    -backtrace no no
    -beam 1e-48 1.000000e-48
    -bestpath yes yes
    -bestpathlw 9.5 9.500000e+00

    -ceplen 13 13
    -cmn current current
    -cmninit 8.0 8.0
    -compallsen no no
    -debug 0
    -dict /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.dic
    -dictcase no no
    -dither no no
    -doublebw no no
    -ds 1 1
    -fdict
    -feat 1s_c_d_dd 1s_c_d_dd
    -featparams
    -fillprob 1e-8 1.000000e-08
    -frate 100 100
    -fsg
    -fsgusealtpron yes yes
    -fsgusefiller yes yes
    -fwdflat yes yes
    -fwdflatbeam 1e-64 1.000000e-64
    -fwdflatefwid 4 4
    -fwdflatlw 8.5 8.500000e+00
    -fwdflatsfwin 25 25
    -fwdflatwbeam 7e-29 7.000000e-29
    -fwdtree yes yes
    -hmm /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle
    -input_endian little little
    -jsgf
    -keyphrase
    -kws
    -kws_delay 10 10
    -kws_plp 1e-1 1.000000e-01
    -kws_threshold 1 1.000000e+00
    -latsize 5000 5000
    -lda
    -ldadim 0 0
    -lifter 0 0
    -lm /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.DMP
    -lmctl
    -lmname
    -logbase 1.0001 1.000100e+00
    -logfn
    -logspec no no
    -lowerf 133.33334 1.333333e+02
    -lpbeam 1e-40 1.000000e-40
    -lponlybeam 7e-29 7.000000e-29
    -lw 6.5 6.500000e+00
    -maxhmmpf 30000 30000
    -maxwpf -1 -1
    -mdef
    -mean
    -mfclogdir
    -min_endfr 0 0
    -mixw
    -mixwfloor 0.0000001 1.000000e-07
    -mllr
    -mmap yes yes
    -ncep 13 13
    -nfft 512 512
    -nfilt 40 40
    -nwpen 1.0 1.000000e+00
    -pbeam 1e-48 1.000000e-48
    -pip 1.0 1.000000e+00
    -pl_beam 1e-10 1.000000e-10
    -pl_pbeam 1e-10 1.000000e-10
    -pl_pip 1.0 1.000000e+00
    -pl_weight 3.0 3.000000e+00
    -pl_window 5 5
    -rawlogdir
    -remove_dc no no
    -remove_noise yes yes
    -remove_silence yes yes
    -round_filters yes yes
    -samprate 16000 1.600000e+04
    -seed -1 -1
    -sendump
    -senlogdir
    -senmgau
    -silprob 0.005 5.000000e-03
    -smoothspec no no
    -svspec
    -tmat
    -tmatfloor 0.0001 1.000000e-04
    -topn 4 4
    -topn_beam 0 0
    -toprule
    -transform legacy legacy
    -unit_area yes yes
    -upperf 6855.4976 6.855498e+03
    -uw 1.0 1.000000e+00
    -vad_postspeech 50 69
    -vad_prespeech 20 10
    -vad_startspeech 10 10
    -vad_threshold 2.0 3.000000e+00
    -var
    -varfloor 0.0001 1.000000e-04
    -varnorm no no
    -verbose no no
    -warp_params
    -warp_type inverse_linear inverse_linear
    -wbeam 7e-29 7.000000e-29
    -wip 0.65 6.500000e-01
    -wlen 0.025625 2.562500e-02

    INFO: cmd_ln.c(703): Parsing command line:
    \
    -nfilt 25 \
    -lowerf 130 \
    -upperf 6800 \
    -feat 1s_c_d_dd \
    -svspec 0-12/13-25/26-38 \
    -agc none \
    -cmn current \
    -varnorm no \
    -transform dct \
    -lifter 22 \
    -cmninit 40

    Current configuration:
    [NAME] [DEFLT] [VALUE]
    -agc none none
    -agcthresh 2.0 2.000000e+00
    -alpha 0.97 9.700000e-01
    -ceplen 13 13
    -cmn current current
    -cmninit 8.0 40
    -dither no no
    -doublebw no no
    -feat 1s_c_d_dd 1s_c_d_dd
    -frate 100 100
    -input_endian little little
    -lda
    -ldadim 0 0
    -lifter 0 22
    -logspec no no
    -lowerf 133.33334 1.300000e+02
    -ncep 13 13
    -nfft 512 512
    -nfilt 40 25
    -remove_dc no no
    -remove_noise yes yes
    -remove_silence yes yes
    -round_filters yes yes
    -samprate 16000 1.600000e+04
    -seed -1 -1
    -smoothspec no no
    -svspec 0-12/13-25/26-38
    -transform legacy dct
    -unit_area yes yes
    -upperf 6855.4976 6.800000e+03
    -vad_postspeech 50 69
    -vad_prespeech 20 10
    -vad_startspeech 10 10
    -vad_threshold 2.0 3.000000e+00
    -varnorm no no
    -verbose no no
    -warp_params
    -warp_type inverse_linear inverse_linear
    -wlen 0.025625 2.562500e-02

    INFO: acmod.c(252): Parsed model-specific feature parameters from /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/feat.params
    INFO: feat.c(715): Initializing feature stream to type: ‘1s_c_d_dd’, ceplen=13, CMN=’current’, VARNORM=’no’, AGC=’none’
    INFO: cmn.c(143): mean[0]= 12.00, mean[1..12]= 0.0
    INFO: acmod.c(171): Using subvector specification 0-12/13-25/26-38
    INFO: mdef.c(518): Reading model definition: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/mdef
    INFO: mdef.c(531): Found byte-order mark BMDF, assuming this is a binary mdef file
    INFO: bin_mdef.c(336): Reading binary model definition: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/mdef
    INFO: bin_mdef.c(516): 46 CI-phone, 168344 CD-phone, 3 emitstate/phone, 138 CI-sen, 6138 Sen, 32881 Sen-Seq
    INFO: tmat.c(206): Reading HMM transition probability matrices: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/transition_matrices
    INFO: acmod.c(124): Attempting to use PTM computation module
    INFO: ms_gauden.c(198): Reading mixture gaussian parameter: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/means
    INFO: ms_gauden.c(292): 1 codebook, 3 feature, size:
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(198): Reading mixture gaussian parameter: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/variances
    INFO: ms_gauden.c(292): 1 codebook, 3 feature, size:
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(354): 0 variance values floored
    INFO: ptm_mgau.c(805): Number of codebooks doesn’t match number of ciphones, doesn’t look like PTM: 1 != 46
    INFO: acmod.c(126): Attempting to use semi-continuous computation module
    INFO: ms_gauden.c(198): Reading mixture gaussian parameter: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/means
    INFO: ms_gauden.c(292): 1 codebook, 3 feature, size:
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(198): Reading mixture gaussian parameter: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/variances
    INFO: ms_gauden.c(292): 1 codebook, 3 feature, size:
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(294): 512×13
    INFO: ms_gauden.c(354): 0 variance values floored
    INFO: s2_semi_mgau.c(904): Loading senones from dump file /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/sendump
    INFO: s2_semi_mgau.c(928): BEGIN FILE FORMAT DESCRIPTION
    INFO: s2_semi_mgau.c(991): Rows: 512, Columns: 6138
    INFO: s2_semi_mgau.c(1023): Using memory-mapped I/O for senones
    INFO: s2_semi_mgau.c(1294): Maximum top-N: 4 Top-N beams: 0 0 0
    INFO: phone_loop_search.c(114): State beam -225 Phone exit beam -225 Insertion penalty 0
    INFO: dict.c(320): Allocating 4115 * 32 bytes (128 KiB) for word entries
    INFO: dict.c(333): Reading main dictionary: /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches/FirstOpenEarsDynamicLanguageModel.dic
    INFO: dict.c(213): Allocated 0 KiB for strings, 0 KiB for phones
    INFO: dict.c(336): 10 words read
    INFO: dict.c(358): Reading filler dictionary: /var/mobile/Containers/Bundle/Application/772400B2-0D97-4986-BA6E-48D839C002FC/OpenEarsSampleApp.app/AcousticModelEnglish.bundle/noisedict
    INFO: dict.c(213): Allocated 0 KiB for strings, 0 KiB for phones
    INFO: dict.c(361): 9 words read
    INFO: dict2pid.c(396): Building PID tables for dictionary
    INFO: dict2pid.c(406): Allocating 46^3 * 2 bytes (190 KiB) for word-initial triphones
    INFO: dict2pid.c(132): Allocated 51152 bytes (49 KiB) for word-final triphones
    INFO: dict2pid.c(196): Allocated 51152 bytes (49 KiB) for single-phone word triphones
    INFO: ngram_model_arpa.c(77): No \data\ mark in LM file
    INFO: ngram_model_dmp.c(166): Will use memory-mapped I/O for LM file
    INFO: ngram_model_dmp.c(220): ngrams 1=11, 2=18, 3=9
    INFO: ngram_model_dmp.c(266): 11 = LM.unigrams(+trailer) read
    INFO: ngram_model_dmp.c(312): 18 = LM.bigrams(+trailer) read
    INFO: ngram_model_dmp.c(338): 9 = LM.trigrams read
    INFO: ngram_model_dmp.c(363): 3 = LM.prob2 entries read
    INFO: ngram_model_dmp.c(383): 3 = LM.bo_wt2 entries read
    INFO: ngram_model_dmp.c(403): 2 = LM.prob3 entries read
    INFO: ngram_model_dmp.c(431): 1 = LM.tseg_base entries read
    INFO: ngram_model_dmp.c(487): 11 = ascii word strings read
    INFO: ngram_search_fwdtree.c(99): 10 unique initial diphones
    INFO: ngram_search_fwdtree.c(148): 0 root, 0 non-root channels, 10 single-phone words
    INFO: ngram_search_fwdtree.c(186): Creating search tree
    INFO: ngram_search_fwdtree.c(192): before: 0 root, 0 non-root channels, 10 single-phone words
    INFO: ngram_search_fwdtree.c(326): after: max nonroot chan increased to 140
    INFO: ngram_search_fwdtree.c(339): after: 10 root, 12 non-root channels, 9 single-phone words
    INFO: ngram_search_fwdflat.c(157): fwdflat: min_ef_width = 4, max_sf_win = 25
    2015-09-20 23:15:46.635 OpenEarsSampleApp[4345:1045401] Listening.
    2015-09-20 23:15:46.636 OpenEarsSampleApp[4345:1045401] Project has these words or phrases in its dictionary:
    EIGHT
    FIVE
    FOUR
    NINE
    ONE
    ONE(2)
    SEVEN
    SIX
    THREE
    TWO

    2015-09-20 23:15:46.636 OpenEarsSampleApp[4345:1045401] Recognition loop has started
    2015-09-20 23:15:46.638 OpenEarsSampleApp[4345:1045344] the path is /var/mobile/Containers/Data/Application/595AF706-7114-41EC-B60F-29E57BD7A1B3/Library/Caches
    2015-09-20 23:15:46.672 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx is now listening.
    2015-09-20 23:15:46.674 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx started.
    2015-09-20 23:15:51.307 OpenEarsSampleApp[4345:1045401] Speech detected…
    2015-09-20 23:15:51.308 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has detected speech.
    2015-09-20 23:15:53.040 OpenEarsSampleApp[4345:1045404] End of speech detected…
    2015-09-20 23:15:53.041 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has detected a second of silence, concluding an utterance.
    INFO: cmn_prior.c(131): cmn_prior_update: from < 40.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 >
    INFO: cmn_prior.c(149): cmn_prior_update: to < 42.12 20.93 -15.38 14.24 -16.31 -4.51 -12.58 -2.26 -7.42 -4.90 -0.29 6.59 -8.43 >
    INFO: ngram_search_fwdtree.c(1553): 1536 words recognized (8/fr)
    INFO: ngram_search_fwdtree.c(1555): 29789 senones evaluated (158/fr)
    INFO: ngram_search_fwdtree.c(1559): 15455 channels searched (82/fr), 1840 1st, 12041 last
    INFO: ngram_search_fwdtree.c(1562): 1923 words for which last channels evaluated (10/fr)
    INFO: ngram_search_fwdtree.c(1564): 466 candidate words for entering last phone (2/fr)
    INFO: ngram_search_fwdtree.c(1567): fwdtree 0.35 CPU 0.187 xRT
    INFO: ngram_search_fwdtree.c(1570): fwdtree 6.16 wall 3.279 xRT
    INFO: ngram_search_fwdflat.c(302): Utterance vocabulary contains 6 words
    INFO: ngram_search_fwdflat.c(948): 1659 words recognized (9/fr)
    INFO: ngram_search_fwdflat.c(950): 24944 senones evaluated (133/fr)
    INFO: ngram_search_fwdflat.c(952): 13878 channels searched (73/fr)
    INFO: ngram_search_fwdflat.c(954): 2173 words searched (11/fr)
    INFO: ngram_search_fwdflat.c(957): 434 word transitions (2/fr)
    INFO: ngram_search_fwdflat.c(960): fwdflat 0.04 CPU 0.020 xRT
    INFO: ngram_search_fwdflat.c(963): fwdflat 0.04 wall 0.021 xRT
    INFO: ngram_search.c(1280): lattice start node <s>.0 end node </s>.138
    INFO: ngram_search.c(1306): Eliminated 5 nodes before end node
    INFO: ngram_search.c(1411): Lattice has 616 nodes, 1752 links
    INFO: ps_lattice.c(1380): Bestpath score: -18594
    INFO: ps_lattice.c(1384): Normalizer P(O) = alpha(</s>:138:186) = -1207819
    INFO: ps_lattice.c(1441): Joint P(O,S) = -1286424 P(S|O) = -78605
    INFO: ngram_search.c(899): bestpath 0.01 CPU 0.004 xRT
    INFO: ngram_search.c(902): bestpath 0.01 wall 0.004 xRT
    2015-09-20 23:15:53.093 OpenEarsSampleApp[4345:1045404] Pocketsphinx heard “THREE FIVE FIVE” with a score of (-78605) and an utterance ID of 0.
    2015-09-20 23:15:53.094 OpenEarsSampleApp[4345:1045344] Flite sending interrupt speech request.
    2015-09-20 23:15:53.094 OpenEarsSampleApp[4345:1045344] Local callback: The received hypothesis is THREE FIVE FIVE with a score of -78605 and an ID of 0
    2015-09-20 23:15:53.096 OpenEarsSampleApp[4345:1045344] I’m running flite
    2015-09-20 23:15:53.209 OpenEarsSampleApp[4345:1045344] I’m done running flite and it took 0.113386 seconds
    2015-09-20 23:15:53.210 OpenEarsSampleApp[4345:1045344] Flite audio player was nil when referenced so attempting to allocate a new audio player.
    2015-09-20 23:15:53.210 OpenEarsSampleApp[4345:1045344] Loading speech data for Flite concluded successfully.
    2015-09-20 23:15:53.301 OpenEarsSampleApp[4345:1045344] Flite sending suspend recognition notification.
    2015-09-20 23:15:53.303 OpenEarsSampleApp[4345:1045344] Local callback: Flite has started speaking
    2015-09-20 23:15:53.308 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has suspended recognition.
    2015-09-20 23:15:55.349 OpenEarsSampleApp[4345:1045344] AVAudioPlayer did finish playing with success flag of 1
    2015-09-20 23:15:55.501 OpenEarsSampleApp[4345:1045344] Flite sending resume recognition notification.
    2015-09-20 23:15:56.003 OpenEarsSampleApp[4345:1045344] Local callback: Flite has finished speaking
    2015-09-20 23:15:56.008 OpenEarsSampleApp[4345:1045344] setSecondsOfSilence wasn’t set, using default of 0.700000.
    2015-09-20 23:15:56.009 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has resumed recognition.
    INFO: cmn_prior.c(131): cmn_prior_update: from < 42.12 20.93 -15.38 14.24 -16.31 -4.51 -12.58 -2.26 -7.42 -4.90 -0.29 6.59 -8.43 >
    INFO: cmn_prior.c(149): cmn_prior_update: to < 42.12 20.93 -15.38 14.24 -16.31 -4.51 -12.58 -2.26 -7.42 -4.90 -0.29 6.59 -8.43 >
    INFO: ngram_search_fwdflat.c(302): Utterance vocabulary contains 0 words
    2015-09-20 23:15:58.301 OpenEarsSampleApp[4345:1045401] Speech detected…
    2015-09-20 23:15:58.302 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has detected speech.
    2015-09-20 23:15:59.832 OpenEarsSampleApp[4345:1045401] End of speech detected…
    2015-09-20 23:15:59.833 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has detected a second of silence, concluding an utterance.
    INFO: cmn_prior.c(131): cmn_prior_update: from < 42.12 20.93 -15.38 14.24 -16.31 -4.51 -12.58 -2.26 -7.42 -4.90 -0.29 6.59 -8.43 >
    INFO: cmn_prior.c(149): cmn_prior_update: to < 42.18 20.48 -12.38 8.20 -16.77 1.20 -14.39 -5.20 -6.01 -4.37 -1.41 5.09 -7.03 >
    INFO: ngram_search_fwdtree.c(1553): 1424 words recognized (9/fr)
    INFO: ngram_search_fwdtree.c(1555): 27335 senones evaluated (167/fr)
    INFO: ngram_search_fwdtree.c(1559): 14207 channels searched (86/fr), 1600 1st, 11096 last
    INFO: ngram_search_fwdtree.c(1562): 1683 words for which last channels evaluated (10/fr)
    INFO: ngram_search_fwdtree.c(1564): 496 candidate words for entering last phone (3/fr)
    INFO: ngram_search_fwdtree.c(1567): fwdtree 0.25 CPU 0.151 xRT
    INFO: ngram_search_fwdtree.c(1570): fwdtree 3.73 wall 2.276 xRT
    INFO: ngram_search_fwdflat.c(302): Utterance vocabulary contains 6 words
    INFO: ngram_search_fwdflat.c(948): 1538 words recognized (9/fr)
    INFO: ngram_search_fwdflat.c(950): 27476 senones evaluated (168/fr)
    INFO: ngram_search_fwdflat.c(952): 16006 channels searched (97/fr)
    INFO: ngram_search_fwdflat.c(954): 2008 words searched (12/fr)
    INFO: ngram_search_fwdflat.c(957): 490 word transitions (2/fr)
    INFO: ngram_search_fwdflat.c(960): fwdflat 0.03 CPU 0.020 xRT
    INFO: ngram_search_fwdflat.c(963): fwdflat 0.04 wall 0.022 xRT
    INFO: ngram_search.c(1280): lattice start node <s>.0 end node </s>.111
    INFO: ngram_search.c(1306): Eliminated 3 nodes before end node
    INFO: ngram_search.c(1411): Lattice has 568 nodes, 2461 links
    INFO: ps_lattice.c(1380): Bestpath score: -15511
    INFO: ps_lattice.c(1384): Normalizer P(O) = alpha(</s>:111:162) = -1087553
    INFO: ps_lattice.c(1441): Joint P(O,S) = -1174585 P(S|O) = -87032
    INFO: ngram_search.c(899): bestpath 0.01 CPU 0.006 xRT
    INFO: ngram_search.c(902): bestpath 0.01 wall 0.005 xRT
    2015-09-20 23:15:59.884 OpenEarsSampleApp[4345:1045401] Pocketsphinx heard “THREE FIVE” with a score of (-87032) and an utterance ID of 1.
    2015-09-20 23:15:59.885 OpenEarsSampleApp[4345:1045344] Flite sending interrupt speech request.
    2015-09-20 23:15:59.885 OpenEarsSampleApp[4345:1045344] Local callback: The received hypothesis is THREE FIVE with a score of -87032 and an ID of 1
    2015-09-20 23:15:59.887 OpenEarsSampleApp[4345:1045344] I’m running flite
    2015-09-20 23:15:59.962 OpenEarsSampleApp[4345:1045344] I’m done running flite and it took 0.074690 seconds
    2015-09-20 23:15:59.962 OpenEarsSampleApp[4345:1045344] Flite audio player was nil when referenced so attempting to allocate a new audio player.
    2015-09-20 23:15:59.962 OpenEarsSampleApp[4345:1045344] Loading speech data for Flite concluded successfully.
    2015-09-20 23:15:59.987 OpenEarsSampleApp[4345:1045344] Flite sending suspend recognition notification.
    2015-09-20 23:15:59.989 OpenEarsSampleApp[4345:1045344] Local callback: Flite has started speaking
    2015-09-20 23:15:59.993 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has suspended recognition.
    2015-09-20 23:16:01.664 OpenEarsSampleApp[4345:1045344] AVAudioPlayer did finish playing with success flag of 1
    2015-09-20 23:16:01.816 OpenEarsSampleApp[4345:1045344] Flite sending resume recognition notification.
    2015-09-20 23:16:02.318 OpenEarsSampleApp[4345:1045344] Local callback: Flite has finished speaking
    2015-09-20 23:16:02.325 OpenEarsSampleApp[4345:1045344] setSecondsOfSilence wasn’t set, using default of 0.700000.
    2015-09-20 23:16:02.326 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has resumed recognition.
    INFO: cmn_prior.c(131): cmn_prior_update: from < 42.18 20.48 -12.38 8.20 -16.77 1.20 -14.39 -5.20 -6.01 -4.37 -1.41 5.09 -7.03 >
    INFO: cmn_prior.c(149): cmn_prior_update: to < 42.18 20.48 -12.38 8.20 -16.77 1.20 -14.39 -5.20 -6.01 -4.37 -1.41 5.09 -7.03 >
    INFO: ngram_search_fwdflat.c(302): Utterance vocabulary contains 0 words
    2015-09-20 23:16:05.724 OpenEarsSampleApp[4345:1045401] Speech detected…
    2015-09-20 23:16:05.725 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has detected speech.
    2015-09-20 23:16:06.364 OpenEarsSampleApp[4345:1045404] End of speech detected…
    INFO: cmn_prior.c(131): cmn_prior_update: from < 42.18 20.48 -12.38 8.20 -16.77 1.20 -14.39 -5.20 -6.01 -4.37 -1.41 5.09 -7.03 >
    2015-09-20 23:16:06.365 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has detected a second of silence, concluding an utterance.
    INFO: cmn_prior.c(149): cmn_prior_update: to < 39.98 17.73 -12.48 7.80 -14.05 1.06 -13.80 -3.82 -5.37 -5.05 -1.22 3.75 -5.35 >
    INFO: ngram_search_fwdtree.c(1553): 766 words recognized (10/fr)
    INFO: ngram_search_fwdtree.c(1555): 10758 senones evaluated (136/fr)
    INFO: ngram_search_fwdtree.c(1559): 5330 channels searched (67/fr), 750 1st, 3921 last
    INFO: ngram_search_fwdtree.c(1562): 813 words for which last channels evaluated (10/fr)
    INFO: ngram_search_fwdtree.c(1564): 203 candidate words for entering last phone (2/fr)
    INFO: ngram_search_fwdtree.c(1567): fwdtree 0.21 CPU 0.266 xRT
    INFO: ngram_search_fwdtree.c(1570): fwdtree 3.85 wall 4.874 xRT
    INFO: ngram_search_fwdflat.c(302): Utterance vocabulary contains 5 words
    INFO: ngram_search_fwdflat.c(948): 762 words recognized (10/fr)
    INFO: ngram_search_fwdflat.c(950): 9527 senones evaluated (121/fr)
    INFO: ngram_search_fwdflat.c(952): 5360 channels searched (67/fr)
    INFO: ngram_search_fwdflat.c(954): 858 words searched (10/fr)
    INFO: ngram_search_fwdflat.c(957): 173 word transitions (2/fr)
    INFO: ngram_search_fwdflat.c(960): fwdflat 0.01 CPU 0.017 xRT
    INFO: ngram_search_fwdflat.c(963): fwdflat 0.02 wall 0.026 xRT
    INFO: ngram_search.c(1280): lattice start node <s>.0 end node </s>.21
    INFO: ngram_search.c(1306): Eliminated 6 nodes before end node
    INFO: ngram_search.c(1411): Lattice has 283 nodes, 269 links
    INFO: ps_lattice.c(1380): Bestpath score: -3872
    INFO: ps_lattice.c(1384): Normalizer P(O) = alpha(</s>:21:77) = 1429791
    INFO: ps_lattice.c(1441): Joint P(O,S) = 1382256 P(S|O) = -47535
    INFO: ngram_search.c(899): bestpath 0.00 CPU 0.002 xRT
    INFO: ngram_search.c(902): bestpath 0.00 wall 0.003 xRT
    2015-09-20 23:16:06.395 OpenEarsSampleApp[4345:1045404] Pocketsphinx heard “TWO” with a score of (-47535) and an utterance ID of 2.
    2015-09-20 23:16:06.395 OpenEarsSampleApp[4345:1045344] Flite sending interrupt speech request.
    2015-09-20 23:16:06.396 OpenEarsSampleApp[4345:1045344] Local callback: The received hypothesis is TWO with a score of -47535 and an ID of 2
    2015-09-20 23:16:06.397 OpenEarsSampleApp[4345:1045344] I’m running flite
    2015-09-20 23:16:06.469 OpenEarsSampleApp[4345:1045344] I’m done running flite and it took 0.071670 seconds
    2015-09-20 23:16:06.470 OpenEarsSampleApp[4345:1045344] Flite audio player was nil when referenced so attempting to allocate a new audio player.
    2015-09-20 23:16:06.470 OpenEarsSampleApp[4345:1045344] Loading speech data for Flite concluded successfully.
    2015-09-20 23:16:06.488 OpenEarsSampleApp[4345:1045344] Flite sending suspend recognition notification.
    2015-09-20 23:16:06.490 OpenEarsSampleApp[4345:1045344] Local callback: Flite has started speaking
    2015-09-20 23:16:06.495 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has suspended recognition.
    2015-09-20 23:16:07.704 OpenEarsSampleApp[4345:1045344] AVAudioPlayer did finish playing with success flag of 1
    2015-09-20 23:16:07.856 OpenEarsSampleApp[4345:1045344] Flite sending resume recognition notification.
    2015-09-20 23:16:08.358 OpenEarsSampleApp[4345:1045344] Local callback: Flite has finished speaking
    2015-09-20 23:16:08.366 OpenEarsSampleApp[4345:1045344] setSecondsOfSilence wasn’t set, using default of 0.700000.
    2015-09-20 23:16:08.366 OpenEarsSampleApp[4345:1045344] Local callback: Pocketsphinx has resumed recognition.
    INFO: cmn_prior.c(131): cmn_prior_update: from < 39.98 17.73 -12.48 7.80 -14.05 1.06 -13.80 -3.82 -5.37 -5.05 -1.22 3.75 -5.35 >
    INFO: cmn_prior.c(149): cmn_prior_update: to < 39.98 17.73 -12.48 7.80 -14.05 1.06 -13.80 -3.82 -5.37 -5.05 -1.22 3.75 -5.35 >
    INFO: ngram_search_fwdflat.c(302): Utterance vocabulary contains 0 words

    #1026841
    Halle Winkler
    Politepix

    OK, I’ll take a look at this after the next update is out.

Viewing 18 posts - 1 through 18 (of 18 total)
  • You must be logged in to reply to this topic.