-
Notifications
You must be signed in to change notification settings - Fork 303
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Empty or incomplete hypotheses #667
Comments
Are you using the latest k2 (i.e., the master branch of k2)? |
I use k2 version 1.19.dev20220922 |
Could you try the latest one from the master? |
I have tried but it didn't reduce empty hypotheses |
I also try greedy search, and it output exact tokens, so that I think the problem is in fast_beam_search_nbest_LG |
Hi has anyone else experienced something like this? head -2 beam_search/errs-test-beam_size_4-epoch-50-avg-25-beam_search-beam-size-4.txt head -2 fast_beam_search/errs-test-beam_15.0_max_contexts_8_max_states_64-epoch-50-avg-25-beam-15.0-max-contexts-8-max-states-64.txt head -2 greedy_search/errs-test-greedy_search-epoch-50-avg-25-context-2-max-sym-per-frame-1.txt head -2 modified_beam_search/errs-test-beam_size_4-epoch-50-avg-25-modified_beam_search-beam-size-4.txt head -2 fast_beam_search_nbest/errs-test-beam_15.0_max_contexts_8_max_states_64_num_paths_100_nbest_scale_0.5-epoch-50-avg-25-beam-15.0-max-contexts-8-max-states-64-nbest-scale-0.5-num-paths-100.txt head -2 fast_beam_search_nbest_LG/errs-test-beam_20.0_max_contexts_8_max_states_64_num_paths_200_nbest_scale_0.5_ngram_lm_scale_0.01-epoch-50-avg-25-beam-20.0-max-contexts-8-max-states-64-nbest-scale-0.5-num-paths-200-ngram-lm-scale-0.01.txt there's a big drop in WER with fast_beam_search_nbest_LG, and no difference when a 2-gram or tri-gram is used |
Could you please check your errs-xxx file and see how many errors are caused by OOV words when LG is used? |
Out of 26379 words of the eval corpus, there are 438 OOV word occurrences w.r.t. the word list in L, which is 1.66% OOV ratio -when I use HLG based decoding, let's say 1best or nbest method, in conformer_ctc, I have a more reasonable 18.7-18.8 WER with the same L and G ex: 293b82f6-2407-4d08-8a27-93dc690c2313: ref=['dat', 'zou', 'niet', 'nodig', 'zijn', 'als', 'hij', 'in', 'deze', 'cockpit', 'zou', 'vliegen'] now, maybe what I could do is train your latest model, where encoder-ctc output is combined with transducer and the HLG decoding can be done on the ctc output and see what I obtain |
Do you mean it is not helpful for HLG decoding? |
that remark referred to fast_beam_search_nbest_LG when I use in G in first pass decoding in HLG with the conformer_ctc and then rescore with a 4-Gram (for example, whole-lattice-rescoring) results are improved (let's say 7-8% relative improvement, in this specific case) |
When I trained conformer stateless streaming mode (transducer_stateless2), in the decode phases, I met a situation like this issue (#403), that is decode with fast_beam_search_nbest_LG and LG graph gives a lot of empty hypotheses.
I tried to fix it by following the solution in (#403), but didn't find any use-max argument.
The text was updated successfully, but these errors were encountered: