“2013-12-13”版本间的差异

来自cslt Wiki
跳转至: 导航搜索
(以内容“== AM development == === Sparse DNN === * Optimal Brain Damage(OBD). # Online OBD held. # OBD + L1 norm start to investigation. * Efficient computing # Using MK...”创建新页面)
 
第15行: 第15行:
 
# There is minor difference between the MKL computing and direct computing. This means computing accuracy does not impact the ASR performance very much. This give some excuse for extremely sparse matrix construction.
 
# There is minor difference between the MKL computing and direct computing. This means computing accuracy does not impact the ASR performance very much. This give some excuse for extremely sparse matrix construction.
  
=== Efficient DNN training ==
+
=== Efficient DNN training ===
  
 
# Moment-based training. NN accuracy decreased with larger moment, but ASR performance increased.  
 
# Moment-based training. NN accuracy decreased with larger moment, but ASR performance increased.  

2013年12月13日 (五) 01:34的版本

AM development

Sparse DNN

  • Optimal Brain Damage(OBD).
  1. Online OBD held.
  2. OBD + L1 norm start to investigation.
  • Efficient computing
  1. Using MKL and CSR storage does not help much for sparse matrix computation. When the sparsity is 20%, the computing costs 2 times of the original time.
  2. Using matrix splitting can improve computing performance for sparse matrix. Using BSR (block sparse row), when the sparsity is 1/6, the same time cost was obtained.
  3. We can re-arrange the matrix structure and compose zero blocks by some smart approaches, leading to better computing speed.
  4. There is minor difference between the MKL computing and direct computing. This means computing accuracy does not impact the ASR performance very much. This give some excuse for extremely sparse matrix construction.

Efficient DNN training

  1. Moment-based training. NN accuracy decreased with larger moment, but ASR performance increased.
  2. Asymmetic window: left 20, right 5. NN accuracy increase by 7%.


Engine optimization

  • Investigating LOUDS FST.


LM development

NN LM

  • 3 iteration 500 M training done. 24 hours per iteration.
  • PPL 189 after 3 iterations.
  • NN-based CSLM merge done (10240*100*10240). The PPL and WER are both worse than the original 10 network outputs.
  • Need to investigate why the merge is not accurate.


Embedded development

  • Embedded stream mode on progress.


Speech QA

  • Text-based QA. 121/199 correction (with answers). 58 no answers(24 no attributes in db, 27 no recorder). 20 with incorrect answers(5 no answers in the db and obtain incorrect from the web, 8 no recorder and obtain incorrect from the web, 3 db error).
  • Speech-based QA. WER=8.70%. SEE=32.0%. Almost English wrong. Remove English SEE=27.1%.
  • Use all the correct input. 45.14% correct in all the input (18*199)
  • Will try to recover some ASR errors using QA.