Asr-language-processing-research-s2s-generation

来自cslt Wiki
2016年11月14日 (一) 07:23Zhangandy讨论 | 贡献的版本

(差异) ←上一版本 | 最后版本 (差异) | 下一版本→ (差异)
跳转至: 导航搜索

Main Idea

People

Yang Feng, Andi Zhang

Time Table

Week Work Plan Work Done
2016/11/07-2016/11/13
  • successfully run the code of "Neural machine translation by jointly learning to align and translate" on gpu
  • start working on model_step_1: linear trans->cosine->linear trans->softmax; start coding if time permits.
90%
2016/11/14-2016/11/20
  • coding on model_step_1
  • run & test the code
  • run NTM on fr-en data set
  • run NTM on paraphrase data set
2016/11/21-2016/11/27
  • continue working on model_step_1
  • start working on model_step_2: lstm->mn->softmax
2016/11/28-2016/12/04
  • coding & debug & run model_step_2
  • start working on model_step_3: joint training
2016/12/05-2016/12/11
  • coding & debug & run model_step_3
2016/12/12-2016/12/18
  • find ways to speed up the model if it is slow
2016/12/19-2016/12/25
  • select memory to optimize result
2016/12/26-2016/12/31
  • run final result
  • check any possible faults

Progress