“Asr-language-processing-research-s2s-generation”版本间的差异

来自cslt Wiki
跳转至: 导航搜索
Time Table
Time Table
 
(相同用户的3个中间修订版本未显示)
第8行: 第8行:
 
!Week !! Work Plan !! Work Done  
 
!Week !! Work Plan !! Work Done  
 
|-
 
|-
| || ||
+
| 2016/11/07-2016/11/13 ||
 +
*successfully run the code of "Neural machine translation by jointly learning to align and translate" on gpu
 +
*start working on model_step_1: linear trans->cosine->linear trans->softmax; start coding if time permits.
 +
||90%
 
|-
 
|-
| || ||
+
| 2016/11/14-2016/11/20||
 +
*coding on model_step_1
 +
*run & test the code
 +
*run NTM on fr-en data set
 +
*run NTM on paraphrase data set
 +
||
 
|-
 
|-
| || ||
+
| 2016/11/21-2016/11/27 ||  
 +
*continue working on model_step_1
 +
*start working on model_step_2: lstm->mn->softmax
 +
||
 
|-
 
|-
| || ||
+
| 2016/11/28-2016/12/04 ||  
 +
*coding & debug & run model_step_2
 +
*start working on model_step_3: joint training
 +
||
 
|-
 
|-
 +
| 2016/12/05-2016/12/11 ||
 +
*coding & debug & run model_step_3
  
 +
||
 +
|-
 +
| 2016/12/12-2016/12/18 ||
 +
*find ways to speed up the model if it is slow
 +
||
 +
|-
 +
| 2016/12/19-2016/12/25 ||
 +
*select memory to optimize result
 +
||
 +
|-
 +
| 2016/12/26-2016/12/31 ||
 +
*run final result
 +
*check any possible faults
 +
||
 +
|-
  
 
|}
 
|}
  
 
==Progress==
 
==Progress==

2016年11月14日 (一) 07:23的最后版本

Main Idea

People

Yang Feng, Andi Zhang

Time Table

Week Work Plan Work Done
2016/11/07-2016/11/13
  • successfully run the code of "Neural machine translation by jointly learning to align and translate" on gpu
  • start working on model_step_1: linear trans->cosine->linear trans->softmax; start coding if time permits.
90%
2016/11/14-2016/11/20
  • coding on model_step_1
  • run & test the code
  • run NTM on fr-en data set
  • run NTM on paraphrase data set
2016/11/21-2016/11/27
  • continue working on model_step_1
  • start working on model_step_2: lstm->mn->softmax
2016/11/28-2016/12/04
  • coding & debug & run model_step_2
  • start working on model_step_3: joint training
2016/12/05-2016/12/11
  • coding & debug & run model_step_3
2016/12/12-2016/12/18
  • find ways to speed up the model if it is slow
2016/12/19-2016/12/25
  • select memory to optimize result
2016/12/26-2016/12/31
  • run final result
  • check any possible faults

Progress