Zhiyuan Thang 15-06-29

来自cslt Wiki
2015年6月30日 (二) 14:31Tangzy讨论 | 贡献的版本

(差异) ←上一版本 | 最后版本 (差异) | 下一版本→ (差异)
跳转至: 导航搜索

Last few weeks (lazy):

Presently got the conclusion that randoming the weights ahead of softmax of DNN before MPE training may be not helpful.

Using language vector to pre-train the hidden layers of DNN gives at least a little improvement.

have not realized leaky Rectifier and temporal Rectifier, CUDA programing may be needed.


This week:

realize the two kinds of Rectifier.