“2017-3-6”版本间的差异
来自cslt Wiki
第4行: | 第4行: | ||
| rowspan="6"|2017/1/3 | | rowspan="6"|2017/1/3 | ||
|Yang Feng || | |Yang Feng || | ||
+ | * ran experiments on the CS-EN data set (200k pairs) with totally identical initialization as the paper. on the sampled 2k training sentences, the bleu is 19.5 (not converged yet). (the bleu on the test set expected to be 26) | ||
+ | * add the alpha and gamma score and do multi-task training. Without multi-task training, the loss didn't decline on the training data, but with multi-task training, the loss did decline. | ||
+ | * prepared for Huilan's inspection. | ||
|| | || | ||
+ | * Analyze the reason that the loss didn't decline with alpha and gamma | ||
+ | * test for multi-task training; | ||
+ | * improve the baseline for the CS-EN | ||
|- | |- | ||
|Jiyuan Zhang || | |Jiyuan Zhang || |
2017年3月6日 (一) 10:56的版本
Date | People | Last Week | This Week |
---|---|---|---|
2017/1/3 | Yang Feng |
|
|
Jiyuan Zhang |
|
| |
Andi Zhang |
|
||
Shiyue Zhang |
|
| |
Peilun Xiao |