<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://index.cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://index.cslt.org/mediawiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Liuaiting</id>
		<title>cslt Wiki - 用户贡献 [zh-cn]</title>
		<link rel="self" type="application/atom+xml" href="http://index.cslt.org/mediawiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Liuaiting"/>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/%E7%89%B9%E6%AE%8A:%E7%94%A8%E6%88%B7%E8%B4%A1%E7%8C%AE/Liuaiting"/>
		<updated>2026-04-14T04:25:03Z</updated>
		<subtitle>用户贡献</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Bi-monthly-2016-08-language_Language_group</id>
		<title>Bi-monthly-2016-08-language Language group</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Bi-monthly-2016-08-language_Language_group"/>
				<updated>2016-08-31T05:50:57Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;AitingLiu [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c0/Bimonthly_report_Aiting.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
Lishiyao [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/90/Lishiyao.pdf ppt]]&lt;br /&gt;
&lt;br /&gt;
Andy Zhang [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b6/Andy%27s_bimonthly_report.pdf slides]]&lt;br /&gt;
&lt;br /&gt;
Jiyuan Zhang [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/98/Bimonthly_report_JiyuanZhang.pptx slides]]&lt;br /&gt;
&lt;br /&gt;
Aodong Li [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/6e/Bi-monthly_report_aodong.pdf slides]]&lt;br /&gt;
&lt;br /&gt;
Group [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/52/NLP-bimonthly-report-20160831.pdf slides]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:Bimonthly_report_Aiting.pdf</id>
		<title>文件:Bimonthly report Aiting.pdf</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:Bimonthly_report_Aiting.pdf"/>
				<updated>2016-08-31T05:50:29Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-31T02:51:39Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Daily Report */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;5&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 || 18:00   || 7+ || test the Chatting model,count the number of startswith &amp;lt;END&amp;gt; sentences,and natural sentences.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 || 18:20  || 8 || learn the CNN to solve MNIST using tensorflow&lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li|| 10:30 || 16:30  || 5 || Read shared paper at the week before the last week&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/30&lt;br /&gt;
|Ziwei Bai|| 9:15 ||18:15   || 8 || 1、prepare for bimonthly report.2、test Chatting Model &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || finished version 1 of chapter 3 with the help of Aiting&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:15  || 18:15  || 8 || assist Andy to complete chapter3, prepare bimonthly report&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:15 || 18:15 || 8 || continue to learn tensorflow and train several models about mnist task&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;5&amp;quot;|2016/08/31&lt;br /&gt;
|Ziwei Bai||9:10 ||   ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   || || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:10 || 18:10 || 8 || bimonthly report&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:10 ||  ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li|| 08:20 ||  ||  || &lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
*surveyed the line works of neural parsing, neural Turing machine and variants of RBMs and fixed the research topic to jointly parsing and understanding with Turing machine&lt;br /&gt;
*learned tensorflow&lt;br /&gt;
*produced two versions of results of query paraphrase&lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
*Code: &lt;br /&gt;
  2. Rare word embedding on word2vec and the related in C&lt;br /&gt;
  3. Help Lantian for ICASSP&lt;br /&gt;
*Paper share: one for bilingual word embedding&lt;br /&gt;
*ML book: first version of chapter 2 with Aiting&lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
*finish part of chapter 5 kernel method&lt;br /&gt;
*retrain &amp;amp; test Chatting Model&lt;br /&gt;
*share 'the use of tensorflow'&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
*shared the principle of word2vec&lt;br /&gt;
*read several papers about comparison of sentence similarity &lt;br /&gt;
*wrote chapter 3 neural network of the ML book&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
*share two papers:&lt;br /&gt;
1.Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
&lt;br /&gt;
2.A Sentence Interaction Network for Modeling Dependence between Sentences&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/6c/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pptx slides]]&lt;br /&gt;
*write chapter2 linear model&lt;br /&gt;
*help Andy complete chapter3 neural network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
*prepare the data for question-pair task.&lt;br /&gt;
*use SRILM to train the language model and Moses to train the task&lt;br /&gt;
*share a paper about LSTM and do a presentation.&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/29 || || || 0.5 hours || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Bi-monthly-2016-08-language_Language_group</id>
		<title>Bi-monthly-2016-08-language Language group</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Bi-monthly-2016-08-language_Language_group"/>
				<updated>2016-08-31T00:51:19Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;AitingLiu [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/26/Bimonthly_report_AitingLiu.pptx ppt]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Bi-monthly-2016-08-language_Language_group</id>
		<title>Bi-monthly-2016-08-language Language group</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Bi-monthly-2016-08-language_Language_group"/>
				<updated>2016-08-31T00:50:52Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：以“AitingLiu [http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/26/Bimonthly_report_AitingLiu.pptx ppt]”为内容创建页面&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;AitingLiu [http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/26/Bimonthly_report_AitingLiu.pptx ppt]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:Bimonthly_report_AitingLiu.pptx</id>
		<title>文件:Bimonthly report AitingLiu.pptx</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:Bimonthly_report_AitingLiu.pptx"/>
				<updated>2016-08-31T00:49:59Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-30T09:13:43Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Daily Report */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;5&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 || 18:00   || 7+ || test the Chatting model,count the number of startswith &amp;lt;END&amp;gt; sentences,and natural sentences.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 || 18:20  || 8 || learn the CNN to solve MNIST using tensorflow&lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li|| 10:30 || 16:30  || 5 || Read shared paper at the week before the last week&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/30&lt;br /&gt;
|Ziwei Bai|| 9:15 ||   ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:15  || 18:15  || 8 || assist Andy to complete chapter3, prepare bimonthly report&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:15 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
*surveyed the line works of neural parsing, neural Turing machine and variants of RBMs and fixed the research topic to jointly parsing and understanding with Turing machine&lt;br /&gt;
*learned tensorflow&lt;br /&gt;
*produced two versions of results of query paraphrase&lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
*Code: &lt;br /&gt;
  1. sequence-to-sequence model with attention mechanism for chatting model and lyrics generation in theano and tensorflow&lt;br /&gt;
  2. Rare word embedding on word2vec and the related in C&lt;br /&gt;
  3. Help Lantian for ICASSP&lt;br /&gt;
*Paper sharing: two shared papers, one for rare word embedding, one for bilingual word embedding&lt;br /&gt;
*ML book: first version of chapter 2, linear model&lt;br /&gt;
*Read several papers about RNN, language model, word2vec, sequence-to-sequence model&lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
*finish part of chapter 5 kernel method&lt;br /&gt;
*retrain &amp;amp; test Chatting Model&lt;br /&gt;
*share 'the use of tensorflow'&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
*shared the principle of word2vec&lt;br /&gt;
*read several papers about comparison of sentence similarity &lt;br /&gt;
*wrote chapter 3 neural network of the ML book&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
*share two papers:&lt;br /&gt;
1.Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
&lt;br /&gt;
2.A Sentence Interaction Network for Modeling Dependence between Sentences&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/6c/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pptx slides]]&lt;br /&gt;
*write chapter2 linear model&lt;br /&gt;
*help Andy complete chapter3 neural network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
*prepare the data for question-pair task.&lt;br /&gt;
*use SRILM to train the language model and Moses to train the task&lt;br /&gt;
*share a paper about LSTM and do a presentation.&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/29 || || || 0.5 hours || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-30T09:11:51Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Daily Report */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 || 18:00   || 7+ || test the Chatting model,count the number of startswith &amp;lt;END&amp;gt; sentences,and natural sentences.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 || 18:20  || 8 || learn the CNN to solve MNIST using tensorflow&lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li|| 10:30 || 16:30  || 5 || Read shared paper at the week before the last week&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/30&lt;br /&gt;
|Ziwei Bai|| 9:15 ||   ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:15  || 18:15  || 8 || assist Andy to complete chapter3, prepare bimonthly report&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:15 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
*surveyed the line works of neural parsing, neural Turing machine and variants of RBMs and fixed the research topic to jointly parsing and understanding with Turing machine&lt;br /&gt;
*learned tensorflow&lt;br /&gt;
*produced two versions of results of query paraphrase&lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
*Code: &lt;br /&gt;
  1. sequence-to-sequence model with attention mechanism for chatting model and lyrics generation in theano and tensorflow&lt;br /&gt;
  2. Rare word embedding on word2vec and the related in C&lt;br /&gt;
  3. Help Lantian for ICASSP&lt;br /&gt;
*Paper sharing: two shared papers, one for rare word embedding, one for bilingual word embedding&lt;br /&gt;
*ML book: first version of chapter 2, linear model&lt;br /&gt;
*Read several papers about RNN, language model, word2vec, sequence-to-sequence model&lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
*finish part of chapter 5 kernel method&lt;br /&gt;
*retrain &amp;amp; test Chatting Model&lt;br /&gt;
*share 'the use of tensorflow'&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
*shared the principle of word2vec&lt;br /&gt;
*read several papers about comparison of sentence similarity &lt;br /&gt;
*wrote chapter 3 neural network of the ML book&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
*share two papers:&lt;br /&gt;
1.Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
&lt;br /&gt;
2.A Sentence Interaction Network for Modeling Dependence between Sentences&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/6c/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pptx slides]]&lt;br /&gt;
*write chapter2 linear model&lt;br /&gt;
*help Andy complete chapter3 neural network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
*prepare the data for question-pair task.&lt;br /&gt;
*use SRILM to train the language model and Moses to train the task&lt;br /&gt;
*share a paper about LSTM and do a presentation.&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/29 || || || 0.5 hours || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-30T06:59:29Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Month Summary */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 || 18:00   || 7+ || test the Chatting model,count the number of startswith &amp;lt;END&amp;gt; sentences,and natural sentences.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 || 18:20  || 8 || learn the CNN to solve MNIST using tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/30&lt;br /&gt;
|Ziwei Bai|| 9:15 ||   ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:15 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
*share two papers:&lt;br /&gt;
1.Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
&lt;br /&gt;
2.A Sentence Interaction Network for Modeling Dependence between Sentences&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/6c/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pptx slides]]&lt;br /&gt;
*write chapter2 linear model&lt;br /&gt;
*help Andy complete chapter3 neural network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
*prepare the data for question-pair task.&lt;br /&gt;
*use SRILM to train the language model and Moses to train the task&lt;br /&gt;
*share a paper about LSTM and do a presentation.&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/29 || || || 0.5 hours || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-30T06:58:42Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Month Summary */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 || 18:00   || 7+ || test the Chatting model,count the number of startswith &amp;lt;END&amp;gt; sentences,and natural sentences.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 || 18:20  || 8 || learn the CNN to solve MNIST using tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/30&lt;br /&gt;
|Ziwei Bai|| 9:15 ||   ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:15 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
*share two papers:&lt;br /&gt;
1.Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
&lt;br /&gt;
2.Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c5/Multilingual_Part-of-Speech_Tagging_with.pdf slides]]&lt;br /&gt;
*write chapter2 linear model&lt;br /&gt;
*help Andy complete chapter3 neural network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
*prepare the data for question-pair task.&lt;br /&gt;
*use SRILM to train the language model and Moses to train the task&lt;br /&gt;
*share a paper about LSTM and do a presentation.&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/29 || || || 0.5 hours || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-30T06:58:06Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Month Summary */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 || 18:00   || 7+ || test the Chatting model,count the number of startswith &amp;lt;END&amp;gt; sentences,and natural sentences.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 || 18:20  || 8 || learn the CNN to solve MNIST using tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/30&lt;br /&gt;
|Ziwei Bai|| 9:15 ||   ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:15 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
*share two papers:&lt;br /&gt;
   1.Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
   2.Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c5/Multilingual_Part-of-Speech_Tagging_with.pdf slides]]&lt;br /&gt;
*write chapter2 linear model&lt;br /&gt;
*help Andy complete chapter3 neural network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
*prepare the data for question-pair task.&lt;br /&gt;
*use SRILM to train the language model and Moses to train the task&lt;br /&gt;
*share a paper about LSTM and do a presentation.&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/29 || || || 0.5 hours || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-30T06:57:07Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Month Summary */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 || 18:00   || 7+ || test the Chatting model,count the number of startswith &amp;lt;END&amp;gt; sentences,and natural sentences.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 || 18:20  || 8 || learn the CNN to solve MNIST using tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/30&lt;br /&gt;
|Ziwei Bai|| 9:15 ||   ||  || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:15 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
*share two papers:&lt;br /&gt;
Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c5/Multilingual_Part-of-Speech_Tagging_with.pdf slides]]&lt;br /&gt;
*write chapter2 linear model&lt;br /&gt;
*help Andy complete chapter3 neural network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
*prepare the data for question-pair task.&lt;br /&gt;
*use SRILM to train the language model and Moses to train the task&lt;br /&gt;
*share a paper about LSTM and do a presentation.&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/29 || || || 0.5 hours || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/NLP_Status_Report_2016-08-29</id>
		<title>NLP Status Report 2016-08-29</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/NLP_Status_Report_2016-08-29"/>
				<updated>2016-08-29T08:54:15Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Date !! People !! Last Week !! This Week&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;6&amp;quot;|2016/08/22&lt;br /&gt;
|Yang Feng ||&lt;br /&gt;
*surveyed the line of neural parsing work&lt;br /&gt;
*arranged the nlp wiki pages&lt;br /&gt;
*arranged the report of tensorflow&lt;br /&gt;
||&lt;br /&gt;
*start the work of neural grammar&lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang ||&lt;br /&gt;
*added punctuation to input&lt;br /&gt;
*added input vector to the attention layer &lt;br /&gt;
*the result of the poem7_49k  [[http://192.168.0.51:5555/cgi-bin/cvss/cvss_request.pl?account=zhangjy&amp;amp;step=view_request&amp;amp;cvssid=559 here]]&lt;br /&gt;
 || &lt;br /&gt;
*writing books&lt;br /&gt;
*an overview of RNN,RBM,LSTM&lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
*Mainly focus on writing chapter 2, Linear Model. &lt;br /&gt;
*Now we have completed Introduction, Polynomial Regression, Linear Regression, Linear Classification, Probabilistic PCA, part of Probabilistic LDA. &lt;br /&gt;
|| &lt;br /&gt;
*Complete the remaining of chapter 2--PLDA and Bayesian Approach.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
*read books, papers, etc. about chapter3 neural networks; &lt;br /&gt;
*wrote the outine of this chapter &lt;br /&gt;
|| &lt;br /&gt;
*start writing my chapter&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;6&amp;quot;|2016/08/29&lt;br /&gt;
|Yang Feng ||&lt;br /&gt;
*continued surveying the line of neural parsing work&lt;br /&gt;
*read papers of variants of RBM and neural Turing machine&lt;br /&gt;
*learned tensorflow&lt;br /&gt;
||&lt;br /&gt;
*read more papers of Turing machine and get the full picture of my idea&lt;br /&gt;
*start the baseline work of neural Turing machine&lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang ||&lt;br /&gt;
*shared an overview of LSTM,RNN,RBM&lt;br /&gt;
*prepared relevant knowledge  for writing book&lt;br /&gt;
*ran models of 58k-hybird and 14k-hybird  [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Result_of_58k_and_14k.pdf result]]&lt;br /&gt;
 || &lt;br /&gt;
*the main focus on writing books&lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
*Finish the first version of chapter 2, Linear Model&lt;br /&gt;
*Help Lantian for ICASSP&lt;br /&gt;
|| &lt;br /&gt;
*Revise the Linear Model chapter (Waiting for teacher Wang's reply)&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
*almost finish the first version of chapter3&lt;br /&gt;
|| &lt;br /&gt;
*hope to finish the first version before '''''Sep 2nd''''', main challenge is to draw the pictures myself&lt;br /&gt;
*revise on the first version&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|| &lt;br /&gt;
*prepare diagrams in chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||&lt;br /&gt;
*Prepare the presentation of Linear Algebra and Probability Theory and Information Theory&lt;br /&gt;
*start to learn tensorflow&lt;br /&gt;
||&lt;br /&gt;
*learn some models by tensorflow&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/NLP_Status_Report_2016-08-29</id>
		<title>NLP Status Report 2016-08-29</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/NLP_Status_Report_2016-08-29"/>
				<updated>2016-08-29T08:53:48Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Date !! People !! Last Week !! This Week&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;6&amp;quot;|2016/08/22&lt;br /&gt;
|Yang Feng ||&lt;br /&gt;
*surveyed the line of neural parsing work&lt;br /&gt;
*arranged the nlp wiki pages&lt;br /&gt;
*arranged the report of tensorflow&lt;br /&gt;
||&lt;br /&gt;
*start the work of neural grammar&lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang ||&lt;br /&gt;
*added punctuation to input&lt;br /&gt;
*added input vector to the attention layer &lt;br /&gt;
*the result of the poem7_49k  [[http://192.168.0.51:5555/cgi-bin/cvss/cvss_request.pl?account=zhangjy&amp;amp;step=view_request&amp;amp;cvssid=559 here]]&lt;br /&gt;
 || &lt;br /&gt;
*writing books&lt;br /&gt;
*an overview of RNN,RBM,LSTM&lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
*Mainly focus on writing chapter 2, Linear Model. &lt;br /&gt;
*Now we have completed Introduction, Polynomial Regression, Linear Regression, Linear Classification, Probabilistic PCA, part of Probabilistic LDA. &lt;br /&gt;
|| &lt;br /&gt;
*Complete the remaining of chapter 2--PLDA and Bayesian Approach.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
*read books, papers, etc. about chapter3 neural networks; &lt;br /&gt;
*wrote the outine of this chapter &lt;br /&gt;
|| &lt;br /&gt;
*start writing my chapter&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;6&amp;quot;|2016/08/29&lt;br /&gt;
|Yang Feng ||&lt;br /&gt;
*continued surveying the line of neural parsing work&lt;br /&gt;
*read papers of variants of RBM and neural Turing machine&lt;br /&gt;
*learned tensorflow&lt;br /&gt;
||&lt;br /&gt;
*read more papers of Turing machine and get the full picture of my idea&lt;br /&gt;
*start the baseline work of neural Turing machine&lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang ||&lt;br /&gt;
*shared an overview of LSTM,RNN,RBM&lt;br /&gt;
*prepared relevant knowledge  for writing book&lt;br /&gt;
*ran models of 58k-hybird and 14k-hybird  [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Result_of_58k_and_14k.pdf result]]&lt;br /&gt;
 || &lt;br /&gt;
*the main focus on writing books&lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
*Finish the first version of chapter 2, Linear Model&lt;br /&gt;
*Help Lantian for ICASSP&lt;br /&gt;
|| &lt;br /&gt;
*Revise the Linear Model chapter (Waiting for teacher Wang's reply)&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
*almost finish the first version of chapter3&lt;br /&gt;
|| &lt;br /&gt;
*hope to finish the first version before '''''Sep 2nd''''', main challenge is to draw the pictures myself&lt;br /&gt;
*revise on the first version&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || || *prepare diagrams in chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||&lt;br /&gt;
*Prepare the presentation of Linear Algebra and Probability Theory and Information Theory&lt;br /&gt;
*start to learn tensorflow&lt;br /&gt;
||&lt;br /&gt;
*learn some models by tensorflow&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-29T08:51:16Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Work Progress */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  || 18:00  || 8 || finish the first edition of my work of ML book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/20&lt;br /&gt;
|Ziwei Bai|| 9:40  || 18:50  || 8 || perfect my work of the book&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/22&lt;br /&gt;
|Ziwei Bai|| 9:15  || 18:15 ||8 || test Chatting model ,retrain word vector&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:15  || 18:15  || 8 || start writing chapter 3, version 0.0.&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/23&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:15||8 || 1、retrain NRM model &amp;amp; word vec 2.review paper‘NRM for STC’&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:00 || 18:15 || 8 || continue writing the ML book. mainly worked on MLP today&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:00  ||  18:15 || 8 || read the book Deep Learning and prepare the presentation of Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/24&lt;br /&gt;
|Ziwei Bai|| 9:05 || 18:05 || 8 || test Chatting model 0-8,modify ASR Python Server Functional Specification&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:05|| 18:05 ||8 || mainly wrote the section of MLP &amp;amp; RBF&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:05  || 18:05 || 8 || finish the preparation of the presentation about Linear Algebra and Probability Theory&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/25&lt;br /&gt;
|Ziwei Bai|| 9:20  ||  21:00 || 7.5 || 1、test Chatting Model 9-14，and upload to cvss。2、learn  Bengio‘s dialogue system by hierarchical NN（except experiment and result）。&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 21:00  || 7.5 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:20  ||  21:00 || 7.5 || revise the PPT and start to learn tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/26&lt;br /&gt;
|Ziwei Bai|| 9:30  ||  18:30 || 8 || prepare for seminar,select 100  posts Suitable for dialogue from weibo data &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:30  || 18:30  || 8 || continue writing chapter3&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || 9:30  ||  18:30 || 8 || self-study tensorflow&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/27&lt;br /&gt;
|Ziwei Bai|| 9:30 || 20:30  || 10 || 1、filter the movie scipts data based on weibo data word vector (30W+pairs) and try to train Chatting Model with the dataset  on the basis of model 31, but load model failed(maybe because the version of tensor flow not consistent).2、find the response Corresponding to the post selected yesterday and test the model (0-34) with it&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/29&lt;br /&gt;
|Ziwei Bai|| 9:20 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:20  || 17:40  || 7+ || continued working on chapter 3, did some revise work &amp;amp; tried to find ways to draw figures&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:20 || 18:20  || 8 || prepare diagrams in the book&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li|| 9:20 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/15 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/16 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/17 || || || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || || 4 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 6 hours || || || || || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/22 || || || || || || 8 hours || 8 hours&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/23 || || 2 hours || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/24 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || || || || || || 8 hours ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/26 || || || 5 hours || || || 8 hours ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-19T09:12:40Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Daily Report */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   || 18:30 || 8  || continue preparing for the MLbook&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai|| 9:00  ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00  || 18:00  || 8 ||learn Hopfield network, prepare for the writing&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||  9:00 || 18:00  || 8 || learn Andrew Ng's online course&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || 2 hours || || || || || ||&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/19 || || 2 hours || || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-18T09:34:53Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Daily Report */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:30 ||18:30 || 8 ||learn the 'propert of kernel function'&amp;amp;learn chaos' chatting model predict program&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   || 18:30  || 8 || prepare for the presentation, learn RBF&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/19&lt;br /&gt;
|Ziwei Bai||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/* || || || || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Reading_table</id>
		<title>Reading table</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Reading_table"/>
				<updated>2016-08-18T05:33:42Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;1&amp;quot;|2014/10/22  ||Zhang Dong Xu|| Why RNN? [[媒体文件:Why_LSTM.pdf|PPT]] [[媒体文件:Learning_Long-Term_Dependencies_with_Gradient_Descent_is_Difficult.pdf|paper 1]],[[媒体文件:LongShortTermMemory.pdf|paper  2]]&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;3&amp;quot;| 2014/12/8 || rowspan='3'|Liu Rong || Yu Zhao, Zhiyuan Liu, Maosong Sun. Phrase Type Sensitive Tensor Indexing Model for Semantic Composition. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_tim.pdf pdf]&lt;br /&gt;
|-&lt;br /&gt;
| Yang Liu, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. Topical Word Embeddings. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_twe.pdf pdf][https://github.com/largelymfs/topical_word_embeddings code]&lt;br /&gt;
|-&lt;br /&gt;
|  Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, Xuan Zhu. Learning Entity and Relation Embeddings for Knowledge Graph Completion. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_transr.pdf pdf][https://github.com/mrlyk423/relation_extraction code]&lt;br /&gt;
&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/10 ||Liu Rong|| &lt;br /&gt;
*Context-Dependent Translation Selection Using Convolutional Neural Network [http://arxiv.org/abs/1503.02357]&lt;br /&gt;
*Syntax-based Deep Matching of Short Texts [http://arxiv.org/abs/1503.02427]&lt;br /&gt;
*Convolutional Neural Network Architectures for Matching Natural Language Sentences[http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf]&lt;br /&gt;
*LSTM: A Search Space Odyssey [http://arxiv.org/pdf/1503.04069.pdf]&lt;br /&gt;
*A Deep Embedding Model for Co-occurrence Learning  [http://arxiv.org/abs/1504.02824]&lt;br /&gt;
*Text segmentation based on semantic word embeddings[http://arxiv.org/abs/1503.05543]&lt;br /&gt;
*semantic parsing via paraphrashings[http://www.cs.tau.ac.il/research/jonathan.berant/homepage_files/publications/ACL14.pdf]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/22 ||Dong Wang|| &lt;br /&gt;
*From Word Embeddings To Document Distances [http://jmlr.org/proceedings/papers/v37/kusnerb15.pdf pdf]&lt;br /&gt;
*[[Asr-read-icml|Reading list for ICML2015]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/29 ||Xiaoxi Wang|| &lt;br /&gt;
* Sequence to Sequence Learning with Neural Networks [http://papers.nips.cc/paper/5346-information-based-learning-by-agents-in-unbounded-state-spaces pdf]&lt;br /&gt;
* Neural Machine Translation by Jointly Learning to Align and Translate [http://arxiv.org/abs/1409.0473 pdf]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Tianyi Luo|| &lt;br /&gt;
* A Hierarchical Knowledge Representation for Expert Finding on Social Media(ACL 2015 short paper) [[http://aclanthology.info/papers/a-hierarchical-knowledge-representation-for-expert-finding-on-social-media pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Dongxu Zhang||&lt;br /&gt;
* Describing Multimedia Content using Attention-based Encoder-Decoder Networks[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/e/e0/Describing_Multimedia_Content_using_Attention-based_Encoder-Decoder_Networks.pdf]&lt;br /&gt;
* Attention-Based Models for Speech Recognition[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/58/Attention-Based_Models_for_Speech_Recognition.pdf] details in speech recognition.&lt;br /&gt;
* Neural Machine Translation by Joint Learning to Align and Translate[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c3/Neural_Machine_Translation_by_Joint_Learning_to_Align_and_Translate.pdf] details in machine translation.&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/07 ||Chao Xing|| &lt;br /&gt;
* Neural Word Embedding as Implicit Matrix Factorization [[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf pdf]]&lt;br /&gt;
* Matrix factorization techniques for recommender systems [[http://www.columbia.edu/~jwp2128/Teaching/W4721/papers/ieeecomputer.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/14 ||Tianyi Luo, Dongxu Zhang, Chao Xing|| &lt;br /&gt;
* MEMORY NETWORKS(ICLR 2015) [[http://arxiv.org/pdf/1410.3916v10.pdf pdf]]&lt;br /&gt;
* End-To-End Memory Networks(NIPS 2015) [[http://arxiv.org/pdf/1503.08895v4.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/20 ||Tianyi Luo, Xiaoxi Wang|| &lt;br /&gt;
* The Kendall and Mallows Kernels for Permutations (ICML 2015) [[http://jmlr.csail.mit.edu/proceedings/papers/v37/jiao15.pdf pdf]]&lt;br /&gt;
* The ordering of expression among a few genes can provide simple cancer biomarkers and signal BRCA1 mutations (BMC Bioinformatics) [[http://www.biomedcentral.com/content/pdf/1471-2105-10-256.pdf pdf]]&lt;br /&gt;
* Reasoning about Entailment with Neural Attention [[http://arxiv.org/pdf/1509.06664v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/28 ||Lantian Li|| &lt;br /&gt;
* Binary Code Ranking with Weighted Hamming Distance [[http://www.cv-foundation.org/openaccess/content_cvpr_2013/papers/Zhang_Binary_Code_Ranking_2013_CVPR_paper.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/05 || Chao Xing, Xiaoxi Wang||&lt;br /&gt;
* Generative Image Modeling Using Spatial LSTMs [[http://arxiv.org/pdf/1506.03478v2.pdf pdf]]&lt;br /&gt;
* Character-level Convolutional Networks for Text Classification [[http://arxiv.org/pdf/1509.01626.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/20 || qixinWang||&lt;br /&gt;
* Are You Talking to a Machine? [[http://arxiv.org/pdf/1505.05612v3.pdf pdf]]&lt;br /&gt;
* m-RNN [[http://arxiv.org/pdf/1412.6632v5.pdf pdf]]&lt;br /&gt;
* PresentationPPT [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/93/PresentationPaper--QixinWang20151120.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Xiaoxi Wang||&lt;br /&gt;
* NEURAL PROGRAMMER-INTERPRETERS [[http://arxiv.org/pdf/1511.06279v2.pdf pdf]]&lt;br /&gt;
* Subset Selection by Pareto Optimization [[http://www.researchgate.net/profile/Yang_Yu87/publication/282632653_Subset_Selection_by_Pareto_Optimization/links/561495d908aed47facee68b5.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Chao Xing ||&lt;br /&gt;
*Random Walks and Neural Network Language Models [[http://www.aclweb.org/anthology/N15-1165 pdf]]&lt;br /&gt;
*SENSEMBED: Learning Sense Embeddings forWord and Relational Similarity[[http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2015_Iacobaccietal.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/4 || Dongxu Zhang, Qixin Wang, Chao Xing ||&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
*Playing Atari with Deep Reinforcement Learning[[http://arxiv.org/pdf/1312.5602v1.pdf pdf]]&lt;br /&gt;
*Word Embedding Revisited A New Representation Learning and Explicit Matrix Factorization Perspective [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/45/Report-1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/11 || Chao Xing, Yiqiao Pan ||&lt;br /&gt;
*Semi-Supervised Word Sense Disambiguation Using Word Embeddings in General and Specific Domains [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/57/Report-12-11-02.pdf pdf]]&lt;br /&gt;
*SENSE2VEC - A FAST AND ACCURATE METHOD FOR WORD SENSE DISAMBIGUATION IN NEURAL WORD EMBEDDINGS [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/f/f0/Report-12-11-03.pdf pdf]]&lt;br /&gt;
*Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space[[http://arxiv.org/pdf/1504.06654v1.pdf pdf]]&lt;br /&gt;
*Distributional Semantics in Use[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Report-12-11-01.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/18 || Tianyi Luo, Dongxu Zhang ||&lt;br /&gt;
*Human-level concept learning through probabilistic program induction(Cognitive Science) [[http://cdn1.almosthuman.cn/wp-content/uploads/2015/12/Human-level-concept-learning-through-probabilistic-program-induction.pdf pdf]]&lt;br /&gt;
*Cluster Analysis of Heterogeneous Rank Data(ICML 2007) [[http://machinelearning.wustl.edu/mlpapers/paper_files/icml2007_BusseOB07.pdf pdf]]&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/25 || Dongxu zhang, Qixin Wang ||&lt;br /&gt;
*Exploiting Multiple Sources for Open-domain Hypernym Discovery[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/d1/Exploiting_Multiple_Sources_for_Open-domain_Hypernym_Discovery.pdf]]&lt;br /&gt;
*learning semantic hierarchies via word embeddings[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/4f/Learning_semantic_hierarchies_via_word_embeddings_acl2014.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/31 || Xiaoxi Wang， Chao Xing||&lt;br /&gt;
* Multilingual Language Processing From Bytes [[http://arxiv.org/pdf/1512.00103v1.pdf pdf]]&lt;br /&gt;
* Towards universal neural nets: Gibbs machines and ACE. [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Report-12-31.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/8 || Qixin Wang， Tianyi Luo||&lt;br /&gt;
*Unveiling the Dreams of Word Embeddings: Towards Language-Driven Image Generation[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a1/Unveiling_the_Dreams_of_Word_Embeddings-_Towards_Language-Driven_Image_Generation.pdf pdf]]&lt;br /&gt;
*Generating Chinese Couplets using a Statistical MT Approach[[http://aclweb.org/anthology/C/C08/C08-1048.pdf pdf]]&lt;br /&gt;
*Generating Chinese Classical Poems with Statistical Machine Translation Models[[http://www.aaai.org/ocs/index.php/AAAI/AAAI12/paper/viewFile/4753/5314 pdf]]&lt;br /&gt;
*Chinese Poetry Generation with Recurrent Neural Networks[[http://www.aclweb.org/old_anthology/D/D14/D14-1074.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/15 || Chao Xing||&lt;br /&gt;
*Learning from Chris Dyer [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cd/Learning_From_Chris_Dyer.pptx ppt]]&lt;br /&gt;
*Learning Word Representations with Hierarchical Sparse Coding [[http://arxiv.org/pdf/1406.2035v2.pdf pdf]]&lt;br /&gt;
*Non-distributional Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-nondist.pdf pdf]]&lt;br /&gt;
*Sparse Overcomplete Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-overcomplete.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/22 || Qixin Wang, Tianyi Luo||&lt;br /&gt;
*Skip_thought_vector [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/aa/Skip_thought_vector.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/29 || Dongxu Zhang||&lt;br /&gt;
*Towards Neural Network-based Reasoning[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cb/Towards_Neural_Network-based_Reasoning.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/3/25 || Jiyuan Zhang||&lt;br /&gt;
*Modeling Temporal Dependencies in High-Dimensional Sequences:Application to Polyphonic Music Generation and Transcription[[http://www-etud.iro.umontreal.ca/~boulanni/ICML2012.pdf pdf]]&lt;br /&gt;
*Composing Music With Recurrent Neural Networks[[http://www.hexahedria.com/2015/08/03/composing-music-with-recurrent-neural-networks/   blog]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/1 || Chao Xing||&lt;br /&gt;
*Generating Text with Deep Reinforcement Learning[[http://arxiv.org/pdf/1510.09202v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/8 || Tianyi Luo||&lt;br /&gt;
*Generating Chinese Classical Poems with RNN[[http://nlp.hivefire.com/articles/share/56982/]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/28 || Chao Xing||&lt;br /&gt;
*Knowledge Base Completion via Search-Based Question Answering [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b1/Knowledge_Base_Completion_via_Search-Based_Question_Answering_-_Report.pdf]]&lt;br /&gt;
*Open Domain Question Answering via Semantic Enrichment [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/Open_Domain_Question_Answering_via_Semantic_Enrichment_-_Report.pdf]]&lt;br /&gt;
*A Neural Conversational Model [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/A_Neural_Conversational_Model_-_Report.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/5/11 || Chao Xing||&lt;br /&gt;
*A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion &lt;br /&gt;
*A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
*Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
*Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
*Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
|-&lt;br /&gt;
|2016/7/28 || Aiting Liu ||&lt;br /&gt;
*Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/4 || &lt;br /&gt;
* Aodong Li&lt;br /&gt;
* Jiyuan Zhang&lt;br /&gt;
* Andi Zhang &lt;br /&gt;
||&lt;br /&gt;
*On the Role of Seed Lexicons in Learning Bilingual Word Embeddings   [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/5b/On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings.pdf pdf]] [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/51/2.0On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings_.pdf slides]]&lt;br /&gt;
*ABCNN- Attention-Based Convolutional Neural Network for Modeling Sentence Pairs &lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/7/76/ABCNN-_Attention-Based_Convolutional_Neural_Network_for_Modeling_Sentence_Pairs_.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/0d/Attention-Based_Convolutianal_Neural_Network_for_Modeling_Sentence_Pairs.pptx slides]]&lt;br /&gt;
*[[Tutorial]]: Introduction to different LMs: NNLM, RNNLM, continuous bag of words, skip-gram&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Models_for_computing_continuous_vector_representations_of_words.pdf slides]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || &lt;br /&gt;
* Shiyao Li &lt;br /&gt;
* Aiting Liu &lt;br /&gt;
||&lt;br /&gt;
*Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
*A Sentence Interaction Network for Modeling Dependence between Sentences&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/6c/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pptx slides]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/25 || &lt;br /&gt;
* Ziwei Bai &lt;br /&gt;
* Jiyuan Zhang&lt;br /&gt;
* Shiyao Li &lt;br /&gt;
||&lt;br /&gt;
*[[Tutorial]]: Tensorflow guidelines and some examples&lt;br /&gt;
*[[Tutorial]]: Introduction to GRU, LSTM, RBM &lt;br /&gt;
* [[Tutorial]] : Linear Algebra, Probability and Information Basics&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pptx</id>
		<title>文件:A Sentence Interaction Network for Modeling Dependence between Sentences.pptx</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pptx"/>
				<updated>2016-08-18T05:31:59Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-18T01:21:57Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Work Progress */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9:30 || 17:30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9:00 || 18:00  || 8 || rewrite 'the property of Gram matrix' and write the second kernel construction method&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang|| 9:00 || 18:00  || 8 || read materials about Radio basis function networks as well as self-organizing map&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/18 &lt;br /&gt;
|Ziwei Bai|| 9:50 || ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||9:30   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/* || || || || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-17T08:52:06Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Work Progress */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9：30 || 17：30  || 7 ||learn kernel construction,write the first construction method.&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||9:30  || 17:30  || 7 || prepare for the ML book, watch Wang's video of neural network&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9：00 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||  ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:00 || 18:00  || 8 || view the information related to the Hopfield network and Boltzmann machine&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/* || || || || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-17T00:34:28Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Daily Report */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9：30 || 17：30  || 7 ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||  ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 || learn Wang's instructional video of Neural Network&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9：00 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||  ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/* || || || || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-17T00:32:47Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Daily Report */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=NLP Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
&lt;br /&gt;
* Yang Feng (冯洋)&lt;br /&gt;
* Jiyuan Zhang （张记袁）&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Andi Zhang (张安迪)&lt;br /&gt;
* Ziwei Bai （白子薇）&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Shiyao Li （李诗瑶）&lt;br /&gt;
&lt;br /&gt;
===Former Members===&lt;br /&gt;
* '''Chao Xing (邢超)'''     :  FreeNeb&lt;br /&gt;
* '''Rong Liu (刘荣)'''      :  优酷&lt;br /&gt;
* '''Xiaoxi Wang (王晓曦)''' :  图灵机器人&lt;br /&gt;
* '''Xi Ma (马习)'''         :  清华大学研究生&lt;br /&gt;
* '''Tianyi Luo (骆天一)'''  ： phd candidate in University of California Santa Cruz&lt;br /&gt;
* '''Qixin Wang (王琪鑫)'''  :  MA candidate in University of California&lt;br /&gt;
* '''DongXu Zhang (张东旭)''': --&lt;br /&gt;
* '''Yiqiao Pan (潘一桥)'''  ： MA candidate in University of Sydney &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Work Progress==&lt;br /&gt;
===Daily Report===&lt;br /&gt;
&lt;br /&gt;
{|class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Person  !! start!! leave !! hours ||status&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/16  &lt;br /&gt;
|Ziwei Bai|| 9：30 || 17：30  || 7 ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||  ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu|| 9:30  || 17:30  || 7 ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;4&amp;quot;|2016/08/17&lt;br /&gt;
|Ziwei Bai|| 9：00 ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang||  ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu||   ||   ||  ||&lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li ||   ||  ||   ||&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Month Summary===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!People!! Summary&lt;br /&gt;
|-&lt;br /&gt;
|Yang Feng || &lt;br /&gt;
|-&lt;br /&gt;
|Jiyuan Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aodong Li || &lt;br /&gt;
|-&lt;br /&gt;
|Ziwei Bai || &lt;br /&gt;
|-&lt;br /&gt;
|Andy Zhang || &lt;br /&gt;
|-&lt;br /&gt;
|Aiting Liu || &lt;br /&gt;
|-&lt;br /&gt;
|Shiyao Li || &lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
===Time Off Table===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Yang Feng !! Jiyuan Zhang !! Aodong Li !! Ziwei Bai !! Andy Zhang !! Aiitng Liu !! Shiyao Li&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/* || || || || || || ||&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
[[nlp-progress 2016/08]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/05/01 -- 08/16 | nlp-progress 2016/05-07]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress 2016/04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Nlp-progress_2016/05/01_--_08/16</id>
		<title>Nlp-progress 2016/05/01 -- 08/16</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Nlp-progress_2016/05/01_--_08/16"/>
				<updated>2016-08-16T07:08:51Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-15 ~ 2016-08-16 :  learn Neural Network &lt;br /&gt;
&lt;br /&gt;
2016-08-08 ~ 2016-08-12 :   &lt;br /&gt;
&lt;br /&gt;
    1.finish the first version of chapter2&lt;br /&gt;
    2.read paper &amp;quot;A Sentence Interaction Network for Modeling Dependence between Sentences&amp;quot;    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
2016-08-05:   write section probabilistic PCA&lt;br /&gt;
&lt;br /&gt;
2016-08-04:   write section softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-05:&lt;br /&gt;
            Give a report on my research on sentence similarity&lt;br /&gt;
&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-15T01:30:55Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-08 ~ 2016-08-12 :   &lt;br /&gt;
&lt;br /&gt;
    1.finish the first version of chapter2&lt;br /&gt;
    2.read paper &amp;quot;A Sentence Interaction Network for Modeling Dependence between Sentences&amp;quot;    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
2016-08-05:   write section probabilistic PCA&lt;br /&gt;
&lt;br /&gt;
2016-08-04:   write section softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-05:&lt;br /&gt;
            Give a report on my research on sentence similarity&lt;br /&gt;
&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-15T01:30:18Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-08 ~ 2016-08-12 :   &lt;br /&gt;
&lt;br /&gt;
finish the first version of chapter2, read paper &amp;quot;A Sentence Interaction Network for Modeling Dependence between Sentences&amp;quot;[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-08-05:   write section probabilistic PCA&lt;br /&gt;
&lt;br /&gt;
2016-08-04:   write section softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-05:&lt;br /&gt;
            Give a report on my research on sentence similarity&lt;br /&gt;
&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-15T01:29:54Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-08 ~ 2016-08-12 :   finish the first version of chapter2, read paper &amp;quot;A Sentence Interaction Network for Modeling Dependence between Sentences&amp;quot;[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-08-05:   write section probabilistic PCA&lt;br /&gt;
&lt;br /&gt;
2016-08-04:   write section softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-05:&lt;br /&gt;
            Give a report on my research on sentence similarity&lt;br /&gt;
&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-15T01:28:58Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-08 ~ 2016-08-12 :   finish the first version of chapter2, read paper [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-08-05:   write section probabilistic PCA&lt;br /&gt;
&lt;br /&gt;
2016-08-04:   write section softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-05:&lt;br /&gt;
            Give a report on my research on sentence similarity&lt;br /&gt;
&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-05T07:27:03Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-05:   write section probabilistic PCA&lt;br /&gt;
&lt;br /&gt;
2016-08-04:   write section softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-04T06:58:47Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression and softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Reading_table</id>
		<title>Reading table</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Reading_table"/>
				<updated>2016-08-04T04:56:02Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;1&amp;quot;|2014/10/22  ||Zhang Dong Xu|| Why RNN? [[媒体文件:Why_LSTM.pdf|PPT]] [[媒体文件:Learning_Long-Term_Dependencies_with_Gradient_Descent_is_Difficult.pdf|paper 1]],[[媒体文件:LongShortTermMemory.pdf|paper  2]]&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;3&amp;quot;| 2014/12/8 || rowspan='3'|Liu Rong || Yu Zhao, Zhiyuan Liu, Maosong Sun. Phrase Type Sensitive Tensor Indexing Model for Semantic Composition. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_tim.pdf pdf]&lt;br /&gt;
|-&lt;br /&gt;
| Yang Liu, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. Topical Word Embeddings. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_twe.pdf pdf][https://github.com/largelymfs/topical_word_embeddings code]&lt;br /&gt;
|-&lt;br /&gt;
|  Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, Xuan Zhu. Learning Entity and Relation Embeddings for Knowledge Graph Completion. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_transr.pdf pdf][https://github.com/mrlyk423/relation_extraction code]&lt;br /&gt;
&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/10 ||Liu Rong|| &lt;br /&gt;
*Context-Dependent Translation Selection Using Convolutional Neural Network [http://arxiv.org/abs/1503.02357]&lt;br /&gt;
*Syntax-based Deep Matching of Short Texts [http://arxiv.org/abs/1503.02427]&lt;br /&gt;
*Convolutional Neural Network Architectures for Matching Natural Language Sentences[http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf]&lt;br /&gt;
*LSTM: A Search Space Odyssey [http://arxiv.org/pdf/1503.04069.pdf]&lt;br /&gt;
*A Deep Embedding Model for Co-occurrence Learning  [http://arxiv.org/abs/1504.02824]&lt;br /&gt;
*Text segmentation based on semantic word embeddings[http://arxiv.org/abs/1503.05543]&lt;br /&gt;
*semantic parsing via paraphrashings[http://www.cs.tau.ac.il/research/jonathan.berant/homepage_files/publications/ACL14.pdf]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/22 ||Dong Wang|| &lt;br /&gt;
*From Word Embeddings To Document Distances [http://jmlr.org/proceedings/papers/v37/kusnerb15.pdf pdf]&lt;br /&gt;
*[[Asr-read-icml|Reading list for ICML2015]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/29 ||Xiaoxi Wang|| &lt;br /&gt;
* Sequence to Sequence Learning with Neural Networks [http://papers.nips.cc/paper/5346-information-based-learning-by-agents-in-unbounded-state-spaces pdf]&lt;br /&gt;
* Neural Machine Translation by Jointly Learning to Align and Translate [http://arxiv.org/abs/1409.0473 pdf]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Tianyi Luo|| &lt;br /&gt;
* A Hierarchical Knowledge Representation for Expert Finding on Social Media(ACL 2015 short paper) [[http://aclanthology.info/papers/a-hierarchical-knowledge-representation-for-expert-finding-on-social-media pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Dongxu Zhang||&lt;br /&gt;
* Describing Multimedia Content using Attention-based Encoder-Decoder Networks[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/e/e0/Describing_Multimedia_Content_using_Attention-based_Encoder-Decoder_Networks.pdf]&lt;br /&gt;
* Attention-Based Models for Speech Recognition[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/58/Attention-Based_Models_for_Speech_Recognition.pdf] details in speech recognition.&lt;br /&gt;
* Neural Machine Translation by Joint Learning to Align and Translate[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c3/Neural_Machine_Translation_by_Joint_Learning_to_Align_and_Translate.pdf] details in machine translation.&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/07 ||Chao Xing|| &lt;br /&gt;
* Neural Word Embedding as Implicit Matrix Factorization [[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf pdf]]&lt;br /&gt;
* Matrix factorization techniques for recommender systems [[http://www.columbia.edu/~jwp2128/Teaching/W4721/papers/ieeecomputer.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/14 ||Tianyi Luo, Dongxu Zhang, Chao Xing|| &lt;br /&gt;
* MEMORY NETWORKS(ICLR 2015) [[http://arxiv.org/pdf/1410.3916v10.pdf pdf]]&lt;br /&gt;
* End-To-End Memory Networks(NIPS 2015) [[http://arxiv.org/pdf/1503.08895v4.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/20 ||Tianyi Luo, Xiaoxi Wang|| &lt;br /&gt;
* The Kendall and Mallows Kernels for Permutations (ICML 2015) [[http://jmlr.csail.mit.edu/proceedings/papers/v37/jiao15.pdf pdf]]&lt;br /&gt;
* The ordering of expression among a few genes can provide simple cancer biomarkers and signal BRCA1 mutations (BMC Bioinformatics) [[http://www.biomedcentral.com/content/pdf/1471-2105-10-256.pdf pdf]]&lt;br /&gt;
* Reasoning about Entailment with Neural Attention [[http://arxiv.org/pdf/1509.06664v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/28 ||Lantian Li|| &lt;br /&gt;
* Binary Code Ranking with Weighted Hamming Distance [[http://www.cv-foundation.org/openaccess/content_cvpr_2013/papers/Zhang_Binary_Code_Ranking_2013_CVPR_paper.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/05 || Chao Xing, Xiaoxi Wang||&lt;br /&gt;
* Generative Image Modeling Using Spatial LSTMs [[http://arxiv.org/pdf/1506.03478v2.pdf pdf]]&lt;br /&gt;
* Character-level Convolutional Networks for Text Classification [[http://arxiv.org/pdf/1509.01626.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/20 || qixinWang||&lt;br /&gt;
* Are You Talking to a Machine? [[http://arxiv.org/pdf/1505.05612v3.pdf pdf]]&lt;br /&gt;
* m-RNN [[http://arxiv.org/pdf/1412.6632v5.pdf pdf]]&lt;br /&gt;
* PresentationPPT [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/93/PresentationPaper--QixinWang20151120.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Xiaoxi Wang||&lt;br /&gt;
* NEURAL PROGRAMMER-INTERPRETERS [[http://arxiv.org/pdf/1511.06279v2.pdf pdf]]&lt;br /&gt;
* Subset Selection by Pareto Optimization [[http://www.researchgate.net/profile/Yang_Yu87/publication/282632653_Subset_Selection_by_Pareto_Optimization/links/561495d908aed47facee68b5.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Chao Xing ||&lt;br /&gt;
*Random Walks and Neural Network Language Models [[http://www.aclweb.org/anthology/N15-1165 pdf]]&lt;br /&gt;
*SENSEMBED: Learning Sense Embeddings forWord and Relational Similarity[[http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2015_Iacobaccietal.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/4 || Dongxu Zhang, Qixin Wang, Chao Xing ||&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
*Playing Atari with Deep Reinforcement Learning[[http://arxiv.org/pdf/1312.5602v1.pdf pdf]]&lt;br /&gt;
*Word Embedding Revisited A New Representation Learning and Explicit Matrix Factorization Perspective [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/45/Report-1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/11 || Chao Xing, Yiqiao Pan ||&lt;br /&gt;
*Semi-Supervised Word Sense Disambiguation Using Word Embeddings in General and Specific Domains [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/57/Report-12-11-02.pdf pdf]]&lt;br /&gt;
*SENSE2VEC - A FAST AND ACCURATE METHOD FOR WORD SENSE DISAMBIGUATION IN NEURAL WORD EMBEDDINGS [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/f/f0/Report-12-11-03.pdf pdf]]&lt;br /&gt;
*Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space[[http://arxiv.org/pdf/1504.06654v1.pdf pdf]]&lt;br /&gt;
*Distributional Semantics in Use[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Report-12-11-01.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/18 || Tianyi Luo, Dongxu Zhang ||&lt;br /&gt;
*Human-level concept learning through probabilistic program induction(Cognitive Science) [[http://cdn1.almosthuman.cn/wp-content/uploads/2015/12/Human-level-concept-learning-through-probabilistic-program-induction.pdf pdf]]&lt;br /&gt;
*Cluster Analysis of Heterogeneous Rank Data(ICML 2007) [[http://machinelearning.wustl.edu/mlpapers/paper_files/icml2007_BusseOB07.pdf pdf]]&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/25 || Dongxu zhang, Qixin Wang ||&lt;br /&gt;
*Exploiting Multiple Sources for Open-domain Hypernym Discovery[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/d1/Exploiting_Multiple_Sources_for_Open-domain_Hypernym_Discovery.pdf]]&lt;br /&gt;
*learning semantic hierarchies via word embeddings[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/4f/Learning_semantic_hierarchies_via_word_embeddings_acl2014.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/31 || Xiaoxi Wang， Chao Xing||&lt;br /&gt;
* Multilingual Language Processing From Bytes [[http://arxiv.org/pdf/1512.00103v1.pdf pdf]]&lt;br /&gt;
* Towards universal neural nets: Gibbs machines and ACE. [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Report-12-31.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/8 || Qixin Wang， Tianyi Luo||&lt;br /&gt;
*Unveiling the Dreams of Word Embeddings: Towards Language-Driven Image Generation[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a1/Unveiling_the_Dreams_of_Word_Embeddings-_Towards_Language-Driven_Image_Generation.pdf pdf]]&lt;br /&gt;
*Generating Chinese Couplets using a Statistical MT Approach[[http://aclweb.org/anthology/C/C08/C08-1048.pdf pdf]]&lt;br /&gt;
*Generating Chinese Classical Poems with Statistical Machine Translation Models[[http://www.aaai.org/ocs/index.php/AAAI/AAAI12/paper/viewFile/4753/5314 pdf]]&lt;br /&gt;
*Chinese Poetry Generation with Recurrent Neural Networks[[http://www.aclweb.org/old_anthology/D/D14/D14-1074.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/15 || Chao Xing||&lt;br /&gt;
*Learning from Chris Dyer [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cd/Learning_From_Chris_Dyer.pptx ppt]]&lt;br /&gt;
*Learning Word Representations with Hierarchical Sparse Coding [[http://arxiv.org/pdf/1406.2035v2.pdf pdf]]&lt;br /&gt;
*Non-distributional Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-nondist.pdf pdf]]&lt;br /&gt;
*Sparse Overcomplete Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-overcomplete.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/22 || Qixin Wang, Tianyi Luo||&lt;br /&gt;
*Skip_thought_vector [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/aa/Skip_thought_vector.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/29 || Dongxu Zhang||&lt;br /&gt;
*Towards Neural Network-based Reasoning[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cb/Towards_Neural_Network-based_Reasoning.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/3/25 || Jiyuan Zhang||&lt;br /&gt;
*Modeling Temporal Dependencies in High-Dimensional Sequences:Application to Polyphonic Music Generation and Transcription[[http://www-etud.iro.umontreal.ca/~boulanni/ICML2012.pdf pdf]]&lt;br /&gt;
*Composing Music With Recurrent Neural Networks[[http://www.hexahedria.com/2015/08/03/composing-music-with-recurrent-neural-networks/   blog]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/1 || Chao Xing||&lt;br /&gt;
*Generating Text with Deep Reinforcement Learning[[http://arxiv.org/pdf/1510.09202v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/8 || Tianyi Luo||&lt;br /&gt;
*Generating Chinese Classical Poems with RNN[[http://nlp.hivefire.com/articles/share/56982/]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/28 || Chao Xing||&lt;br /&gt;
*Knowledge Base Completion via Search-Based Question Answering [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b1/Knowledge_Base_Completion_via_Search-Based_Question_Answering_-_Report.pdf]]&lt;br /&gt;
*Open Domain Question Answering via Semantic Enrichment [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/Open_Domain_Question_Answering_via_Semantic_Enrichment_-_Report.pdf]]&lt;br /&gt;
*A Neural Conversational Model [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/A_Neural_Conversational_Model_-_Report.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/5/11 || Chao Xing||&lt;br /&gt;
*A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion &lt;br /&gt;
*A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
*Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
*Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
*Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
|-&lt;br /&gt;
|2016/7/28 || Aiting Liu ||&lt;br /&gt;
*Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/17/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx slides]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/4 || Aodong Li, Jiyuan Zhang, Andi Zhang ||&lt;br /&gt;
*On the Role of Seed Lexicons in Learning Bilingual Word Embeddings   [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/5b/On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings.pdf pdf]]&lt;br /&gt;
*ABCNN- Attention-Based Convolutional Neural Network for Modeling Sentence Pairs &lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/7/76/ABCNN-_Attention-Based_Convolutional_Neural_Network_for_Modeling_Sentence_Pairs_.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Introduction to different LMs: NNLM, RNNLM, bag of words, skip-gram&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/11 || Shiyao Li, Ziwei Bai ||&lt;br /&gt;
*Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Tensorflow guidelines and some examples&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || Jiyuan Zhang ||&lt;br /&gt;
*[[Tutorial]]: Introduction to GRU, LSTM, RBM &lt;br /&gt;
* [[Tutorial]] : Linear Algebra, Probability and Information Basics&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-03T05:33:30Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-08-01T05:20:11Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Reading_table</id>
		<title>Reading table</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Reading_table"/>
				<updated>2016-08-01T01:24:39Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;1&amp;quot;|2014/10/22  ||Zhang Dong Xu|| Why RNN? [[媒体文件:Why_LSTM.pdf|PPT]] [[媒体文件:Learning_Long-Term_Dependencies_with_Gradient_Descent_is_Difficult.pdf|paper 1]],[[媒体文件:LongShortTermMemory.pdf|paper  2]]&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;3&amp;quot;| 2014/12/8 || rowspan='3'|Liu Rong || Yu Zhao, Zhiyuan Liu, Maosong Sun. Phrase Type Sensitive Tensor Indexing Model for Semantic Composition. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_tim.pdf pdf]&lt;br /&gt;
|-&lt;br /&gt;
| Yang Liu, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. Topical Word Embeddings. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_twe.pdf pdf][https://github.com/largelymfs/topical_word_embeddings code]&lt;br /&gt;
|-&lt;br /&gt;
|  Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, Xuan Zhu. Learning Entity and Relation Embeddings for Knowledge Graph Completion. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_transr.pdf pdf][https://github.com/mrlyk423/relation_extraction code]&lt;br /&gt;
&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/10 ||Liu Rong|| &lt;br /&gt;
*Context-Dependent Translation Selection Using Convolutional Neural Network [http://arxiv.org/abs/1503.02357]&lt;br /&gt;
*Syntax-based Deep Matching of Short Texts [http://arxiv.org/abs/1503.02427]&lt;br /&gt;
*Convolutional Neural Network Architectures for Matching Natural Language Sentences[http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf]&lt;br /&gt;
*LSTM: A Search Space Odyssey [http://arxiv.org/pdf/1503.04069.pdf]&lt;br /&gt;
*A Deep Embedding Model for Co-occurrence Learning  [http://arxiv.org/abs/1504.02824]&lt;br /&gt;
*Text segmentation based on semantic word embeddings[http://arxiv.org/abs/1503.05543]&lt;br /&gt;
*semantic parsing via paraphrashings[http://www.cs.tau.ac.il/research/jonathan.berant/homepage_files/publications/ACL14.pdf]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/22 ||Dong Wang|| &lt;br /&gt;
*From Word Embeddings To Document Distances [http://jmlr.org/proceedings/papers/v37/kusnerb15.pdf pdf]&lt;br /&gt;
*[[Asr-read-icml|Reading list for ICML2015]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/29 ||Xiaoxi Wang|| &lt;br /&gt;
* Sequence to Sequence Learning with Neural Networks [http://papers.nips.cc/paper/5346-information-based-learning-by-agents-in-unbounded-state-spaces pdf]&lt;br /&gt;
* Neural Machine Translation by Jointly Learning to Align and Translate [http://arxiv.org/abs/1409.0473 pdf]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Tianyi Luo|| &lt;br /&gt;
* A Hierarchical Knowledge Representation for Expert Finding on Social Media(ACL 2015 short paper) [[http://aclanthology.info/papers/a-hierarchical-knowledge-representation-for-expert-finding-on-social-media pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Dongxu Zhang||&lt;br /&gt;
* Describing Multimedia Content using Attention-based Encoder-Decoder Networks[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/e/e0/Describing_Multimedia_Content_using_Attention-based_Encoder-Decoder_Networks.pdf]&lt;br /&gt;
* Attention-Based Models for Speech Recognition[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/58/Attention-Based_Models_for_Speech_Recognition.pdf] details in speech recognition.&lt;br /&gt;
* Neural Machine Translation by Joint Learning to Align and Translate[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c3/Neural_Machine_Translation_by_Joint_Learning_to_Align_and_Translate.pdf] details in machine translation.&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/07 ||Chao Xing|| &lt;br /&gt;
* Neural Word Embedding as Implicit Matrix Factorization [[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf pdf]]&lt;br /&gt;
* Matrix factorization techniques for recommender systems [[http://www.columbia.edu/~jwp2128/Teaching/W4721/papers/ieeecomputer.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/14 ||Tianyi Luo, Dongxu Zhang, Chao Xing|| &lt;br /&gt;
* MEMORY NETWORKS(ICLR 2015) [[http://arxiv.org/pdf/1410.3916v10.pdf pdf]]&lt;br /&gt;
* End-To-End Memory Networks(NIPS 2015) [[http://arxiv.org/pdf/1503.08895v4.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/20 ||Tianyi Luo, Xiaoxi Wang|| &lt;br /&gt;
* The Kendall and Mallows Kernels for Permutations (ICML 2015) [[http://jmlr.csail.mit.edu/proceedings/papers/v37/jiao15.pdf pdf]]&lt;br /&gt;
* The ordering of expression among a few genes can provide simple cancer biomarkers and signal BRCA1 mutations (BMC Bioinformatics) [[http://www.biomedcentral.com/content/pdf/1471-2105-10-256.pdf pdf]]&lt;br /&gt;
* Reasoning about Entailment with Neural Attention [[http://arxiv.org/pdf/1509.06664v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/28 ||Lantian Li|| &lt;br /&gt;
* Binary Code Ranking with Weighted Hamming Distance [[http://www.cv-foundation.org/openaccess/content_cvpr_2013/papers/Zhang_Binary_Code_Ranking_2013_CVPR_paper.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/05 || Chao Xing, Xiaoxi Wang||&lt;br /&gt;
* Generative Image Modeling Using Spatial LSTMs [[http://arxiv.org/pdf/1506.03478v2.pdf pdf]]&lt;br /&gt;
* Character-level Convolutional Networks for Text Classification [[http://arxiv.org/pdf/1509.01626.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/20 || qixinWang||&lt;br /&gt;
* Are You Talking to a Machine? [[http://arxiv.org/pdf/1505.05612v3.pdf pdf]]&lt;br /&gt;
* m-RNN [[http://arxiv.org/pdf/1412.6632v5.pdf pdf]]&lt;br /&gt;
* PresentationPPT [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/93/PresentationPaper--QixinWang20151120.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Xiaoxi Wang||&lt;br /&gt;
* NEURAL PROGRAMMER-INTERPRETERS [[http://arxiv.org/pdf/1511.06279v2.pdf pdf]]&lt;br /&gt;
* Subset Selection by Pareto Optimization [[http://www.researchgate.net/profile/Yang_Yu87/publication/282632653_Subset_Selection_by_Pareto_Optimization/links/561495d908aed47facee68b5.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Chao Xing ||&lt;br /&gt;
*Random Walks and Neural Network Language Models [[http://www.aclweb.org/anthology/N15-1165 pdf]]&lt;br /&gt;
*SENSEMBED: Learning Sense Embeddings forWord and Relational Similarity[[http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2015_Iacobaccietal.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/4 || Dongxu Zhang, Qixin Wang, Chao Xing ||&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
*Playing Atari with Deep Reinforcement Learning[[http://arxiv.org/pdf/1312.5602v1.pdf pdf]]&lt;br /&gt;
*Word Embedding Revisited A New Representation Learning and Explicit Matrix Factorization Perspective [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/45/Report-1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/11 || Chao Xing, Yiqiao Pan ||&lt;br /&gt;
*Semi-Supervised Word Sense Disambiguation Using Word Embeddings in General and Specific Domains [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/57/Report-12-11-02.pdf pdf]]&lt;br /&gt;
*SENSE2VEC - A FAST AND ACCURATE METHOD FOR WORD SENSE DISAMBIGUATION IN NEURAL WORD EMBEDDINGS [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/f/f0/Report-12-11-03.pdf pdf]]&lt;br /&gt;
*Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space[[http://arxiv.org/pdf/1504.06654v1.pdf pdf]]&lt;br /&gt;
*Distributional Semantics in Use[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Report-12-11-01.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/18 || Tianyi Luo, Dongxu Zhang ||&lt;br /&gt;
*Human-level concept learning through probabilistic program induction(Cognitive Science) [[http://cdn1.almosthuman.cn/wp-content/uploads/2015/12/Human-level-concept-learning-through-probabilistic-program-induction.pdf pdf]]&lt;br /&gt;
*Cluster Analysis of Heterogeneous Rank Data(ICML 2007) [[http://machinelearning.wustl.edu/mlpapers/paper_files/icml2007_BusseOB07.pdf pdf]]&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/25 || Dongxu zhang, Qixin Wang ||&lt;br /&gt;
*Exploiting Multiple Sources for Open-domain Hypernym Discovery[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/d1/Exploiting_Multiple_Sources_for_Open-domain_Hypernym_Discovery.pdf]]&lt;br /&gt;
*learning semantic hierarchies via word embeddings[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/4f/Learning_semantic_hierarchies_via_word_embeddings_acl2014.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/31 || Xiaoxi Wang， Chao Xing||&lt;br /&gt;
* Multilingual Language Processing From Bytes [[http://arxiv.org/pdf/1512.00103v1.pdf pdf]]&lt;br /&gt;
* Towards universal neural nets: Gibbs machines and ACE. [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Report-12-31.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/8 || Qixin Wang， Tianyi Luo||&lt;br /&gt;
*Unveiling the Dreams of Word Embeddings: Towards Language-Driven Image Generation[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a1/Unveiling_the_Dreams_of_Word_Embeddings-_Towards_Language-Driven_Image_Generation.pdf pdf]]&lt;br /&gt;
*Generating Chinese Couplets using a Statistical MT Approach[[http://aclweb.org/anthology/C/C08/C08-1048.pdf pdf]]&lt;br /&gt;
*Generating Chinese Classical Poems with Statistical Machine Translation Models[[http://www.aaai.org/ocs/index.php/AAAI/AAAI12/paper/viewFile/4753/5314 pdf]]&lt;br /&gt;
*Chinese Poetry Generation with Recurrent Neural Networks[[http://www.aclweb.org/old_anthology/D/D14/D14-1074.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/15 || Chao Xing||&lt;br /&gt;
*Learning from Chris Dyer [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cd/Learning_From_Chris_Dyer.pptx ppt]]&lt;br /&gt;
*Learning Word Representations with Hierarchical Sparse Coding [[http://arxiv.org/pdf/1406.2035v2.pdf pdf]]&lt;br /&gt;
*Non-distributional Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-nondist.pdf pdf]]&lt;br /&gt;
*Sparse Overcomplete Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-overcomplete.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/22 || Qixin Wang, Tianyi Luo||&lt;br /&gt;
*Skip_thought_vector [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/aa/Skip_thought_vector.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/29 || Dongxu Zhang||&lt;br /&gt;
*Towards Neural Network-based Reasoning[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cb/Towards_Neural_Network-based_Reasoning.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/3/25 || Jiyuan Zhang||&lt;br /&gt;
*Modeling Temporal Dependencies in High-Dimensional Sequences:Application to Polyphonic Music Generation and Transcription[[http://www-etud.iro.umontreal.ca/~boulanni/ICML2012.pdf pdf]]&lt;br /&gt;
*Composing Music With Recurrent Neural Networks[[http://www.hexahedria.com/2015/08/03/composing-music-with-recurrent-neural-networks/   blog]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/1 || Chao Xing||&lt;br /&gt;
*Generating Text with Deep Reinforcement Learning[[http://arxiv.org/pdf/1510.09202v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/8 || Tianyi Luo||&lt;br /&gt;
*Generating Chinese Classical Poems with RNN[[http://nlp.hivefire.com/articles/share/56982/]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/28 || Chao Xing||&lt;br /&gt;
*Knowledge Base Completion via Search-Based Question Answering [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b1/Knowledge_Base_Completion_via_Search-Based_Question_Answering_-_Report.pdf]]&lt;br /&gt;
*Open Domain Question Answering via Semantic Enrichment [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/Open_Domain_Question_Answering_via_Semantic_Enrichment_-_Report.pdf]]&lt;br /&gt;
*A Neural Conversational Model [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/A_Neural_Conversational_Model_-_Report.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/5/11 || Chao Xing||&lt;br /&gt;
*A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion &lt;br /&gt;
*A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
*Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
*Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
*Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
|-&lt;br /&gt;
|2016/7/28 || Aiting Liu ||&lt;br /&gt;
*Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://Intrinsic Subspace Evaluation of Word Embedding Representations.pptx]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/4 || Aodong Li, Jiyuan Zhang, Andi Li ||&lt;br /&gt;
*On the Role of Seed Lexicons in Learning Bilingual Word Embeddings   [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/5b/On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings.pdf pdf]]&lt;br /&gt;
*ABCNN- Attention-Based Convolutional Neural Network for Modeling Sentence Pairs &lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/7/76/ABCNN-_Attention-Based_Convolutional_Neural_Network_for_Modeling_Sentence_Pairs_.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Introduction to different LMs: NNLM, RNNLM, bag of words, skip-gram&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/11 || Shiyao Li, Ziwei Bai ||&lt;br /&gt;
*Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Tensorflow guidelines and some examples&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || Jiyuan Zhang ||&lt;br /&gt;
*[[Tutorial]]: Introduction to GRU, LSTM, RBM &lt;br /&gt;
* [[Tutorial]] : Linear Algebra, Probability and Information Basics&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Reading_table</id>
		<title>Reading table</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Reading_table"/>
				<updated>2016-08-01T01:23:42Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;1&amp;quot;|2014/10/22  ||Zhang Dong Xu|| Why RNN? [[媒体文件:Why_LSTM.pdf|PPT]] [[媒体文件:Learning_Long-Term_Dependencies_with_Gradient_Descent_is_Difficult.pdf|paper 1]],[[媒体文件:LongShortTermMemory.pdf|paper  2]]&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;3&amp;quot;| 2014/12/8 || rowspan='3'|Liu Rong || Yu Zhao, Zhiyuan Liu, Maosong Sun. Phrase Type Sensitive Tensor Indexing Model for Semantic Composition. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_tim.pdf pdf]&lt;br /&gt;
|-&lt;br /&gt;
| Yang Liu, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. Topical Word Embeddings. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_twe.pdf pdf][https://github.com/largelymfs/topical_word_embeddings code]&lt;br /&gt;
|-&lt;br /&gt;
|  Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, Xuan Zhu. Learning Entity and Relation Embeddings for Knowledge Graph Completion. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_transr.pdf pdf][https://github.com/mrlyk423/relation_extraction code]&lt;br /&gt;
&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/10 ||Liu Rong|| &lt;br /&gt;
*Context-Dependent Translation Selection Using Convolutional Neural Network [http://arxiv.org/abs/1503.02357]&lt;br /&gt;
*Syntax-based Deep Matching of Short Texts [http://arxiv.org/abs/1503.02427]&lt;br /&gt;
*Convolutional Neural Network Architectures for Matching Natural Language Sentences[http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf]&lt;br /&gt;
*LSTM: A Search Space Odyssey [http://arxiv.org/pdf/1503.04069.pdf]&lt;br /&gt;
*A Deep Embedding Model for Co-occurrence Learning  [http://arxiv.org/abs/1504.02824]&lt;br /&gt;
*Text segmentation based on semantic word embeddings[http://arxiv.org/abs/1503.05543]&lt;br /&gt;
*semantic parsing via paraphrashings[http://www.cs.tau.ac.il/research/jonathan.berant/homepage_files/publications/ACL14.pdf]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/22 ||Dong Wang|| &lt;br /&gt;
*From Word Embeddings To Document Distances [http://jmlr.org/proceedings/papers/v37/kusnerb15.pdf pdf]&lt;br /&gt;
*[[Asr-read-icml|Reading list for ICML2015]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/29 ||Xiaoxi Wang|| &lt;br /&gt;
* Sequence to Sequence Learning with Neural Networks [http://papers.nips.cc/paper/5346-information-based-learning-by-agents-in-unbounded-state-spaces pdf]&lt;br /&gt;
* Neural Machine Translation by Jointly Learning to Align and Translate [http://arxiv.org/abs/1409.0473 pdf]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Tianyi Luo|| &lt;br /&gt;
* A Hierarchical Knowledge Representation for Expert Finding on Social Media(ACL 2015 short paper) [[http://aclanthology.info/papers/a-hierarchical-knowledge-representation-for-expert-finding-on-social-media pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Dongxu Zhang||&lt;br /&gt;
* Describing Multimedia Content using Attention-based Encoder-Decoder Networks[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/e/e0/Describing_Multimedia_Content_using_Attention-based_Encoder-Decoder_Networks.pdf]&lt;br /&gt;
* Attention-Based Models for Speech Recognition[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/58/Attention-Based_Models_for_Speech_Recognition.pdf] details in speech recognition.&lt;br /&gt;
* Neural Machine Translation by Joint Learning to Align and Translate[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c3/Neural_Machine_Translation_by_Joint_Learning_to_Align_and_Translate.pdf] details in machine translation.&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/07 ||Chao Xing|| &lt;br /&gt;
* Neural Word Embedding as Implicit Matrix Factorization [[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf pdf]]&lt;br /&gt;
* Matrix factorization techniques for recommender systems [[http://www.columbia.edu/~jwp2128/Teaching/W4721/papers/ieeecomputer.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/14 ||Tianyi Luo, Dongxu Zhang, Chao Xing|| &lt;br /&gt;
* MEMORY NETWORKS(ICLR 2015) [[http://arxiv.org/pdf/1410.3916v10.pdf pdf]]&lt;br /&gt;
* End-To-End Memory Networks(NIPS 2015) [[http://arxiv.org/pdf/1503.08895v4.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/20 ||Tianyi Luo, Xiaoxi Wang|| &lt;br /&gt;
* The Kendall and Mallows Kernels for Permutations (ICML 2015) [[http://jmlr.csail.mit.edu/proceedings/papers/v37/jiao15.pdf pdf]]&lt;br /&gt;
* The ordering of expression among a few genes can provide simple cancer biomarkers and signal BRCA1 mutations (BMC Bioinformatics) [[http://www.biomedcentral.com/content/pdf/1471-2105-10-256.pdf pdf]]&lt;br /&gt;
* Reasoning about Entailment with Neural Attention [[http://arxiv.org/pdf/1509.06664v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/28 ||Lantian Li|| &lt;br /&gt;
* Binary Code Ranking with Weighted Hamming Distance [[http://www.cv-foundation.org/openaccess/content_cvpr_2013/papers/Zhang_Binary_Code_Ranking_2013_CVPR_paper.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/05 || Chao Xing, Xiaoxi Wang||&lt;br /&gt;
* Generative Image Modeling Using Spatial LSTMs [[http://arxiv.org/pdf/1506.03478v2.pdf pdf]]&lt;br /&gt;
* Character-level Convolutional Networks for Text Classification [[http://arxiv.org/pdf/1509.01626.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/20 || qixinWang||&lt;br /&gt;
* Are You Talking to a Machine? [[http://arxiv.org/pdf/1505.05612v3.pdf pdf]]&lt;br /&gt;
* m-RNN [[http://arxiv.org/pdf/1412.6632v5.pdf pdf]]&lt;br /&gt;
* PresentationPPT [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/93/PresentationPaper--QixinWang20151120.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Xiaoxi Wang||&lt;br /&gt;
* NEURAL PROGRAMMER-INTERPRETERS [[http://arxiv.org/pdf/1511.06279v2.pdf pdf]]&lt;br /&gt;
* Subset Selection by Pareto Optimization [[http://www.researchgate.net/profile/Yang_Yu87/publication/282632653_Subset_Selection_by_Pareto_Optimization/links/561495d908aed47facee68b5.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Chao Xing ||&lt;br /&gt;
*Random Walks and Neural Network Language Models [[http://www.aclweb.org/anthology/N15-1165 pdf]]&lt;br /&gt;
*SENSEMBED: Learning Sense Embeddings forWord and Relational Similarity[[http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2015_Iacobaccietal.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/4 || Dongxu Zhang, Qixin Wang, Chao Xing ||&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
*Playing Atari with Deep Reinforcement Learning[[http://arxiv.org/pdf/1312.5602v1.pdf pdf]]&lt;br /&gt;
*Word Embedding Revisited A New Representation Learning and Explicit Matrix Factorization Perspective [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/45/Report-1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/11 || Chao Xing, Yiqiao Pan ||&lt;br /&gt;
*Semi-Supervised Word Sense Disambiguation Using Word Embeddings in General and Specific Domains [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/57/Report-12-11-02.pdf pdf]]&lt;br /&gt;
*SENSE2VEC - A FAST AND ACCURATE METHOD FOR WORD SENSE DISAMBIGUATION IN NEURAL WORD EMBEDDINGS [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/f/f0/Report-12-11-03.pdf pdf]]&lt;br /&gt;
*Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space[[http://arxiv.org/pdf/1504.06654v1.pdf pdf]]&lt;br /&gt;
*Distributional Semantics in Use[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Report-12-11-01.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/18 || Tianyi Luo, Dongxu Zhang ||&lt;br /&gt;
*Human-level concept learning through probabilistic program induction(Cognitive Science) [[http://cdn1.almosthuman.cn/wp-content/uploads/2015/12/Human-level-concept-learning-through-probabilistic-program-induction.pdf pdf]]&lt;br /&gt;
*Cluster Analysis of Heterogeneous Rank Data(ICML 2007) [[http://machinelearning.wustl.edu/mlpapers/paper_files/icml2007_BusseOB07.pdf pdf]]&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/25 || Dongxu zhang, Qixin Wang ||&lt;br /&gt;
*Exploiting Multiple Sources for Open-domain Hypernym Discovery[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/d1/Exploiting_Multiple_Sources_for_Open-domain_Hypernym_Discovery.pdf]]&lt;br /&gt;
*learning semantic hierarchies via word embeddings[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/4f/Learning_semantic_hierarchies_via_word_embeddings_acl2014.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/31 || Xiaoxi Wang， Chao Xing||&lt;br /&gt;
* Multilingual Language Processing From Bytes [[http://arxiv.org/pdf/1512.00103v1.pdf pdf]]&lt;br /&gt;
* Towards universal neural nets: Gibbs machines and ACE. [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Report-12-31.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/8 || Qixin Wang， Tianyi Luo||&lt;br /&gt;
*Unveiling the Dreams of Word Embeddings: Towards Language-Driven Image Generation[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a1/Unveiling_the_Dreams_of_Word_Embeddings-_Towards_Language-Driven_Image_Generation.pdf pdf]]&lt;br /&gt;
*Generating Chinese Couplets using a Statistical MT Approach[[http://aclweb.org/anthology/C/C08/C08-1048.pdf pdf]]&lt;br /&gt;
*Generating Chinese Classical Poems with Statistical Machine Translation Models[[http://www.aaai.org/ocs/index.php/AAAI/AAAI12/paper/viewFile/4753/5314 pdf]]&lt;br /&gt;
*Chinese Poetry Generation with Recurrent Neural Networks[[http://www.aclweb.org/old_anthology/D/D14/D14-1074.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/15 || Chao Xing||&lt;br /&gt;
*Learning from Chris Dyer [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cd/Learning_From_Chris_Dyer.pptx ppt]]&lt;br /&gt;
*Learning Word Representations with Hierarchical Sparse Coding [[http://arxiv.org/pdf/1406.2035v2.pdf pdf]]&lt;br /&gt;
*Non-distributional Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-nondist.pdf pdf]]&lt;br /&gt;
*Sparse Overcomplete Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-overcomplete.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/22 || Qixin Wang, Tianyi Luo||&lt;br /&gt;
*Skip_thought_vector [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/aa/Skip_thought_vector.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/29 || Dongxu Zhang||&lt;br /&gt;
*Towards Neural Network-based Reasoning[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cb/Towards_Neural_Network-based_Reasoning.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/3/25 || Jiyuan Zhang||&lt;br /&gt;
*Modeling Temporal Dependencies in High-Dimensional Sequences:Application to Polyphonic Music Generation and Transcription[[http://www-etud.iro.umontreal.ca/~boulanni/ICML2012.pdf pdf]]&lt;br /&gt;
*Composing Music With Recurrent Neural Networks[[http://www.hexahedria.com/2015/08/03/composing-music-with-recurrent-neural-networks/   blog]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/1 || Chao Xing||&lt;br /&gt;
*Generating Text with Deep Reinforcement Learning[[http://arxiv.org/pdf/1510.09202v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/8 || Tianyi Luo||&lt;br /&gt;
*Generating Chinese Classical Poems with RNN[[http://nlp.hivefire.com/articles/share/56982/]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/28 || Chao Xing||&lt;br /&gt;
*Knowledge Base Completion via Search-Based Question Answering [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b1/Knowledge_Base_Completion_via_Search-Based_Question_Answering_-_Report.pdf]]&lt;br /&gt;
*Open Domain Question Answering via Semantic Enrichment [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/Open_Domain_Question_Answering_via_Semantic_Enrichment_-_Report.pdf]]&lt;br /&gt;
*A Neural Conversational Model [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/A_Neural_Conversational_Model_-_Report.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/5/11 || Chao Xing||&lt;br /&gt;
*A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion &lt;br /&gt;
*A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
*Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
*Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
*Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
|-&lt;br /&gt;
|2016/7/28 || Aiting Liu ||&lt;br /&gt;
*Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[slides[http://Intrinsic Subspace Evaluation of Word Embedding Representations.pptx]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/4 || Aodong Li, Jiyuan Zhang, Andi Li ||&lt;br /&gt;
*On the Role of Seed Lexicons in Learning Bilingual Word Embeddings   [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/5b/On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings.pdf pdf]]&lt;br /&gt;
*ABCNN- Attention-Based Convolutional Neural Network for Modeling Sentence Pairs &lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/7/76/ABCNN-_Attention-Based_Convolutional_Neural_Network_for_Modeling_Sentence_Pairs_.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Introduction to different LMs: NNLM, RNNLM, bag of words, skip-gram&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/11 || Shiyao Li, Ziwei Bai ||&lt;br /&gt;
*Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Tensorflow guidelines and some examples&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || Jiyuan Zhang ||&lt;br /&gt;
*[[Tutorial]]: Introduction to GRU, LSTM, RBM &lt;br /&gt;
* [[Tutorial]] : Linear Algebra, Probability and Information Basics&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Reading_table</id>
		<title>Reading table</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Reading_table"/>
				<updated>2016-08-01T01:23:17Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;1&amp;quot;|2014/10/22  ||Zhang Dong Xu|| Why RNN? [[媒体文件:Why_LSTM.pdf|PPT]] [[媒体文件:Learning_Long-Term_Dependencies_with_Gradient_Descent_is_Difficult.pdf|paper 1]],[[媒体文件:LongShortTermMemory.pdf|paper  2]]&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;3&amp;quot;| 2014/12/8 || rowspan='3'|Liu Rong || Yu Zhao, Zhiyuan Liu, Maosong Sun. Phrase Type Sensitive Tensor Indexing Model for Semantic Composition. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_tim.pdf pdf]&lt;br /&gt;
|-&lt;br /&gt;
| Yang Liu, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. Topical Word Embeddings. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_twe.pdf pdf][https://github.com/largelymfs/topical_word_embeddings code]&lt;br /&gt;
|-&lt;br /&gt;
|  Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, Xuan Zhu. Learning Entity and Relation Embeddings for Knowledge Graph Completion. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_transr.pdf pdf][https://github.com/mrlyk423/relation_extraction code]&lt;br /&gt;
&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/10 ||Liu Rong|| &lt;br /&gt;
*Context-Dependent Translation Selection Using Convolutional Neural Network [http://arxiv.org/abs/1503.02357]&lt;br /&gt;
*Syntax-based Deep Matching of Short Texts [http://arxiv.org/abs/1503.02427]&lt;br /&gt;
*Convolutional Neural Network Architectures for Matching Natural Language Sentences[http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf]&lt;br /&gt;
*LSTM: A Search Space Odyssey [http://arxiv.org/pdf/1503.04069.pdf]&lt;br /&gt;
*A Deep Embedding Model for Co-occurrence Learning  [http://arxiv.org/abs/1504.02824]&lt;br /&gt;
*Text segmentation based on semantic word embeddings[http://arxiv.org/abs/1503.05543]&lt;br /&gt;
*semantic parsing via paraphrashings[http://www.cs.tau.ac.il/research/jonathan.berant/homepage_files/publications/ACL14.pdf]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/22 ||Dong Wang|| &lt;br /&gt;
*From Word Embeddings To Document Distances [http://jmlr.org/proceedings/papers/v37/kusnerb15.pdf pdf]&lt;br /&gt;
*[[Asr-read-icml|Reading list for ICML2015]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/29 ||Xiaoxi Wang|| &lt;br /&gt;
* Sequence to Sequence Learning with Neural Networks [http://papers.nips.cc/paper/5346-information-based-learning-by-agents-in-unbounded-state-spaces pdf]&lt;br /&gt;
* Neural Machine Translation by Jointly Learning to Align and Translate [http://arxiv.org/abs/1409.0473 pdf]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Tianyi Luo|| &lt;br /&gt;
* A Hierarchical Knowledge Representation for Expert Finding on Social Media(ACL 2015 short paper) [[http://aclanthology.info/papers/a-hierarchical-knowledge-representation-for-expert-finding-on-social-media pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Dongxu Zhang||&lt;br /&gt;
* Describing Multimedia Content using Attention-based Encoder-Decoder Networks[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/e/e0/Describing_Multimedia_Content_using_Attention-based_Encoder-Decoder_Networks.pdf]&lt;br /&gt;
* Attention-Based Models for Speech Recognition[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/58/Attention-Based_Models_for_Speech_Recognition.pdf] details in speech recognition.&lt;br /&gt;
* Neural Machine Translation by Joint Learning to Align and Translate[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c3/Neural_Machine_Translation_by_Joint_Learning_to_Align_and_Translate.pdf] details in machine translation.&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/07 ||Chao Xing|| &lt;br /&gt;
* Neural Word Embedding as Implicit Matrix Factorization [[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf pdf]]&lt;br /&gt;
* Matrix factorization techniques for recommender systems [[http://www.columbia.edu/~jwp2128/Teaching/W4721/papers/ieeecomputer.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/14 ||Tianyi Luo, Dongxu Zhang, Chao Xing|| &lt;br /&gt;
* MEMORY NETWORKS(ICLR 2015) [[http://arxiv.org/pdf/1410.3916v10.pdf pdf]]&lt;br /&gt;
* End-To-End Memory Networks(NIPS 2015) [[http://arxiv.org/pdf/1503.08895v4.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/20 ||Tianyi Luo, Xiaoxi Wang|| &lt;br /&gt;
* The Kendall and Mallows Kernels for Permutations (ICML 2015) [[http://jmlr.csail.mit.edu/proceedings/papers/v37/jiao15.pdf pdf]]&lt;br /&gt;
* The ordering of expression among a few genes can provide simple cancer biomarkers and signal BRCA1 mutations (BMC Bioinformatics) [[http://www.biomedcentral.com/content/pdf/1471-2105-10-256.pdf pdf]]&lt;br /&gt;
* Reasoning about Entailment with Neural Attention [[http://arxiv.org/pdf/1509.06664v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/28 ||Lantian Li|| &lt;br /&gt;
* Binary Code Ranking with Weighted Hamming Distance [[http://www.cv-foundation.org/openaccess/content_cvpr_2013/papers/Zhang_Binary_Code_Ranking_2013_CVPR_paper.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/05 || Chao Xing, Xiaoxi Wang||&lt;br /&gt;
* Generative Image Modeling Using Spatial LSTMs [[http://arxiv.org/pdf/1506.03478v2.pdf pdf]]&lt;br /&gt;
* Character-level Convolutional Networks for Text Classification [[http://arxiv.org/pdf/1509.01626.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/20 || qixinWang||&lt;br /&gt;
* Are You Talking to a Machine? [[http://arxiv.org/pdf/1505.05612v3.pdf pdf]]&lt;br /&gt;
* m-RNN [[http://arxiv.org/pdf/1412.6632v5.pdf pdf]]&lt;br /&gt;
* PresentationPPT [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/93/PresentationPaper--QixinWang20151120.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Xiaoxi Wang||&lt;br /&gt;
* NEURAL PROGRAMMER-INTERPRETERS [[http://arxiv.org/pdf/1511.06279v2.pdf pdf]]&lt;br /&gt;
* Subset Selection by Pareto Optimization [[http://www.researchgate.net/profile/Yang_Yu87/publication/282632653_Subset_Selection_by_Pareto_Optimization/links/561495d908aed47facee68b5.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Chao Xing ||&lt;br /&gt;
*Random Walks and Neural Network Language Models [[http://www.aclweb.org/anthology/N15-1165 pdf]]&lt;br /&gt;
*SENSEMBED: Learning Sense Embeddings forWord and Relational Similarity[[http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2015_Iacobaccietal.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/4 || Dongxu Zhang, Qixin Wang, Chao Xing ||&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
*Playing Atari with Deep Reinforcement Learning[[http://arxiv.org/pdf/1312.5602v1.pdf pdf]]&lt;br /&gt;
*Word Embedding Revisited A New Representation Learning and Explicit Matrix Factorization Perspective [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/45/Report-1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/11 || Chao Xing, Yiqiao Pan ||&lt;br /&gt;
*Semi-Supervised Word Sense Disambiguation Using Word Embeddings in General and Specific Domains [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/57/Report-12-11-02.pdf pdf]]&lt;br /&gt;
*SENSE2VEC - A FAST AND ACCURATE METHOD FOR WORD SENSE DISAMBIGUATION IN NEURAL WORD EMBEDDINGS [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/f/f0/Report-12-11-03.pdf pdf]]&lt;br /&gt;
*Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space[[http://arxiv.org/pdf/1504.06654v1.pdf pdf]]&lt;br /&gt;
*Distributional Semantics in Use[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Report-12-11-01.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/18 || Tianyi Luo, Dongxu Zhang ||&lt;br /&gt;
*Human-level concept learning through probabilistic program induction(Cognitive Science) [[http://cdn1.almosthuman.cn/wp-content/uploads/2015/12/Human-level-concept-learning-through-probabilistic-program-induction.pdf pdf]]&lt;br /&gt;
*Cluster Analysis of Heterogeneous Rank Data(ICML 2007) [[http://machinelearning.wustl.edu/mlpapers/paper_files/icml2007_BusseOB07.pdf pdf]]&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/25 || Dongxu zhang, Qixin Wang ||&lt;br /&gt;
*Exploiting Multiple Sources for Open-domain Hypernym Discovery[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/d1/Exploiting_Multiple_Sources_for_Open-domain_Hypernym_Discovery.pdf]]&lt;br /&gt;
*learning semantic hierarchies via word embeddings[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/4f/Learning_semantic_hierarchies_via_word_embeddings_acl2014.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/31 || Xiaoxi Wang， Chao Xing||&lt;br /&gt;
* Multilingual Language Processing From Bytes [[http://arxiv.org/pdf/1512.00103v1.pdf pdf]]&lt;br /&gt;
* Towards universal neural nets: Gibbs machines and ACE. [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Report-12-31.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/8 || Qixin Wang， Tianyi Luo||&lt;br /&gt;
*Unveiling the Dreams of Word Embeddings: Towards Language-Driven Image Generation[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a1/Unveiling_the_Dreams_of_Word_Embeddings-_Towards_Language-Driven_Image_Generation.pdf pdf]]&lt;br /&gt;
*Generating Chinese Couplets using a Statistical MT Approach[[http://aclweb.org/anthology/C/C08/C08-1048.pdf pdf]]&lt;br /&gt;
*Generating Chinese Classical Poems with Statistical Machine Translation Models[[http://www.aaai.org/ocs/index.php/AAAI/AAAI12/paper/viewFile/4753/5314 pdf]]&lt;br /&gt;
*Chinese Poetry Generation with Recurrent Neural Networks[[http://www.aclweb.org/old_anthology/D/D14/D14-1074.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/15 || Chao Xing||&lt;br /&gt;
*Learning from Chris Dyer [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cd/Learning_From_Chris_Dyer.pptx ppt]]&lt;br /&gt;
*Learning Word Representations with Hierarchical Sparse Coding [[http://arxiv.org/pdf/1406.2035v2.pdf pdf]]&lt;br /&gt;
*Non-distributional Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-nondist.pdf pdf]]&lt;br /&gt;
*Sparse Overcomplete Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-overcomplete.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/22 || Qixin Wang, Tianyi Luo||&lt;br /&gt;
*Skip_thought_vector [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/aa/Skip_thought_vector.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/29 || Dongxu Zhang||&lt;br /&gt;
*Towards Neural Network-based Reasoning[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cb/Towards_Neural_Network-based_Reasoning.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/3/25 || Jiyuan Zhang||&lt;br /&gt;
*Modeling Temporal Dependencies in High-Dimensional Sequences:Application to Polyphonic Music Generation and Transcription[[http://www-etud.iro.umontreal.ca/~boulanni/ICML2012.pdf pdf]]&lt;br /&gt;
*Composing Music With Recurrent Neural Networks[[http://www.hexahedria.com/2015/08/03/composing-music-with-recurrent-neural-networks/   blog]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/1 || Chao Xing||&lt;br /&gt;
*Generating Text with Deep Reinforcement Learning[[http://arxiv.org/pdf/1510.09202v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/8 || Tianyi Luo||&lt;br /&gt;
*Generating Chinese Classical Poems with RNN[[http://nlp.hivefire.com/articles/share/56982/]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/28 || Chao Xing||&lt;br /&gt;
*Knowledge Base Completion via Search-Based Question Answering [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b1/Knowledge_Base_Completion_via_Search-Based_Question_Answering_-_Report.pdf]]&lt;br /&gt;
*Open Domain Question Answering via Semantic Enrichment [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/Open_Domain_Question_Answering_via_Semantic_Enrichment_-_Report.pdf]]&lt;br /&gt;
*A Neural Conversational Model [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/A_Neural_Conversational_Model_-_Report.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/5/11 || Chao Xing||&lt;br /&gt;
*A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion &lt;br /&gt;
*A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
*Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
*Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
*Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
|-&lt;br /&gt;
|2016/7/28 || Aiting Liu ||&lt;br /&gt;
*Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://Intrinsic Subspace Evaluation of Word Embedding Representations.pptx pptx]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/4 || Aodong Li, Jiyuan Zhang, Andi Li ||&lt;br /&gt;
*On the Role of Seed Lexicons in Learning Bilingual Word Embeddings   [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/5b/On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings.pdf pdf]]&lt;br /&gt;
*ABCNN- Attention-Based Convolutional Neural Network for Modeling Sentence Pairs &lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/7/76/ABCNN-_Attention-Based_Convolutional_Neural_Network_for_Modeling_Sentence_Pairs_.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Introduction to different LMs: NNLM, RNNLM, bag of words, skip-gram&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/11 || Shiyao Li, Ziwei Bai ||&lt;br /&gt;
*Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Tensorflow guidelines and some examples&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || Jiyuan Zhang ||&lt;br /&gt;
*[[Tutorial]]: Introduction to GRU, LSTM, RBM &lt;br /&gt;
* [[Tutorial]] : Linear Algebra, Probability and Information Basics&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Reading_table</id>
		<title>Reading table</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Reading_table"/>
				<updated>2016-08-01T01:22:15Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;1&amp;quot;|2014/10/22  ||Zhang Dong Xu|| Why RNN? [[媒体文件:Why_LSTM.pdf|PPT]] [[媒体文件:Learning_Long-Term_Dependencies_with_Gradient_Descent_is_Difficult.pdf|paper 1]],[[媒体文件:LongShortTermMemory.pdf|paper  2]]&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;3&amp;quot;| 2014/12/8 || rowspan='3'|Liu Rong || Yu Zhao, Zhiyuan Liu, Maosong Sun. Phrase Type Sensitive Tensor Indexing Model for Semantic Composition. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_tim.pdf pdf]&lt;br /&gt;
|-&lt;br /&gt;
| Yang Liu, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. Topical Word Embeddings. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_twe.pdf pdf][https://github.com/largelymfs/topical_word_embeddings code]&lt;br /&gt;
|-&lt;br /&gt;
|  Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, Xuan Zhu. Learning Entity and Relation Embeddings for Knowledge Graph Completion. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_transr.pdf pdf][https://github.com/mrlyk423/relation_extraction code]&lt;br /&gt;
&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/10 ||Liu Rong|| &lt;br /&gt;
*Context-Dependent Translation Selection Using Convolutional Neural Network [http://arxiv.org/abs/1503.02357]&lt;br /&gt;
*Syntax-based Deep Matching of Short Texts [http://arxiv.org/abs/1503.02427]&lt;br /&gt;
*Convolutional Neural Network Architectures for Matching Natural Language Sentences[http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf]&lt;br /&gt;
*LSTM: A Search Space Odyssey [http://arxiv.org/pdf/1503.04069.pdf]&lt;br /&gt;
*A Deep Embedding Model for Co-occurrence Learning  [http://arxiv.org/abs/1504.02824]&lt;br /&gt;
*Text segmentation based on semantic word embeddings[http://arxiv.org/abs/1503.05543]&lt;br /&gt;
*semantic parsing via paraphrashings[http://www.cs.tau.ac.il/research/jonathan.berant/homepage_files/publications/ACL14.pdf]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/22 ||Dong Wang|| &lt;br /&gt;
*From Word Embeddings To Document Distances [http://jmlr.org/proceedings/papers/v37/kusnerb15.pdf pdf]&lt;br /&gt;
*[[Asr-read-icml|Reading list for ICML2015]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/29 ||Xiaoxi Wang|| &lt;br /&gt;
* Sequence to Sequence Learning with Neural Networks [http://papers.nips.cc/paper/5346-information-based-learning-by-agents-in-unbounded-state-spaces pdf]&lt;br /&gt;
* Neural Machine Translation by Jointly Learning to Align and Translate [http://arxiv.org/abs/1409.0473 pdf]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Tianyi Luo|| &lt;br /&gt;
* A Hierarchical Knowledge Representation for Expert Finding on Social Media(ACL 2015 short paper) [[http://aclanthology.info/papers/a-hierarchical-knowledge-representation-for-expert-finding-on-social-media pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Dongxu Zhang||&lt;br /&gt;
* Describing Multimedia Content using Attention-based Encoder-Decoder Networks[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/e/e0/Describing_Multimedia_Content_using_Attention-based_Encoder-Decoder_Networks.pdf]&lt;br /&gt;
* Attention-Based Models for Speech Recognition[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/58/Attention-Based_Models_for_Speech_Recognition.pdf] details in speech recognition.&lt;br /&gt;
* Neural Machine Translation by Joint Learning to Align and Translate[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c3/Neural_Machine_Translation_by_Joint_Learning_to_Align_and_Translate.pdf] details in machine translation.&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/07 ||Chao Xing|| &lt;br /&gt;
* Neural Word Embedding as Implicit Matrix Factorization [[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf pdf]]&lt;br /&gt;
* Matrix factorization techniques for recommender systems [[http://www.columbia.edu/~jwp2128/Teaching/W4721/papers/ieeecomputer.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/14 ||Tianyi Luo, Dongxu Zhang, Chao Xing|| &lt;br /&gt;
* MEMORY NETWORKS(ICLR 2015) [[http://arxiv.org/pdf/1410.3916v10.pdf pdf]]&lt;br /&gt;
* End-To-End Memory Networks(NIPS 2015) [[http://arxiv.org/pdf/1503.08895v4.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/20 ||Tianyi Luo, Xiaoxi Wang|| &lt;br /&gt;
* The Kendall and Mallows Kernels for Permutations (ICML 2015) [[http://jmlr.csail.mit.edu/proceedings/papers/v37/jiao15.pdf pdf]]&lt;br /&gt;
* The ordering of expression among a few genes can provide simple cancer biomarkers and signal BRCA1 mutations (BMC Bioinformatics) [[http://www.biomedcentral.com/content/pdf/1471-2105-10-256.pdf pdf]]&lt;br /&gt;
* Reasoning about Entailment with Neural Attention [[http://arxiv.org/pdf/1509.06664v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/28 ||Lantian Li|| &lt;br /&gt;
* Binary Code Ranking with Weighted Hamming Distance [[http://www.cv-foundation.org/openaccess/content_cvpr_2013/papers/Zhang_Binary_Code_Ranking_2013_CVPR_paper.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/05 || Chao Xing, Xiaoxi Wang||&lt;br /&gt;
* Generative Image Modeling Using Spatial LSTMs [[http://arxiv.org/pdf/1506.03478v2.pdf pdf]]&lt;br /&gt;
* Character-level Convolutional Networks for Text Classification [[http://arxiv.org/pdf/1509.01626.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/20 || qixinWang||&lt;br /&gt;
* Are You Talking to a Machine? [[http://arxiv.org/pdf/1505.05612v3.pdf pdf]]&lt;br /&gt;
* m-RNN [[http://arxiv.org/pdf/1412.6632v5.pdf pdf]]&lt;br /&gt;
* PresentationPPT [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/93/PresentationPaper--QixinWang20151120.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Xiaoxi Wang||&lt;br /&gt;
* NEURAL PROGRAMMER-INTERPRETERS [[http://arxiv.org/pdf/1511.06279v2.pdf pdf]]&lt;br /&gt;
* Subset Selection by Pareto Optimization [[http://www.researchgate.net/profile/Yang_Yu87/publication/282632653_Subset_Selection_by_Pareto_Optimization/links/561495d908aed47facee68b5.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Chao Xing ||&lt;br /&gt;
*Random Walks and Neural Network Language Models [[http://www.aclweb.org/anthology/N15-1165 pdf]]&lt;br /&gt;
*SENSEMBED: Learning Sense Embeddings forWord and Relational Similarity[[http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2015_Iacobaccietal.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/4 || Dongxu Zhang, Qixin Wang, Chao Xing ||&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
*Playing Atari with Deep Reinforcement Learning[[http://arxiv.org/pdf/1312.5602v1.pdf pdf]]&lt;br /&gt;
*Word Embedding Revisited A New Representation Learning and Explicit Matrix Factorization Perspective [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/45/Report-1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/11 || Chao Xing, Yiqiao Pan ||&lt;br /&gt;
*Semi-Supervised Word Sense Disambiguation Using Word Embeddings in General and Specific Domains [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/57/Report-12-11-02.pdf pdf]]&lt;br /&gt;
*SENSE2VEC - A FAST AND ACCURATE METHOD FOR WORD SENSE DISAMBIGUATION IN NEURAL WORD EMBEDDINGS [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/f/f0/Report-12-11-03.pdf pdf]]&lt;br /&gt;
*Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space[[http://arxiv.org/pdf/1504.06654v1.pdf pdf]]&lt;br /&gt;
*Distributional Semantics in Use[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Report-12-11-01.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/18 || Tianyi Luo, Dongxu Zhang ||&lt;br /&gt;
*Human-level concept learning through probabilistic program induction(Cognitive Science) [[http://cdn1.almosthuman.cn/wp-content/uploads/2015/12/Human-level-concept-learning-through-probabilistic-program-induction.pdf pdf]]&lt;br /&gt;
*Cluster Analysis of Heterogeneous Rank Data(ICML 2007) [[http://machinelearning.wustl.edu/mlpapers/paper_files/icml2007_BusseOB07.pdf pdf]]&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/25 || Dongxu zhang, Qixin Wang ||&lt;br /&gt;
*Exploiting Multiple Sources for Open-domain Hypernym Discovery[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/d1/Exploiting_Multiple_Sources_for_Open-domain_Hypernym_Discovery.pdf]]&lt;br /&gt;
*learning semantic hierarchies via word embeddings[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/4f/Learning_semantic_hierarchies_via_word_embeddings_acl2014.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/31 || Xiaoxi Wang， Chao Xing||&lt;br /&gt;
* Multilingual Language Processing From Bytes [[http://arxiv.org/pdf/1512.00103v1.pdf pdf]]&lt;br /&gt;
* Towards universal neural nets: Gibbs machines and ACE. [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Report-12-31.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/8 || Qixin Wang， Tianyi Luo||&lt;br /&gt;
*Unveiling the Dreams of Word Embeddings: Towards Language-Driven Image Generation[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a1/Unveiling_the_Dreams_of_Word_Embeddings-_Towards_Language-Driven_Image_Generation.pdf pdf]]&lt;br /&gt;
*Generating Chinese Couplets using a Statistical MT Approach[[http://aclweb.org/anthology/C/C08/C08-1048.pdf pdf]]&lt;br /&gt;
*Generating Chinese Classical Poems with Statistical Machine Translation Models[[http://www.aaai.org/ocs/index.php/AAAI/AAAI12/paper/viewFile/4753/5314 pdf]]&lt;br /&gt;
*Chinese Poetry Generation with Recurrent Neural Networks[[http://www.aclweb.org/old_anthology/D/D14/D14-1074.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/15 || Chao Xing||&lt;br /&gt;
*Learning from Chris Dyer [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cd/Learning_From_Chris_Dyer.pptx ppt]]&lt;br /&gt;
*Learning Word Representations with Hierarchical Sparse Coding [[http://arxiv.org/pdf/1406.2035v2.pdf pdf]]&lt;br /&gt;
*Non-distributional Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-nondist.pdf pdf]]&lt;br /&gt;
*Sparse Overcomplete Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-overcomplete.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/22 || Qixin Wang, Tianyi Luo||&lt;br /&gt;
*Skip_thought_vector [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/aa/Skip_thought_vector.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/29 || Dongxu Zhang||&lt;br /&gt;
*Towards Neural Network-based Reasoning[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cb/Towards_Neural_Network-based_Reasoning.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/3/25 || Jiyuan Zhang||&lt;br /&gt;
*Modeling Temporal Dependencies in High-Dimensional Sequences:Application to Polyphonic Music Generation and Transcription[[http://www-etud.iro.umontreal.ca/~boulanni/ICML2012.pdf pdf]]&lt;br /&gt;
*Composing Music With Recurrent Neural Networks[[http://www.hexahedria.com/2015/08/03/composing-music-with-recurrent-neural-networks/   blog]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/1 || Chao Xing||&lt;br /&gt;
*Generating Text with Deep Reinforcement Learning[[http://arxiv.org/pdf/1510.09202v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/8 || Tianyi Luo||&lt;br /&gt;
*Generating Chinese Classical Poems with RNN[[http://nlp.hivefire.com/articles/share/56982/]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/28 || Chao Xing||&lt;br /&gt;
*Knowledge Base Completion via Search-Based Question Answering [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b1/Knowledge_Base_Completion_via_Search-Based_Question_Answering_-_Report.pdf]]&lt;br /&gt;
*Open Domain Question Answering via Semantic Enrichment [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/Open_Domain_Question_Answering_via_Semantic_Enrichment_-_Report.pdf]]&lt;br /&gt;
*A Neural Conversational Model [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/A_Neural_Conversational_Model_-_Report.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/5/11 || Chao Xing||&lt;br /&gt;
*A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion &lt;br /&gt;
*A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
*Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
*Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
*Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
|-&lt;br /&gt;
|2016/7/28 || Aiting Liu ||&lt;br /&gt;
*Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[[http://Intrinsic Subspace Evaluation of Word Embedding Representations.pptx slides]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/4 || Aodong Li, Jiyuan Zhang, Andi Li ||&lt;br /&gt;
*On the Role of Seed Lexicons in Learning Bilingual Word Embeddings   [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/5b/On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings.pdf pdf]]&lt;br /&gt;
*ABCNN- Attention-Based Convolutional Neural Network for Modeling Sentence Pairs &lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/7/76/ABCNN-_Attention-Based_Convolutional_Neural_Network_for_Modeling_Sentence_Pairs_.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Introduction to different LMs: NNLM, RNNLM, bag of words, skip-gram&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/11 || Shiyao Li, Ziwei Bai ||&lt;br /&gt;
*Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Tensorflow guidelines and some examples&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || Jiyuan Zhang ||&lt;br /&gt;
*[[Tutorial]]: Introduction to GRU, LSTM, RBM &lt;br /&gt;
* [[Tutorial]] : Linear Algebra, Probability and Information Basics&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Reading_table</id>
		<title>Reading table</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Reading_table"/>
				<updated>2016-08-01T01:20:58Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;1&amp;quot;|2014/10/22  ||Zhang Dong Xu|| Why RNN? [[媒体文件:Why_LSTM.pdf|PPT]] [[媒体文件:Learning_Long-Term_Dependencies_with_Gradient_Descent_is_Difficult.pdf|paper 1]],[[媒体文件:LongShortTermMemory.pdf|paper  2]]&lt;br /&gt;
|-&lt;br /&gt;
| rowspan=&amp;quot;3&amp;quot;| 2014/12/8 || rowspan='3'|Liu Rong || Yu Zhao, Zhiyuan Liu, Maosong Sun. Phrase Type Sensitive Tensor Indexing Model for Semantic Composition. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_tim.pdf pdf]&lt;br /&gt;
|-&lt;br /&gt;
| Yang Liu, Zhiyuan Liu, Tat-Seng Chua, Maosong Sun. Topical Word Embeddings. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_twe.pdf pdf][https://github.com/largelymfs/topical_word_embeddings code]&lt;br /&gt;
|-&lt;br /&gt;
|  Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, Xuan Zhu. Learning Entity and Relation Embeddings for Knowledge Graph Completion. AAAI'15. [http://nlp.csai.tsinghua.edu.cn/~lzy/publications/aaai2015_transr.pdf pdf][https://github.com/mrlyk423/relation_extraction code]&lt;br /&gt;
&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/10 ||Liu Rong|| &lt;br /&gt;
*Context-Dependent Translation Selection Using Convolutional Neural Network [http://arxiv.org/abs/1503.02357]&lt;br /&gt;
*Syntax-based Deep Matching of Short Texts [http://arxiv.org/abs/1503.02427]&lt;br /&gt;
*Convolutional Neural Network Architectures for Matching Natural Language Sentences[http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf]&lt;br /&gt;
*LSTM: A Search Space Odyssey [http://arxiv.org/pdf/1503.04069.pdf]&lt;br /&gt;
*A Deep Embedding Model for Co-occurrence Learning  [http://arxiv.org/abs/1504.02824]&lt;br /&gt;
*Text segmentation based on semantic word embeddings[http://arxiv.org/abs/1503.05543]&lt;br /&gt;
*semantic parsing via paraphrashings[http://www.cs.tau.ac.il/research/jonathan.berant/homepage_files/publications/ACL14.pdf]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/22 ||Dong Wang|| &lt;br /&gt;
*From Word Embeddings To Document Distances [http://jmlr.org/proceedings/papers/v37/kusnerb15.pdf pdf]&lt;br /&gt;
*[[Asr-read-icml|Reading list for ICML2015]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/07/29 ||Xiaoxi Wang|| &lt;br /&gt;
* Sequence to Sequence Learning with Neural Networks [http://papers.nips.cc/paper/5346-information-based-learning-by-agents-in-unbounded-state-spaces pdf]&lt;br /&gt;
* Neural Machine Translation by Jointly Learning to Align and Translate [http://arxiv.org/abs/1409.0473 pdf]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Tianyi Luo|| &lt;br /&gt;
* A Hierarchical Knowledge Representation for Expert Finding on Social Media(ACL 2015 short paper) [[http://aclanthology.info/papers/a-hierarchical-knowledge-representation-for-expert-finding-on-social-media pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/05 ||Dongxu Zhang||&lt;br /&gt;
* Describing Multimedia Content using Attention-based Encoder-Decoder Networks[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/e/e0/Describing_Multimedia_Content_using_Attention-based_Encoder-Decoder_Networks.pdf]&lt;br /&gt;
* Attention-Based Models for Speech Recognition[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/58/Attention-Based_Models_for_Speech_Recognition.pdf] details in speech recognition.&lt;br /&gt;
* Neural Machine Translation by Joint Learning to Align and Translate[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/c3/Neural_Machine_Translation_by_Joint_Learning_to_Align_and_Translate.pdf] details in machine translation.&lt;br /&gt;
|-&lt;br /&gt;
|2015/08/07 ||Chao Xing|| &lt;br /&gt;
* Neural Word Embedding as Implicit Matrix Factorization [[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf pdf]]&lt;br /&gt;
* Matrix factorization techniques for recommender systems [[http://www.columbia.edu/~jwp2128/Teaching/W4721/papers/ieeecomputer.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/14 ||Tianyi Luo, Dongxu Zhang, Chao Xing|| &lt;br /&gt;
* MEMORY NETWORKS(ICLR 2015) [[http://arxiv.org/pdf/1410.3916v10.pdf pdf]]&lt;br /&gt;
* End-To-End Memory Networks(NIPS 2015) [[http://arxiv.org/pdf/1503.08895v4.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/20 ||Tianyi Luo, Xiaoxi Wang|| &lt;br /&gt;
* The Kendall and Mallows Kernels for Permutations (ICML 2015) [[http://jmlr.csail.mit.edu/proceedings/papers/v37/jiao15.pdf pdf]]&lt;br /&gt;
* The ordering of expression among a few genes can provide simple cancer biomarkers and signal BRCA1 mutations (BMC Bioinformatics) [[http://www.biomedcentral.com/content/pdf/1471-2105-10-256.pdf pdf]]&lt;br /&gt;
* Reasoning about Entailment with Neural Attention [[http://arxiv.org/pdf/1509.06664v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/10/28 ||Lantian Li|| &lt;br /&gt;
* Binary Code Ranking with Weighted Hamming Distance [[http://www.cv-foundation.org/openaccess/content_cvpr_2013/papers/Zhang_Binary_Code_Ranking_2013_CVPR_paper.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/05 || Chao Xing, Xiaoxi Wang||&lt;br /&gt;
* Generative Image Modeling Using Spatial LSTMs [[http://arxiv.org/pdf/1506.03478v2.pdf pdf]]&lt;br /&gt;
* Character-level Convolutional Networks for Text Classification [[http://arxiv.org/pdf/1509.01626.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/20 || qixinWang||&lt;br /&gt;
* Are You Talking to a Machine? [[http://arxiv.org/pdf/1505.05612v3.pdf pdf]]&lt;br /&gt;
* m-RNN [[http://arxiv.org/pdf/1412.6632v5.pdf pdf]]&lt;br /&gt;
* PresentationPPT [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/93/PresentationPaper--QixinWang20151120.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Xiaoxi Wang||&lt;br /&gt;
* NEURAL PROGRAMMER-INTERPRETERS [[http://arxiv.org/pdf/1511.06279v2.pdf pdf]]&lt;br /&gt;
* Subset Selection by Pareto Optimization [[http://www.researchgate.net/profile/Yang_Yu87/publication/282632653_Subset_Selection_by_Pareto_Optimization/links/561495d908aed47facee68b5.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/11/27 || Chao Xing ||&lt;br /&gt;
*Random Walks and Neural Network Language Models [[http://www.aclweb.org/anthology/N15-1165 pdf]]&lt;br /&gt;
*SENSEMBED: Learning Sense Embeddings forWord and Relational Similarity[[http://wwwusers.di.uniroma1.it/~navigli/pubs/ACL_2015_Iacobaccietal.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/4 || Dongxu Zhang, Qixin Wang, Chao Xing ||&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
*Playing Atari with Deep Reinforcement Learning[[http://arxiv.org/pdf/1312.5602v1.pdf pdf]]&lt;br /&gt;
*Word Embedding Revisited A New Representation Learning and Explicit Matrix Factorization Perspective [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/45/Report-1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/11 || Chao Xing, Yiqiao Pan ||&lt;br /&gt;
*Semi-Supervised Word Sense Disambiguation Using Word Embeddings in General and Specific Domains [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/57/Report-12-11-02.pdf pdf]]&lt;br /&gt;
*SENSE2VEC - A FAST AND ACCURATE METHOD FOR WORD SENSE DISAMBIGUATION IN NEURAL WORD EMBEDDINGS [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/f/f0/Report-12-11-03.pdf pdf]]&lt;br /&gt;
*Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space[[http://arxiv.org/pdf/1504.06654v1.pdf pdf]]&lt;br /&gt;
*Distributional Semantics in Use[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Report-12-11-01.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/18 || Tianyi Luo, Dongxu Zhang ||&lt;br /&gt;
*Human-level concept learning through probabilistic program induction(Cognitive Science) [[http://cdn1.almosthuman.cn/wp-content/uploads/2015/12/Human-level-concept-learning-through-probabilistic-program-induction.pdf pdf]]&lt;br /&gt;
*Cluster Analysis of Heterogeneous Rank Data(ICML 2007) [[http://machinelearning.wustl.edu/mlpapers/paper_files/icml2007_BusseOB07.pdf pdf]]&lt;br /&gt;
*Building a shared world: Mapping distributional to model-theoretic semantic spaces[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/da/Building_a_shared_world.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/25 || Dongxu zhang, Qixin Wang ||&lt;br /&gt;
*Exploiting Multiple Sources for Open-domain Hypernym Discovery[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/d1/Exploiting_Multiple_Sources_for_Open-domain_Hypernym_Discovery.pdf]]&lt;br /&gt;
*learning semantic hierarchies via word embeddings[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/4/4f/Learning_semantic_hierarchies_via_word_embeddings_acl2014.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2015/12/31 || Xiaoxi Wang， Chao Xing||&lt;br /&gt;
* Multilingual Language Processing From Bytes [[http://arxiv.org/pdf/1512.00103v1.pdf pdf]]&lt;br /&gt;
* Towards universal neural nets: Gibbs machines and ACE. [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/3b/Report-12-31.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/8 || Qixin Wang， Tianyi Luo||&lt;br /&gt;
*Unveiling the Dreams of Word Embeddings: Towards Language-Driven Image Generation[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a1/Unveiling_the_Dreams_of_Word_Embeddings-_Towards_Language-Driven_Image_Generation.pdf pdf]]&lt;br /&gt;
*Generating Chinese Couplets using a Statistical MT Approach[[http://aclweb.org/anthology/C/C08/C08-1048.pdf pdf]]&lt;br /&gt;
*Generating Chinese Classical Poems with Statistical Machine Translation Models[[http://www.aaai.org/ocs/index.php/AAAI/AAAI12/paper/viewFile/4753/5314 pdf]]&lt;br /&gt;
*Chinese Poetry Generation with Recurrent Neural Networks[[http://www.aclweb.org/old_anthology/D/D14/D14-1074.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/15 || Chao Xing||&lt;br /&gt;
*Learning from Chris Dyer [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cd/Learning_From_Chris_Dyer.pptx ppt]]&lt;br /&gt;
*Learning Word Representations with Hierarchical Sparse Coding [[http://arxiv.org/pdf/1406.2035v2.pdf pdf]]&lt;br /&gt;
*Non-distributional Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-nondist.pdf pdf]]&lt;br /&gt;
*Sparse Overcomplete Word Vector Representations [[http://www.cs.cmu.edu/~mfaruqui/papers/acl15-overcomplete.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/22 || Qixin Wang, Tianyi Luo||&lt;br /&gt;
*Skip_thought_vector [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/aa/Skip_thought_vector.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/1/29 || Dongxu Zhang||&lt;br /&gt;
*Towards Neural Network-based Reasoning[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/c/cb/Towards_Neural_Network-based_Reasoning.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/3/25 || Jiyuan Zhang||&lt;br /&gt;
*Modeling Temporal Dependencies in High-Dimensional Sequences:Application to Polyphonic Music Generation and Transcription[[http://www-etud.iro.umontreal.ca/~boulanni/ICML2012.pdf pdf]]&lt;br /&gt;
*Composing Music With Recurrent Neural Networks[[http://www.hexahedria.com/2015/08/03/composing-music-with-recurrent-neural-networks/   blog]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/1 || Chao Xing||&lt;br /&gt;
*Generating Text with Deep Reinforcement Learning[[http://arxiv.org/pdf/1510.09202v1.pdf pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/8 || Tianyi Luo||&lt;br /&gt;
*Generating Chinese Classical Poems with RNN[[http://nlp.hivefire.com/articles/share/56982/]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/4/28 || Chao Xing||&lt;br /&gt;
*Knowledge Base Completion via Search-Based Question Answering [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b1/Knowledge_Base_Completion_via_Search-Based_Question_Answering_-_Report.pdf]]&lt;br /&gt;
*Open Domain Question Answering via Semantic Enrichment [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/Open_Domain_Question_Answering_via_Semantic_Enrichment_-_Report.pdf]]&lt;br /&gt;
*A Neural Conversational Model [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/1/15/A_Neural_Conversational_Model_-_Report.pdf]]&lt;br /&gt;
|-&lt;br /&gt;
|2016/5/11 || Chao Xing||&lt;br /&gt;
*A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion &lt;br /&gt;
*A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
*Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
*Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
*Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
|-&lt;br /&gt;
|2016/7/28 || Aiting Liu ||&lt;br /&gt;
*Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
[http://Intrinsic Subspace Evaluation of Word Embedding Representations.pptx ]&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/4 || Aodong Li, Jiyuan Zhang, Andi Li ||&lt;br /&gt;
*On the Role of Seed Lexicons in Learning Bilingual Word Embeddings   [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/5/5b/On_the_Role_of_Seed_Lexicons_in_Learning_Bilingual_Word_Embeddings.pdf pdf]]&lt;br /&gt;
*ABCNN- Attention-Based Convolutional Neural Network for Modeling Sentence Pairs &lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/7/76/ABCNN-_Attention-Based_Convolutional_Neural_Network_for_Modeling_Sentence_Pairs_.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Introduction to different LMs: NNLM, RNNLM, bag of words, skip-gram&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/11 || Shiyao Li, Ziwei Bai ||&lt;br /&gt;
*Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss&lt;br /&gt;
[[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/a/a4/Multilingual_part-of-speech_tagging_with_bidirectional_long_short-term_memory_models_and_auxiliary_loss.pdf pdf]]&lt;br /&gt;
*[[Tutorial]]: Tensorflow guidelines and some examples&lt;br /&gt;
|-&lt;br /&gt;
|2016/8/18 || Jiyuan Zhang ||&lt;br /&gt;
*[[Tutorial]]: Introduction to GRU, LSTM, RBM &lt;br /&gt;
* [[Tutorial]] : Linear Algebra, Probability and Information Basics&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx</id>
		<title>文件:Intrinsic Subspace Evaluation of Word Embedding Representations.pptx</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/%E6%96%87%E4%BB%B6:Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pptx"/>
				<updated>2016-08-01T01:15:00Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-28T07:35:20Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-28:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-21~ 2016-07-22:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-26T07:11:59Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-26:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-21~ 2016-07-22:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-26T07:11:03Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-26:    read paper http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-21~ 2016-07-22:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/2016_Summer_Seminar_for_Machine_learning</id>
		<title>2016 Summer Seminar for Machine learning</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/2016_Summer_Seminar_for_Machine_learning"/>
				<updated>2016-07-22T06:36:42Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
*Location: FIT-1-304&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Date !! Speaker!! Title !!  Owner !! Materials  &lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/04  ||Dong Wang  || Machine learning overview ||  || [http://wangd.cslt.org/talks/seminar/2016-sum-ml/chpt1.%20Overview%20of%20Machine%20Learning.pptx slides][http://arch.cslt.org/video/2016/sum-ML/lesson1-2.mp4 video(part 2)][http://cs229.stanford.edu/section/cs229-linalg.pdf Algebra review] [http://cs229.stanford.edu/section/cs229-prob.pdf probability review] [http://cs229.stanford.edu/section/gaussians.pdf Gaussian distribution][http://cs229.stanford.edu/notes/cs229-notes4.pdf Learning theory]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/05  ||Dong Wang  || Linear models || Aiting Liu|| [http://wangd.cslt.org/talks/seminar/2016-sum-ml/chpt2.%20Linear%20Model.pptx slides][http://arch.cslt.org/video/2016/sum-ML/lesson2-1.mp4 video(part 1)] [http://arch.cslt.org/video/2016/sum-ML/lesson2-2.mp4 video(part 2)] [http://cs229.stanford.edu/notes/cs229-notes1.pdf NG's lecture 1] [http://cs229.stanford.edu/notes/cs229-notes2.pdf NG's lecture 2]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/08    ||Dong Wang  || Neural networks ||  || [http://wangd.cslt.org/talks/seminar/2016-sum-ml/chpt3.%20Neural%20Network.pptx slides] [http://arch.cslt.org/video/2016/sum-ML/lesson3-1.mp4 video(part 1)] [http://arch.cslt.org/video/2016/sum-ML/lesson3-2.mp4 video(part 2)]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/11    ||Dong Wang  || Deep learning (1)|| Zhiyuan Caixia Hang Luo|| [http://wangd.cslt.org/talks/seminar/2016-sum-ml/chpt4.%20Deep%20learning-1.pptx slides] [http://arch.cslt.org/video/2016/sum-ML/lesson4-1_Deep_learning.m4v video(part 1)] [http://arch.cslt.org/video/2016/sum-ML/lesson4-2_Deep_learning.m4v video(part 2)] [http://www.iro.umontreal.ca/~bengioy/talks/DL-Tutorial-NIPS2015.pdf NIPS 2015 tutorial]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/12   ||Dong Wang  || Deep learning (2)|| Zhiyuan Caixia  Hang Luo || [http://wangd.cslt.org/talks/seminar/2016-sum-ml/chpt5.%20Deep%20learning-2.pptx slides] [http://arch.cslt.org/video/2016/sum-ML/lesson5-1_Deep_learning.m4v video(part 1)] [http://arch.cslt.org/video/2016/sum-ML/lesson5-2_Deep_learning.m4v video(part 2)] [http://www.icassp2016.org/SP16_PlenaryDeng_Slides.pdf Li Deng's ICASSP16 keynote]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/13    ||Caixia Wang  || Kernel methods ||  || [http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/07/KM1.pdf slides]  [http://arch.cslt.org/video/2016/sum-ML/lesson6_Kernel_method.m4v video]  [http://cslt.riit.tsinghua.edu.cn/mediawiki/images/b/b5/Kernel_Methods_for_Pattern_Analysis.pdf Kernel method book] [http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/95/Pattern_Recognition_and_Machine_Learning.pdf pattern recognition 6-7]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/18    ||Yang Feng  || Graphical model (1) ||Jingyi Lin Ying Shi || [http://cslt.riit.tsinghua.edu.cn/mediawiki/images/d/dd/Chpt7._Graphical_models-bayesian_approach.pdf slides] [http://arch.cslt.org/video/2016/sum-ML/lesson7_Graphical_model.m4v video] &lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/21    ||Dong Wang  || Graphical model (2) ||Jingyi Lin Ying Shi||[http://wangd.cslt.org/talks/seminar/2016-sum-ml/chpt%208.%20Graphical%20models-2.pptx slides] [http://wangd.cslt.org/talks/seminar/2016-sum-ml/chapt8/Graphical%20models-learningInference.pptx Yang's slides] [http://wangd.cslt.org/talks/seminar/2016-sum-ml/chapt8/GraphicalModel_Jordan.pdf Jordan's lecture]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/25    ||Dong Wang  || Unsupervised learning || ||&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/26    ||Dong Wang  || Non parametric models || || [http://cs229.stanford.edu/section/cs229-gaussian_processes.pdf Gaussian process]&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/27    ||Dong Wang  || Reinforcement learning ||  ||&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/28    ||Maoning Wang  || Evolutionary learning || ||&lt;br /&gt;
|-&lt;br /&gt;
| 2016/07/29    ||Dong Wang  || Optimization ||  || [http://cs229.stanford.edu/section/cs229-cvxopt.pdf Convex optimization I] [http://cs229.stanford.edu/section/cs229-cvxopt.pdf Convex optimization II]&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-22T05:42:17Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-22:    run lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-18 ~ 2016-07-20:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-21T05:50:06Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-18 ~ 2016-07-20:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-21T05:49:38Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
                         preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-18 ~ 2016-07-20:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-21T05:49:06Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
    preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-18 ~ 2016-07-20:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-21T05:48:55Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-18 ~ 2016-07-20:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-21T05:47:47Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-18 ~ 2016-07-20:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php/Schedule</id>
		<title>Schedule</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php/Schedule"/>
				<updated>2016-07-21T05:45:48Z</updated>
		
		<summary type="html">&lt;p&gt;Liuaiting：/* Aiting Liu */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Text Processing Team Schedule=&lt;br /&gt;
&lt;br /&gt;
==Members==&lt;br /&gt;
===Former Members===&lt;br /&gt;
* Rong Liu (刘荣) : 优酷&lt;br /&gt;
* Xiaoxi Wang (王晓曦) : 图灵机器人&lt;br /&gt;
* Xi Ma (马习) : 清华大学研究生&lt;br /&gt;
* DongXu Zhang (张东旭) : --&lt;br /&gt;
* Yiqiao Pan (潘一桥)：继续读研&lt;br /&gt;
&lt;br /&gt;
===Current Members===&lt;br /&gt;
* Tianyi Luo (骆天一)&lt;br /&gt;
* Chao Xing (邢超)&lt;br /&gt;
* Qixin Wang (王琪鑫)&lt;br /&gt;
* Aodong Li (李傲冬)&lt;br /&gt;
* Aiting Liu (刘艾婷)&lt;br /&gt;
* Ziwei Bai (白子薇)&lt;br /&gt;
&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    preprocess the 20,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get twenty thousand songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-18 ~ 2016-07-20:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;br /&gt;
&lt;br /&gt;
==Past progress==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-05]]&lt;br /&gt;
&lt;br /&gt;
[[nlp-progress-2016-04]]&lt;/div&gt;</summary>
		<author><name>Liuaiting</name></author>	</entry>

	</feed>