<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://index.cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://index.cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=Nlp-progress_201605-20160816</id>
		<title>Nlp-progress 201605-20160816 - 版本历史</title>
		<link rel="self" type="application/atom+xml" href="http://index.cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=Nlp-progress_201605-20160816"/>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php?title=Nlp-progress_201605-20160816&amp;action=history"/>
		<updated>2026-04-09T09:15:33Z</updated>
		<subtitle>本wiki的该页面的版本历史</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php?title=Nlp-progress_201605-20160816&amp;diff=21721&amp;oldid=prev</id>
		<title>Fengyang：以“ ==Work Process== ===Paper Share=== ====2016-06-23==== Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D1...”为内容创建页面</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php?title=Nlp-progress_201605-20160816&amp;diff=21721&amp;oldid=prev"/>
				<updated>2016-08-16T06:33:21Z</updated>
		
		<summary type="html">&lt;p&gt;以“ ==Work Process== ===Paper Share=== ====2016-06-23==== Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D1...”为内容创建页面&lt;/p&gt;
&lt;p&gt;&lt;b&gt;新页面&lt;/b&gt;&lt;/p&gt;&lt;div&gt;&lt;br /&gt;
==Work Process==&lt;br /&gt;
===Paper Share===&lt;br /&gt;
====2016-06-23====&lt;br /&gt;
Learning Better Embeddings for Rare Words Using Distributional Representations [http://aclweb.org/anthology/D15-1033 pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Attention Networks for Document Classification [https://www.cs.cmu.edu/~diyiy/docs/naacl16.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Hierarchical Recurrent Neural Network for Document Modeling [http://www.aclweb.org/anthology/D/D15/D15-1106.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Learning Distributed Representations of Sentences from Unlabelled Data [http://arxiv.org/pdf/1602.03483.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
Speech Synthesis Based on HiddenMarkov Models [http://www.research.ed.ac.uk/portal/files/15269212/Speech_Synthesis_Based_on_Hidden_Markov_Models.pdf pdf]&lt;br /&gt;
&lt;br /&gt;
===Research Task===&lt;br /&gt;
====Binary Word Embedding(Aiting)====&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/9/97/Binary.pdf binary]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:  find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-01:  complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-28:  complete the word2vec model in tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
====Ordered Word Embedding(Aodong)====&lt;br /&gt;
&lt;br /&gt;
: 2016-07-25, 26, 27, 28, 29 : Share the ACL paper and implement rare word embedding&lt;br /&gt;
: 2016-07-18, 19, 20, 21, 22 : Find a scratch paper and ask for the corpora&lt;br /&gt;
: 2016-07-13, 14, 15 : Debug the mixed version of Rare Word Embedding&lt;br /&gt;
: 2016-07-12 : Complete the mixed initialization version of Rare Word Embedding and start training&lt;br /&gt;
: 2016-07-11 : &lt;br /&gt;
    Improve the predict process of chatting model&lt;br /&gt;
    Changing some hyperparameters of the chatting model to speed up the training process&lt;br /&gt;
: 2016-07-09, 10 : Try to carry out paper's low-freq word experiment, and do some readings both from PRML and Hang Li's paper.&lt;br /&gt;
: 2016-07-08 : Do model selections and the model finally set off on the server.&lt;br /&gt;
: 2016-07-07 : Complete the chatting model and run on Huilian's server, in order to overcome the GPU memory problem.&lt;br /&gt;
: 2016-07-06 : Finally complete translate model in tensorflow!!!! Now I am dealing with the out of memory problem. &lt;br /&gt;
: 2016-07-05 : &lt;br /&gt;
    Code predict process&lt;br /&gt;
    Although I've got a low cost value, the predict result does not compatible as expected, even the input of predict process from training set&lt;br /&gt;
    When I tried the Weibo data, program collapsed with an out of memory error.&lt;br /&gt;
: 2016-07-04 : Complete Coding training process&lt;br /&gt;
: 2016-07-01, 02: The cost function is very bumpy, debug it, while it's quite difficult!&lt;br /&gt;
: 2016-06-27, 28, 29 : Coding&lt;br /&gt;
: 2016-06-26 : Code tf's GRU and attention model&lt;br /&gt;
: 2016-06-25 : Read tf's source code rnn_cell.py and seq2seq.py&lt;br /&gt;
: 2016-06-24 : &lt;br /&gt;
    Code spearman correlation coefficient and experiment&lt;br /&gt;
    Read Li's paper &amp;quot;Neural Responding Machine for Short-Text Conversation&amp;quot;&lt;br /&gt;
: 2016-06-23 : &lt;br /&gt;
    Share paper &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
    experiment and receive new task&lt;br /&gt;
: 2016-06-22 : &lt;br /&gt;
    Experiment on low-frequency words&lt;br /&gt;
    Roughly read &amp;quot;Online Learning of Interpretable Word Embeddings&amp;quot;&lt;br /&gt;
    Roughly read &amp;quot;Learning Better Embeddings for Rare Words Using Distributional Representations&amp;quot;&lt;br /&gt;
: 2016-06-21 : Experiment and calculate cosine distance between words&lt;br /&gt;
: 2016-06-20 : Something went wrong with my program and fix it, so I have to start it all over again&lt;br /&gt;
: 2016-06-04 : Experiment the semantic&amp;amp;syntactic analysis of retrained word vector&lt;br /&gt;
: 2016-06-03 : Complete coding retrain process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-02 : Complete coding predict process of low-freq word and experiment the semantic&amp;amp;syntactic analysis&lt;br /&gt;
: 2016-06-01 : Read &amp;quot;Distributed Representations of Words and Phrases and their Compositionality&amp;quot;&lt;br /&gt;
: 2016-05-31 : &lt;br /&gt;
    Read Mikolov's ppt about his word embedding papers&lt;br /&gt;
    test the randomness of word2vec and there is nothing different in single thread while rerunning the program&lt;br /&gt;
    Download dataset &amp;quot;microsoft syntactic test set&amp;quot;, &amp;quot;wordsim353&amp;quot;, and &amp;quot;simlex-999&amp;quot;&lt;br /&gt;
: 2016-05-30 : Read &amp;quot;Hierarchical Probabilistic Neural Network Language Model&amp;quot; and &amp;quot;word2vec Explained: Deriving Mikolov's Negative-Sampling Word-Embedding Method&amp;quot;&lt;br /&gt;
: 2016-05-27 : Reread word2vec paper and read C-version word2vec. &lt;br /&gt;
: 2016-05-24 : Understand word2vec in TensorFlow, and because of some uncompleted functions, I determine to adapt the source of C-versioned word2vec. &lt;br /&gt;
: 2016-05-23 : &lt;br /&gt;
    Basic setup of TensorFlow&lt;br /&gt;
    Read code of word2vec in TensorFlow&lt;br /&gt;
: 2016-05-22 : &lt;br /&gt;
    Learn about algorithms in word2vec&lt;br /&gt;
    Read low-freq word papar and learn about 6 strategies&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/3/39/How_to_deal_with_low_frequency_words.pdf low_freq]&lt;br /&gt;
&lt;br /&gt;
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2c/Lowv.pdf order_rep]&lt;br /&gt;
&lt;br /&gt;
====Matrix Factorization(Ziwei)====&lt;br /&gt;
[http://papers.nips.cc/paper/5477-neural-word-embedding-as-implicit-matrix-factorization.pdf matrix-factorization]&lt;br /&gt;
&lt;br /&gt;
2016-06-23：&lt;br /&gt;
          prepare for report&lt;br /&gt;
2016-05-28：&lt;br /&gt;
          learn the code 'matrix-factorization.py','count_word_frequence.py',and 'reduce_rawtext_matrix_factorization.py'&lt;br /&gt;
          problem:I have no idea how to run the program and where the data.&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
           read the code 'map_rawtext_matrix_factorization.py'&lt;br /&gt;
2016-05-22：&lt;br /&gt;
           learn the rest of  paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the ‘abstract’ and ‘introduction’ of paper ‘Neural word Embedding as implicit matrix factorization’&lt;br /&gt;
&lt;br /&gt;
===Question answering system===&lt;br /&gt;
&lt;br /&gt;
====Chao Xing====&lt;br /&gt;
2016-05-30 ~ 2016-06-04 :&lt;br /&gt;
             Deliver CDSSM model to huilan.&lt;br /&gt;
2016-05-29 :&lt;br /&gt;
             Package chatting model in practice. &lt;br /&gt;
2016-05-28 :&lt;br /&gt;
             Modify bugs...&lt;br /&gt;
2016-05-27 :&lt;br /&gt;
             Train large scale model, find some problem.&lt;br /&gt;
2016-05-26 :&lt;br /&gt;
             Modify test program for large scale testing process.&lt;br /&gt;
2016-05-24 : &lt;br /&gt;
             Build CDSSM model in huilan's machine.&lt;br /&gt;
2016-05-23 : &lt;br /&gt;
             Find three things to do.&lt;br /&gt;
             1. Cost function change to maximize QA+ - QA-.&lt;br /&gt;
             2. Different parameters space in Q space and A space.&lt;br /&gt;
             3. HRNN separate to two tricky things : use output layer or use hidden layer as decoder's softmax layer's input.&lt;br /&gt;
2016-05-22 :&lt;br /&gt;
             1. Investigate different loss functions in chatting model.&lt;br /&gt;
2016-05-21 :&lt;br /&gt;
             1. Hand out different research task to intern students.&lt;br /&gt;
2016-05-20 : &lt;br /&gt;
             1. Testing denosing rnn generation model.&lt;br /&gt;
2016-05-19 : &lt;br /&gt;
             1. Discover for denosing rnn.&lt;br /&gt;
2016-05-18 :&lt;br /&gt;
             1. Modify model for crawler data.&lt;br /&gt;
2016-05-17 :&lt;br /&gt;
             1. Code &amp;amp; Test HRNN model.&lt;br /&gt;
2016-05-16 : &lt;br /&gt;
             1. Work done for CDSSM model.&lt;br /&gt;
2016-05-15 :&lt;br /&gt;
             1. Test CDSSM model package version.&lt;br /&gt;
2016-05-13 :&lt;br /&gt;
             1. Coding done CDSSM model package version. Wait to test.&lt;br /&gt;
2016-05-12 : &lt;br /&gt;
             1. Begin to package CDSSM model for huilan.&lt;br /&gt;
2016-05-11 : &lt;br /&gt;
             1. Prepare for paper sharing.&lt;br /&gt;
             2. Finish CDSSM model in chatting process.&lt;br /&gt;
             3. Start setup model &amp;amp; experiment in dialogue system.&lt;br /&gt;
2016-05-10 : &lt;br /&gt;
             1. Finish test CDSSM model in chatting, find original data has some problem.&lt;br /&gt;
             2. Read paper:&lt;br /&gt;
                    A Hierarchical Recurrent Encoder-Decoder for Generative Context-Aware Query Suggestion&lt;br /&gt;
                    A Neural Network Approach to Context-Sensitive Generation of Conversational Responses&lt;br /&gt;
                    Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models&lt;br /&gt;
                    Neural Responding Machine for Short-Text Conversation&lt;br /&gt;
2016-05-09 : &lt;br /&gt;
             1. Test CDSSM model in chatting model.&lt;br /&gt;
             2. Read paper : &lt;br /&gt;
                    Learning from Real Users Rating Dialogue Success with Neural Networks for Reinforcement Learning in Spoken Dialogue Systems&lt;br /&gt;
                    SimpleDS A Simple Deep Reinforcement Learning Dialogue System&lt;br /&gt;
             3. Code RNN by myself in tensorflow.&lt;br /&gt;
2016-05-08 : &lt;br /&gt;
             Fix some problem in dialogue system team, and continue read some papers in dialogue system.&lt;br /&gt;
2016-05-07 : &lt;br /&gt;
             Read some papers in dialogue system.&lt;br /&gt;
2016-05-06 : &lt;br /&gt;
             Try to fix RNN-DSSM model in tensorflow. Failure..&lt;br /&gt;
2016-05-05 : &lt;br /&gt;
             Coding for RNN-DSSM in tensorflow. Face an error when running rnn-dssm model in cpu : memory keep increasing. &lt;br /&gt;
             Tensorflow's version in huilan is 0.7.0 and install by pip, this cause using error in creating gpu graph,&lt;br /&gt;
             one possible solution is build tensorflow from source code.&lt;br /&gt;
&lt;br /&gt;
====Aiting Liu====&lt;br /&gt;
&lt;br /&gt;
2016-08-08 ~ 2016-08-12 :   &lt;br /&gt;
&lt;br /&gt;
    1.finish the first version of chapter2&lt;br /&gt;
    2.read paper &amp;quot;A Sentence Interaction Network for Modeling Dependence between Sentences&amp;quot;    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/0/04/A_Sentence_Interaction_Network_for_Modeling_Dependence_between_Sentences.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
2016-08-05:   write section probabilistic PCA&lt;br /&gt;
&lt;br /&gt;
2016-08-04:   write section softmax regression &lt;br /&gt;
&lt;br /&gt;
2016-08-03:   write section logistic regression&lt;br /&gt;
&lt;br /&gt;
2016-08-02:   write section polynomial fitting and linear regression&lt;br /&gt;
&lt;br /&gt;
2016-08-01:    learn linear model , determine the content of the chapter2&lt;br /&gt;
&lt;br /&gt;
2016-07-28 ~ 2016-07-29:    learn lesson linear model&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-27:    read paper Intrinsic Subspace Evaluation of Word Embedding Representations    [[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/6/68/Intrinsic_Subspace_Evaluation_of_Word_Embedding_Representations.pdf pdf]]&lt;br /&gt;
&lt;br /&gt;
2016-07-22 ~ 2016-07-23:    run and modify lyrics generation model&lt;br /&gt;
&lt;br /&gt;
2016-07-19 ~ 2016-07-21:    &lt;br /&gt;
&lt;br /&gt;
        preprocess the 200,000 lyrics &lt;br /&gt;
&lt;br /&gt;
2016-07-18:    get 200,000 songs from http://www.cnlyric.com/ ( singer list from a-z)&lt;br /&gt;
&lt;br /&gt;
2016-07-08:   preprocess the lyrics from baidu music&lt;br /&gt;
&lt;br /&gt;
2016-07-07:   get 27963 songs from http://www.cnlyric.com/ (singerlist from A/B/C)&lt;br /&gt;
&lt;br /&gt;
2016-07-06:   try to get lyrics from http://www.kuwo.cn/ http://www.kugou.com/ http://y.qq.com/#type=index http://music.163.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-05:   write lyrics spider, and get 56306 songs from http://music.baidu.com/&lt;br /&gt;
&lt;br /&gt;
2016-07-04:   learn tensorflow&lt;br /&gt;
&lt;br /&gt;
2016-07-01:   submit APSIPA2016  paper&lt;br /&gt;
&lt;br /&gt;
2016-06-30:   perfection paper&lt;br /&gt;
&lt;br /&gt;
2016-06-29:   complete the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-26:   modify the ordered word embedding's paper&lt;br /&gt;
&lt;br /&gt;
2016-06-25:   complete ordered word embedding experiment,get 54 figures&lt;br /&gt;
&lt;br /&gt;
2016-06-23:   read Bengio's paper https://arxiv.org/pdf/1605.06069v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-22:   read Bengio's paper http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
&lt;br /&gt;
2016-06-13:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Classification.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-12:&lt;br /&gt;
&lt;br /&gt;
    [[文件:Similarity.jpg]]&lt;br /&gt;
&lt;br /&gt;
2016-06-05:   complete the binary word embedding, find out that tensorflow does not provide logical derivation method.&lt;br /&gt;
&lt;br /&gt;
2016-06-04:   write the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-06-01:&lt;br /&gt;
&lt;br /&gt;
        1.Record demo video of our Personalized Chatterbot&lt;br /&gt;
        2.program the binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-31:  debugging our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-30:  complete our Personalized Chatterbot&lt;br /&gt;
&lt;br /&gt;
2016-05-29:&lt;br /&gt;
&lt;br /&gt;
        1.scan Chao's code and modify it&lt;br /&gt;
        2.run the modified program to get the eight hundred thousand sentences's whole matrix&lt;br /&gt;
&lt;br /&gt;
2016-05-28:&lt;br /&gt;
&lt;br /&gt;
        1.complete the word2vec model in tensorflow&lt;br /&gt;
        2.complete the first version of binary word embedding model&lt;br /&gt;
&lt;br /&gt;
2016-05-25:  .write my own version of word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-23:&lt;br /&gt;
&lt;br /&gt;
        1.get tensorflow's word2vec model from(https://github.com/tensorflow/tensorflow/tree/master/tensorflow/models/embedding)&lt;br /&gt;
        2.learn word2vec_basic model&lt;br /&gt;
        3.run word2vec.py and word2vec_optimized.py,we need a Chinese evaluation dataset if we want to use it directly&lt;br /&gt;
&lt;br /&gt;
2016-05-22：&lt;br /&gt;
&lt;br /&gt;
        1.find the tf.logical_xor(x,y) method in tensorflow to compute Hamming distance.&lt;br /&gt;
        2.learn tensorflow's word2vec model&lt;br /&gt;
&lt;br /&gt;
2016-05-21：&lt;br /&gt;
&lt;br /&gt;
        1.read Lantian's paper 'Binary Speaker Embedding'&lt;br /&gt;
        2.try to find a formula in tensorflow to compute Hamming distance.&lt;br /&gt;
&lt;br /&gt;
2016-05-18：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles and process them into a specific format(12.6M)&lt;br /&gt;
           (1.Sex and the City 2.Gossip Girl 3.Desperate Housewives 4.The IT Crowd 5.Empire 6.2 Broke Girls)&lt;br /&gt;
&lt;br /&gt;
2016-05-16：Process the data collected from the interview site,interview books and American TV subtitles(38.2M+23.2M)&lt;br /&gt;
&lt;br /&gt;
2016-05-11：&lt;br /&gt;
&lt;br /&gt;
            Fetch American TV subtitles&lt;br /&gt;
           (1.Friends 2.Big Bang Theory 3.The descendant of the Sun 4.Modern Family 5.House M.D. 6.Grey's Anatomy)&lt;br /&gt;
&lt;br /&gt;
2016-05-08：Fetch data from 'http://news.ifeng.com/' and 'http://www.xinhuanet.com/'(13.4M)&lt;br /&gt;
&lt;br /&gt;
2016-05-07：Fetch data from 'http://fangtan.china.com.cn/' and interview books (10M)&lt;br /&gt;
&lt;br /&gt;
2016-05-04：Establish the overall framework of our chat robot,and continue to build database&lt;br /&gt;
&lt;br /&gt;
====Ziwei Bai====&lt;br /&gt;
2016-07-29：&lt;br /&gt;
           download &amp;amp; learn latex&lt;br /&gt;
2016-07-25 ~2016-07-28：&lt;br /&gt;
           1、debug the based-RNN TTS（not ideal）&lt;br /&gt;
           2、run the based-RNN TTS&lt;br /&gt;
           3、write template&lt;br /&gt;
2016-07-21~ 2016-07-23:&lt;br /&gt;
           build RNN model for TTS&lt;br /&gt;
2016-07-18 ~ 2016-07-19:&lt;br /&gt;
           1、run bottleneck model with different parameters&lt;br /&gt;
           2、 prepare Bi-weekly report&lt;br /&gt;
           3、draw a map to compare different model,&lt;br /&gt;
2016-07-14 ~ 2016-07-15：&lt;br /&gt;
           build bottleneck model（non linear layer : sigmoid relu tanh）&lt;br /&gt;
2016-07-12 ~ 2016-07-13:&lt;br /&gt;
           modify the TTS program&lt;br /&gt;
           1、separate classify and transfer&lt;br /&gt;
           2、separate lf0 and mgc&lt;br /&gt;
2016-07-11：&lt;br /&gt;
           finish the patent&lt;br /&gt;
2016-07-07 ~ 2016-07-08:&lt;br /&gt;
           1、program LSTM with tensorflow （still has some bug）&lt;br /&gt;
           2、learn paper 'Fast,Compact,and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers fot Mobile Devices'&lt;br /&gt;
2016-07-06:&lt;br /&gt;
           finish the second edition of patent&lt;br /&gt;
2016-07-05：&lt;br /&gt;
           finish the fisrt edition of patent&lt;br /&gt;
2016-07-04：&lt;br /&gt;
           1、debug and run the chatting model with softmax &lt;br /&gt;
           2、determine model for patent of ‘LSTM-based modern text to the poetry conversion technology’&lt;br /&gt;
2016-07-01：&lt;br /&gt;
           the model updated yesterday can't converge，try to learn tf.sampled_softmax_loss()&lt;br /&gt;
2016-06-30:&lt;br /&gt;
           convert our chatting model from Negative sample to softmax and convert the cost from cosine to cross-entropy&lt;br /&gt;
           tf.softmax()&lt;br /&gt;
2016-06-29：&lt;br /&gt;
           learn paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
2016-06-23：&lt;br /&gt;
           learn paper ‘Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models’&lt;br /&gt;
           http://arxiv.org/pdf/1507.04808v3.pdf&lt;br /&gt;
2016-06-22：&lt;br /&gt;
          1、construct vector for word cut by jieba&lt;br /&gt;
          2、retrain the cdssm model with new word vector(still run)&lt;br /&gt;
2016-06-04：&lt;br /&gt;
          1、modify the interface for QA system&lt;br /&gt;
          2、pull together the interface and QA system&lt;br /&gt;
2016-06-01：&lt;br /&gt;
          1、add  data source and Performance Test results in work report&lt;br /&gt;
          2、learn pyQt&lt;br /&gt;
&lt;br /&gt;
2016-05-30：&lt;br /&gt;
            complete the work report&lt;br /&gt;
2016-05-29：&lt;br /&gt;
           write code for inputting a question ,return a answer sets whose question is most similar to the input question&lt;br /&gt;
2016-05-25:&lt;br /&gt;
           1、learn DSSM&lt;br /&gt;
           2、 complete the first edition of work report&lt;br /&gt;
           3、construct basic Q&amp;amp;A（name，age，job and so on）               &lt;br /&gt;
2016-05-23：&lt;br /&gt;
           write code for searching question in 'zhihu.sogou.com' and searching answer in zhihu&lt;br /&gt;
2016-05-21：&lt;br /&gt;
           learn the second half of paper 'A Neural Conversational Model'&lt;br /&gt;
2016-05-18:&lt;br /&gt;
           1、crawl QA pairs from http://www.chinalife.com.cn/publish/zhuzhan/index.html and http://www.pingan.com/&lt;br /&gt;
           2、find  paper 'A Neural Conversational Model' from google scholar and learn the first half of it.&lt;br /&gt;
2016-05-16:&lt;br /&gt;
            1、find datasets in paper 'Neural Responding Machine for Short-Text Conversation'&lt;br /&gt;
            2、reconstruct 15 scripts into our expected formula &lt;br /&gt;
2016-05-15:&lt;br /&gt;
            1、find 130 scripts&lt;br /&gt;
            2、 reconstruct 11 scripts into our expected formula &lt;br /&gt;
            problem：many files cann't distinguish between dialogue and scenario describes by program. &lt;br /&gt;
&lt;br /&gt;
2016-05-11:&lt;br /&gt;
             1、read paper“Movie-DiC: a Movie Dialogue Corpus for Research and Development”&lt;br /&gt;
             2、reconstruct a new film scripts into our expected formula &lt;br /&gt;
&lt;br /&gt;
2016-05-08:   convert the pdf we found yesterday into txt，and reconstruct the data into our expected formula   &lt;br /&gt;
&lt;br /&gt;
2016-05-07:   Finding 9 Drama scripts and 20 film scripts  &lt;br /&gt;
&lt;br /&gt;
2016-05-04：Finding and dealing with the data for QA system&lt;br /&gt;
&lt;br /&gt;
====Andi Zhang====&lt;br /&gt;
2016-08-05:&lt;br /&gt;
            Give a report on my research on sentence similarity&lt;br /&gt;
&lt;br /&gt;
2016-08-04:&lt;br /&gt;
            Give a representation on NNLMs; review papers read earlier this week.&lt;br /&gt;
&lt;br /&gt;
2016-08-03:&lt;br /&gt;
            Read the paper ''Multi-Perspective Sentence Similarity Modeling with Convolutional Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-02:&lt;br /&gt;
            Read the paper ''Modeling Interestingness with Deep Neural Networks''&lt;br /&gt;
&lt;br /&gt;
2016-08-01:&lt;br /&gt;
            Read papers about ABCNN for modeling sentence pairs&lt;br /&gt;
&lt;br /&gt;
2016-07-25 ~ 2016-07-29:&lt;br /&gt;
            Read papers about the theories and realization of NNLM, RNNLM &amp;amp; word2vec, prepared for a representation of this topic&lt;br /&gt;
&lt;br /&gt;
2016-07-22:&lt;br /&gt;
            Read papers about CBOW &amp;amp; Skip-gram&lt;br /&gt;
&lt;br /&gt;
===Generation Model (Aodong li)===&lt;br /&gt;
&lt;br /&gt;
: 2016-05-21 : Complete my biweekly report and take over new tasks -- low-frequency words&lt;br /&gt;
: 2016-05-20 : &lt;br /&gt;
    Optimize my code to speed up&lt;br /&gt;
    Train the models with GPU&lt;br /&gt;
    However, it does not converge :(&lt;br /&gt;
: 2016-05-19 : Code a simple version of keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-18 : Debug keywords-to-sequence model and train the model&lt;br /&gt;
: 2016-05-17 : make technical details clear and code keywords-to-sequence model&lt;br /&gt;
: 2016-05-16 : Denoise and segment more lyrics and prepare for keywords to sequence model&lt;br /&gt;
: 2016-05-15 : Train some different models and analyze performance: song to song, paragraph to paragraph, etc.&lt;br /&gt;
: 2016-05-12 : complete sequence to sequence model's prediction process and the whole standard sequence to sequence lstm-based model v0.0&lt;br /&gt;
: 2016-05-11 : complete sequence to sequence model's training process in Theano&lt;br /&gt;
: 2016-05-10 : complete sequence to sequence lstm-based model in Theano&lt;br /&gt;
: 2016-05-09 : try to code sequence to sequence model &lt;br /&gt;
: 2016-05-08 : &lt;br /&gt;
    denoise and train word vectors of  Lijun Deng's lyrics (110+ pieces)&lt;br /&gt;
    decide on using raw sequence to sequence model&lt;br /&gt;
: 2016-05-07 : &lt;br /&gt;
    study attention-based model&lt;br /&gt;
    learn some details about the poem generation model&lt;br /&gt;
    change my focus onto lyrics generation model&lt;br /&gt;
: 2016-05-06 : read the paper about poem generation and learn about LSTM&lt;br /&gt;
: 2016-05-05 : check in and have an overview of generation model&lt;br /&gt;
&lt;br /&gt;
===jiyuan zhang===&lt;br /&gt;
: 2016-05-01~06 :modify input format and run lstmrbm model (16-beat,32-beat,bar)&lt;br /&gt;
: 2016-05-09~13:&lt;br /&gt;
   Modify model parameters  and run model ，the result is not ideal  yet &lt;br /&gt;
   According to teacher Wang's opinion, in the generation stage,replace random generation with the maximum probability generation&lt;br /&gt;
&lt;br /&gt;
: 2016-05-24~27 :check the blog's codes  and  understand  the model and input format details  on the blog&lt;/div&gt;</summary>
		<author><name>Fengyang</name></author>	</entry>

	</feed>