“ASR:2015-10-12”版本间的差异
来自cslt Wiki
(以“==Speech Processing == === AM development === ==== Environment ==== * grid-12 GPU is transferred to grid-18 * grid-14 is unstable ==== RNN AM==== *train monophone...”为内容创建页面) |
(→Speech Processing) |
||
第3行: | 第3行: | ||
==== Environment ==== | ==== Environment ==== | ||
− | * | + | * repair laptop |
− | + | ||
==== RNN AM==== | ==== RNN AM==== | ||
第13行: | 第13行: | ||
* train RNN MPE using large dataset--mengyuan | * train RNN MPE using large dataset--mengyuan | ||
:* diverge problem | :* diverge problem | ||
+ | :* try adaptation method | ||
+ | |||
====Learning rate tunning==== | ====Learning rate tunning==== | ||
− | |||
− | |||
− | |||
− | |||
* sequence training -Xiangyu | * sequence training -Xiangyu | ||
第36行: | 第34行: | ||
===Ivector&Dvector based ASR=== | ===Ivector&Dvector based ASR=== | ||
− | |||
− | |||
* dark knowledge | * dark knowledge | ||
:* has much worse performance than baseline (EER: base 29% dark knowledge 48%) | :* has much worse performance than baseline (EER: base 29% dark knowledge 48%) | ||
− | + | * binary ivector | |
− | + | ||
− | * binary ivector | + | |
* metric learning | * metric learning | ||
===language vector=== | ===language vector=== | ||
* hold | * hold | ||
− | * write a paper--zhiyuan | + | * write a paper--zhiyuan |
+ | * language vector is added to multi hidden layers--zhiyuan | ||
* RNN language vector | * RNN language vector | ||
− | :* hold | + | :*hold |
− | ===multi-GPU | + | ===multi-GPU=== |
* multi-stream training --Sheng Su | * multi-stream training --Sheng Su | ||
− | :* two GPUs work well, but four GPUs divergent | + | :*two GPUs work well, but four GPUs divergent |
* solve the problem of buffer-- Sheng Su | * solve the problem of buffer-- Sheng Su | ||
+ | * kaldi-nnet3 --Xuewei | ||
+ | |||
+ | ===multi-task=== | ||
+ | * write code according to selt-information neural structure learning --mengyuan | ||
+ | * speech rate learning --xiangyu | ||
===Neutral picture style transfer== | ===Neutral picture style transfer== |
2015年10月12日 (一) 09:07的最后版本
目录
- 1 Speech Processing
- 2 =Neutral picture style transfer
- 3 Text Processing
- 4 financial group
Speech Processing
AM development
Environment
- repair laptop
RNN AM
- train monophone RNN --zhiyuan
- decode using 5-gram
- the train method of batch
- test using another test set
- train RNN MPE using large dataset--mengyuan
- diverge problem
- try adaptation method
Learning rate tunning
- sequence training -Xiangyu
Mic-Array
- hold
- compute EER with kaldi
Data selection unsupervised learning
- hold
- acoustic feature based submodular using Pingan dataset --zhiyong
- write code to speed up --zhiyong
- curriculum learning --zhiyong
RNN-DAE(Deep based Auto-Encode-RNN)
- RNN-DAE has worse performance than DNN-DAE because training dataset is small
- extract real room impulse to generate WSJ reverberation data, and then train RNN-DAE
Ivector&Dvector based ASR
- dark knowledge
- has much worse performance than baseline (EER: base 29% dark knowledge 48%)
- binary ivector
- metric learning
language vector
- hold
- write a paper--zhiyuan
- language vector is added to multi hidden layers--zhiyuan
- RNN language vector
- hold
multi-GPU
- multi-stream training --Sheng Su
- two GPUs work well, but four GPUs divergent
- solve the problem of buffer-- Sheng Su
- kaldi-nnet3 --Xuewei
multi-task
- write code according to selt-information neural structure learning --mengyuan
- speech rate learning --xiangyu
=Neutral picture style transfer
- hold
- reproduced the result of the paper "A neutral algorithm of artistic style" --Zhiyuan, Xuewei
- while subject to the GPU's memory, limited to inception net with sgd optimizer (VGG network with the default L-BFGS optimizer consumes very much memory, which is better)
Multi-task learning
- train model using speech rate --xiangyu
- speech recognition plus speaker reconition --xiangyu,lantian,zhiyuan
Text Processing
RNN LM
- character-lm rnn(hold)
- lstm+rnn
- check the lstm-rnnlm code about how to Initialize and update learning rate.(hold)
Neural Based Document Classification
- (hold)
RNN Rank Task
- Test.
- Paper: RNN Rank Net.
- (hold)
- Output rank information.
Graph RNN
- Entity path embeded to entity.
- (hold)
RNN Word Segment
- Set bound to word segment.
- (hold)
Seq to Seq(09-15)
- Review papers.
- Reproduce baseline. (08-03 <--> 08-17)
Order representation
- Nested Dropout
- semi-linear --> neural based auto-encoder.
- modify the objective function(hold)
Balance Representation
- Find error signal
Recommendation
- Reproduce baseline.
- LDA matrix dissovle.
- LDA (Text classification & Recommendation System) --> AAAI
RNN based QA
- Read Source Code.
- Attention based QA.
- Coding.
RNN Poem Process
- Seq based BP.
- (hold)
Text Group Intern Project
Buddhist Process
- (hold)
RNN Poem Process
- Done by Haichao yu & Chaoyuan zuo Mentor : Tianyi Luo.
RNN Document Vector
- (hold)
Image Baseline
- Demo Release.
- Paper Report.
- Read CNN Paper.
Text Intuitive Idea
Trace Learning
- (Hold)
Match RNN
- (Hold)
financial group
model research
- RNN
- online model, update everyday
- modify cost function and learning method
- add more feature
rule combination
- GA method to optimize the model
basic rule
- classical tenth model
multiple-factor
- add more factor
- use sparse model
display
- bug fixed
- buy rule fixed
data
- data api
- download the future data and factor data