|
|
| (5位用户的7个中间修订版本未显示) |
| 第58行: |
第58行: |
| | * double check mixed Hubert code: | | * double check mixed Hubert code: |
| | ** fix some bugs (time-mask.etc) | | ** fix some bugs (time-mask.etc) |
| − | ** time-mask vs. feat mask: (Top-1 acc, EER): (27.98%, 23.17%) vs.(23.19%, 25.99%) | + | ** feat-mask vs. time-mask: (Top-1 acc, EER): (27.98%, 23.17%) vs.(23.19%, 25.99%) |
| | ** softmax+CE --> sigmoid+BCE still have problem. | | ** softmax+CE --> sigmoid+BCE still have problem. |
| | || | | || |
| 第86行: |
第86行: |
| | |Zehua Liu | | |Zehua Liu |
| | || | | || |
| − | * | + | *Finish VTS document with Xiaolou |
| | + | *Reorganize my previous code about VSR-LLM |
| | + | *Run some exps(still training) |
| | || | | || |
| | * | | * |
| 第97行: |
第99行: |
| | |Pengqi Li | | |Pengqi Li |
| | || | | || |
| − | * | + | * Implement TAO and LayerCAM on the verification task[https://z1et6d3xtb.feishu.cn/docx/EBB3dcGzioCEoaxh8vUchVPgn9c] |
| | + | * Evaluate it reliability. |
| | || | | || |
| | * | | * |
| 第136行: |
第139行: |
| | || | | || |
| | * Model quantification phase document | | * Model quantification phase document |
| − | paper reading
| + | * paper reading |
| | || | | || |
| | * | | * |
| 第147行: |
第150行: |
| | |Junhui Chen | | |Junhui Chen |
| | || | | || |
| − | * | + | * Some vb1 filter exps with NS as Wan Lin marked. |
| | + | * Prepare vb1 test data and code, ready for vb2 training. |
| | || | | || |
| | * | | * |
| 第182行: |
第186行: |
| | |Wenqiang Du | | |Wenqiang Du |
| | || | | || |
| − | * | + | * optimize primary school handbook(14/45) |
| | + | * Some of the company's work |
| | + | ** Training of New Dialect Models |
| | + | ** Project application |
| | || | | || |
| | * | | * |
| People |
This Week |
Next Week |
Task Tracking (DeadLine)
|
| Dong Wang
|
- AIGraph high education version
- Prepare AIgraph Large Model version
- NMI paper publication staff
|
|
|
| Lantian Li
|
- AI-Graph EN (1/4)
- Huawei Project Proposal v1.0
- First Lesson on 24-fall AI Undergraduates
|
|
|
| Ying Shi
|
- Huawei project proposal
- Optimize the Text-enroll KWS code
- improve readability.
- remove redundant code.
|
|
|
| Zhenghai You
|
- Exploring the generality of spk aug on different data and structures[1]
|
|
|
| Junming Yuan
|
- double check mixed Hubert code:
- fix some bugs (time-mask.etc)
- feat-mask vs. time-mask: (Top-1 acc, EER): (27.98%, 23.17%) vs.(23.19%, 25.99%)
- softmax+CE --> sigmoid+BCE still have problem.
|
|
|
| Xiaolou Li
|
- Writing VTS documents
- Paper Reading & Preparing for Report
- Exp on LRS3
- LLM: LLaMA2 -> LLaMA3.1 (30h ↓0.4%)
- Grouping LLaMA2: (443h ↑0.5%, 30h ↓2.5%)
- Rethinking the method to inject information (ablation study first)
|
|
|
| Zehua Liu
|
- Finish VTS document with Xiaolou
- Reorganize my previous code about VSR-LLM
- Run some exps(still training)
|
|
|
| Pengqi Li
|
- Implement TAO and LayerCAM on the verification task[2]
- Evaluate it reliability.
|
|
|
| Wan Lin
|
- VC2 pre-train; VB1+VC2 mix-tuning
- Data filter in VB1: 1.25% EER in vox1-o
- VB1 pre-train; VC2 fine-tuning
- VB1 pre-train: 2.61% EER in vox1-o
- VC2 fine-tuning: maybe couldn't reach better performance
|
|
|
| Tianhao Wang
|
- IS24 paper reading & weekly report
- sound separartion project proposal
|
|
|
| Zhenyu Zhou
|
- Model quantification phase document
- paper reading
|
|
|
| Junhui Chen
|
- Some vb1 filter exps with NS as Wan Lin marked.
- Prepare vb1 test data and code, ready for vb2 training.
|
|
|
| Jiaying Wang
|
|
|
|
| Yu Zhang
|
- Dataset collection from THS
- Retraining R^2 SAC paper, with same env still failed (TCN ACC: 0.708, RECALL: 0.183), will check with Han this week
- Paper reading and some plan (report this Fri)
|
|
|
| Wenqiang Du
|
- optimize primary school handbook(14/45)
- Some of the company's work
- Training of New Dialect Models
- Project application
|
|
|
| Yang Wei
|
- Train text enroll KWS model with aishell and kespeech data.
- Prepare live broadcast
|
|
|
| Lily
|
- Prepare for holiday course(October 2nd、3rd) and online-course
- AI radiance's daily work
|
|
|
| Turi
|
- Trained conformer on Sagalee data excluding utterances containing digits
- Achieved 21.28% WER, 2.65 WER reduction
- Preparing KWS data from Sagalee dataset using MFA
|
|
|
| Yue Gu
|
- paper writing
- open the code
- prepare for the presentation
|
|
|
| Qi Qu
|
- KWS:
- Finding ideal thresholds for b0-models in predefined scenes: Chinese Mandarin, Cantonese, Uyghur and Kazakh.
- Finding ideal thresholds for b6-models with fixed b0-model thresholds.
- AED:
- Fixing parameters of Fbank feature extraction for CED and retraining classifiers.
|
|
|