“2025-03-03”版本间的差异

来自cslt Wiki
跳转至: 导航搜索
第208行: 第208行:
 
|Turi
 
|Turi
 
||
 
||
*  
+
* Thesis writing
 +
* Result with LM[https://z1et6d3xtb.feishu.cn/docx/JvDsd8zR4oMwnyxQEQdckpMjn7m?from=from_copylink]
 
||
 
||
 
*  
 
*  

2025年3月3日 (一) 10:53的版本

People This Week Next Week Task Tracking (DeadLine)
Dong Wang
Lantian Li
Ying Shi
  • Prepare Ascend Sever environment
  • training Conditional Chain overlap ASR model with Hierachical-Transformer here
Zhenghai You
Junming Yuan
  • Finish MPC-HuBERT pretrain.
  • Double check the related experimental code.
    • MT-HuBERT(in progress) & Cocktail-HuBERT need re-pretrain.
    • The results of other baseline in here
Xiaolou Li
Zehua Liu
  • Paper Reading and Sharing in last Friday
  • Writing Vision Language Model code
  • Writing NSFC document
Pengqi Li
  • Prepare the AI course for Tsinghua University Junior High School.
  • Using t-SNE to visualize the factorized content vector.
    • Next step is to color(speaker information importance or not) each point.
Wan Lin
  • try some methods for clean performance(no improvement)
  • supply experiments for other tests
Tianhao Wang
  • sound separation: 2-mix and 3-mix model training
  • weekly report
  • subset data training
Xiaoxue Luo
  • generation of multi-mix audio data and did some test experiments.
  • read papers
Zhenyu Zhou
  • finish graduation thesis
Junhui Chen
  • Reproducing speaker diarization method for NS (debugging...)
  • read paper
Jiaying Wang
  • debug ctc loss part[1]
Yu Zhang
Wenqiang Du
  • Primary handbook's PPT (24/44)
  • Continue to check Primary and middle handbook(Completed this week)
  • Speech cloning sample for the company
Yang Wei
  • Tuning text enroll kws model for dialect data with linear layer. (recall: 65%->85%->94%)
Turi
  • Thesis writing
  • Result with LM[2]
Yue Gu
  • finish some exps, but nothing is improved.
  • finish a proposal,I will present it recently
Qi Qu
  • Applying pre-prod eval routine on text-enroll KWS models: the ideal thresholds for each keyword vary significantly.