|
|
(3位用户的3个中间修订版本未显示) |
第76行: |
第76行: |
| || | | || |
| * Debug the Chinese VTS (in training already) | | * Debug the Chinese VTS (in training already) |
| + | * Process the data of CVS3 |
| * Write the report of VTS project (main work) | | * Write the report of VTS project (main work) |
| || | | || |
第111行: |
第112行: |
| |Wan Lin | | |Wan Lin |
| || | | || |
− | * | + | * NS:detection (edit code with Chen) |
| + | ** EER of the model decrease faster in the previous epochs |
| || | | || |
| * | | * |
第144行: |
第146行: |
| |Zhenyu Zhou | | |Zhenyu Zhou |
| || | | || |
− | * | + | *Reproduction:conditional TasNet [https://z1et6d3xtb.feishu.cn/docx/D2UQdxMBvojkF9xCXGfcFBLGned] |
| || | | || |
| * | | * |
第203行: |
第205行: |
| |Yang Wei | | |Yang Wei |
| || | | || |
− | * | + | * Write text enroll KWS model document. |
| + | * Prepare data and code for Aibabel data finetuning. |
| || | | || |
| * | | * |
People |
This Week |
Next Week |
Task Tracking (DeadLine)
|
Dong Wang
|
- AI Medical sector 2 chapters done
|
|
|
Lantian Li
|
- Submit three papers supporting ICCIP 2024.
- Go on designing 2025 AI daily posts
- Attend CSTR 40th anniversary
|
|
|
Ying Shi
|
- Stop strategy for Cohort Overlap ASR here
|
|
|
Zhenghai You
|
- Huawei project (Unsuccessful IRA) [1]
- Summarize SPK-AUG experiments[2]
|
|
|
Junming Yuan
|
- paper reading
- prepare to reproduce cocktail HuBERT (in progress)
|
|
|
Chen Chen
|
|
|
|
Xiaolou Li
|
- Debug the Chinese VTS (in training already)
- Process the data of CVS3
- Write the report of VTS project (main work)
|
|
|
Zehua Liu
|
- In-Context-Learning(if sentence is very long,context seems fail)still finding reason
- (context<30s)45.30% | 44.69% (context = 30s) | 46.02%(context = 120s)
- Writing VTS project document
|
|
|
Pengqi Li
|
- New Process of consistency of TAO and LayerCAM.[3]
|
|
|
Wan Lin
|
- NS:detection (edit code with Chen)
- EER of the model decrease faster in the previous epochs
|
|
|
Tianhao Wang
|
- investigating some new approach for target sound separation
- prepare the code for LoRA tuned CLAP
|
|
|
Xiaoxue Luo
|
|
|
|
Zhenyu Zhou
|
- Reproduction:conditional TasNet [4]
|
|
|
Junhui Chen
|
- NS with frame-level detection loss
- use silero-vad
- Model is training, seems EER decrease faster.
|
|
|
Jiaying Wang
|
|
|
|
Yu Zhang
|
- SocioDojo
- with cash ratio risk aware, and change information sources, seems have a decent risk control over Nasdaq 100 index [5]
- Some paper reading and report in RoyalFlush, get some idea (mainly about LLM for time series task)
|
|
|
Wenqiang Du
|
- Training of New Dialect Models(Yi language )
|
|
|
Yang Wei
|
- Write text enroll KWS model document.
- Prepare data and code for Aibabel data finetuning.
|
|
|
Lily
|
|
|
|
Turi
|
- LoRA finetuning (Result is not good)
- Data cleaning
|
|
|
Yue Gu
|
- read several paper about speech tokenizer. I want to design a encoder, which processes different size feature frame and construct several different codebooks, to extract personality from the varing speech speed. It is still in progress.
- paper writing
|
|
|
Qi Qu
|
- KWS:
- Yi (Liangshan, Sichuan) dataset prepared for training; dataset to be annotated for testing.
- Experiments on model quantization for NPU devices: i16 quantization arrives at a balance between accuracy and efficiency (~2ms per inference, compared to ~250ms for non-quantized); more calibration data needed for further confirmation.
- Full-featured demo (recording + feature extraction + model inference) for NPU devices in development.
|
|
|