<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://index.cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://index.cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=%E8%AE%BA%E6%96%87%E5%88%97%E8%A1%A8</id>
		<title>论文列表 - 版本历史</title>
		<link rel="self" type="application/atom+xml" href="http://index.cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=%E8%AE%BA%E6%96%87%E5%88%97%E8%A1%A8"/>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php?title=%E8%AE%BA%E6%96%87%E5%88%97%E8%A1%A8&amp;action=history"/>
		<updated>2026-04-06T16:51:01Z</updated>
		<subtitle>本wiki的该页面的版本历史</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php?title=%E8%AE%BA%E6%96%87%E5%88%97%E8%A1%A8&amp;diff=43936&amp;oldid=prev</id>
		<title>2026年2月13日 (五) 15:34 Cslt</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php?title=%E8%AE%BA%E6%96%87%E5%88%97%E8%A1%A8&amp;diff=43936&amp;oldid=prev"/>
				<updated>2026-02-13T15:34:25Z</updated>
		
		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table class='diff diff-contentalign-left'&gt;
				&lt;col class='diff-marker' /&gt;
				&lt;col class='diff-content' /&gt;
				&lt;col class='diff-marker' /&gt;
				&lt;col class='diff-content' /&gt;
				&lt;tr style='vertical-align: top;'&gt;
				&lt;td colspan='2' style=&quot;background-color: white; color:black; text-align: center;&quot;&gt;←上一版本&lt;/td&gt;
				&lt;td colspan='2' style=&quot;background-color: white; color:black; text-align: center;&quot;&gt;2026年2月13日 (五) 15:34的版本&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;第1行：&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;第1行：&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[1] Haoran Sun, Chen Chen, Lantian Li, Dong Wang, CycleFlow: Purify Information Factors by Cycle Loss, Odyssey 2022 (Best Student Paper)&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[1] Haoran Sun, Chen Chen, Lantian Li, Dong Wang, CycleFlow: Purify Information Factors by Cycle Loss, Odyssey 2022 (Best Student Paper)&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[2] Haoran Sun, Dong Wang, Lantian Li, Chen Chen, Thomas Fang Zheng, Random Cycle Loss and Its Application to Voice Conversion, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.45, no.8, August, 2023.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[2] Haoran Sun, Dong Wang, Lantian Li, Chen Chen, Thomas Fang Zheng, Random Cycle Loss and Its Application to Voice Conversion, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.45, no.8, August, 2023.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[3] Chen Chen, Dong Wang, Thomas Fang Zheng, CN-CVS: A Mandarin Audio-Visual Dataset for Large Vocabulary Continuous Visual to Speech Synthesis, ICASSP 2023.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[3] Chen Chen, Dong Wang, Thomas Fang Zheng, CN-CVS: A Mandarin Audio-Visual Dataset for Large Vocabulary Continuous Visual to Speech Synthesis, ICASSP 2023.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[4] Chen Chen, Xiaolou Li, Zehua Liu, Lantian Li, Dong Wang, Quantitative Analysis of Audio-Visual Tasks: An Information-Theoretic Perspective, ISCSLP 2024.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[4] Chen Chen, Xiaolou Li, Zehua Liu, Lantian Li, Dong Wang, Quantitative Analysis of Audio-Visual Tasks: An Information-Theoretic Perspective, ISCSLP 2024.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[5] Liu Z, Li X, Chen C, et al. CNVSRC 2024: The Second Chinese Continuous Visual Speech Recognition Challenge[J]. Interspeech 2025&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[5] Liu Z, Li X, Chen C, et al. CNVSRC 2024: The Second Chinese Continuous Visual Speech Recognition Challenge[J]. Interspeech 2025&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[6] Chen Chen, Zehua Liu, Xiaolou Li, Lantian Li, Dong Wang, CNVSRC 2023: The First Chinese Continuous Visual Speech Recognition Challenge, Interspeech 2024&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[6] Chen Chen, Zehua Liu, Xiaolou Li, Lantian Li, Dong Wang, CNVSRC 2023: The First Chinese Continuous Visual Speech Recognition Challenge, Interspeech 2024&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[7] Ying Shi, Lantian Li, Shi Yin, Dong Wang, Jiqing Han, Serialized Output Training by Learned Dominance, Interspeech 2024&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[7] Ying Shi, Lantian Li, Shi Yin, Dong Wang, Jiqing Han, Serialized Output Training by Learned Dominance, Interspeech 2024&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[8] Yuan J, Shi Y, Wang D, et al. MT-HuBERT: Self-Supervised Mix-Training for Few-Shot Keyword Spotting in Mixed Speech[J]. ICASSP 2026.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[8] Yuan J, Shi Y, Wang D, et al. MT-HuBERT: Self-Supervised Mix-Training for Few-Shot Keyword Spotting in Mixed Speech[J]. ICASSP 2026.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[9] Junming Yuan, Ying Shi, Lantian Li, Dong Wang, Askar Hamdulla, Few-Shot Keyword Spotting from Mixed Speech, Interspeech 2024&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[9] Junming Yuan, Ying Shi, Lantian Li, Dong Wang, Askar Hamdulla, Few-Shot Keyword Spotting from Mixed Speech, Interspeech 2024&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[10] Shi, Y., Wang, D.*, Li, L., Han, J., Yin, S. (2023) Spot Keywords From Very Noisy and Mixed Speech. Proc. INTERSPEECH 2023, 1488-1492&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[10] Shi, Y., Wang, D.*, Li, L., Han, J., Yin, S. (2023) Spot Keywords From Very Noisy and Mixed Speech. Proc. INTERSPEECH 2023, 1488-1492&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[11] Lin W, Chen J, Wang T, et al. Neural Scoring: A Refreshed End-to-End Approach for Speaker Verification in Complex Conditions[J]. IEEE Signal Processing Letters, 2025.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[11] Lin W, Chen J, Wang T, et al. Neural Scoring: A Refreshed End-to-End Approach for Speaker Verification in Complex Conditions[J]. IEEE Signal Processing Letters, 2025.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[12] Ying Shi, Lantian Li, Dong Wang, Jiqing Han, Keyword Guided Target Speech Recognition, IEEE Signal Processing Letters, 2024&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[12] Ying Shi, Lantian Li, Dong Wang, Jiqing Han, Keyword Guided Target Speech Recognition, IEEE Signal Processing Letters, 2024&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[13] Zehua Liu, Xiaolou Li, Li Guo, Lantian Li, Dong Wang, Leveraging Large Language Models in Visual Speech Recognition: Model Scaling, Context-Aware Decoding, and Iterative Polishing, APSIPA 2025.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[13] Zehua Liu, Xiaolou Li, Li Guo, Lantian Li, Dong Wang, Leveraging Large Language Models in Visual Speech Recognition: Model Scaling, Context-Aware Decoding, and Iterative Polishing, APSIPA 2025.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[14] 陈琛，语音转换任务中的信息解耦优化方法研究，硕士论文，清华大学，2024.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[14] 陈琛，语音转换任务中的信息解耦优化方法研究，硕士论文，清华大学，2024.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[15] Liu Z, Li X, Chen C, et al. AlignVSR: Audio-visual cross-modal alignment for visual speech recognition[C]//Proceedings of the 2025 11th International Conference on Communication and Information Processing. 2025: 161-165.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[15] Liu Z, Li X, Chen C, et al. AlignVSR: Audio-visual cross-modal alignment for visual speech recognition[C]//Proceedings of the 2025 11th International Conference on Communication and Information Processing. 2025: 161-165.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[16] Cai Y, Li L, Abel A, et al. Maximum Gaussianality training for deep speaker vector normalization[J]. Pattern Recognition, 2024, 145: 109977&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[16] Cai Y, Li L, Abel A, et al. Maximum Gaussianality training for deep speaker vector normalization[J]. Pattern Recognition, 2024, 145: 109977&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[17] Lantian Li, Ruiqi Liu, Jiawen Kang, Yue Fa, Hao Cui, Yunqi Cai, Ravichander Vipperla, Thomas Fang Zheng and Dong Wang. &amp;quot;CN-Celeb: multi-genre speaker recognition&amp;quot;, Speech Communication, 2022.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[17] Lantian Li, Ruiqi Liu, Jiawen Kang, Yue Fa, Hao Cui, Yunqi Cai, Ravichander Vipperla, Thomas Fang Zheng and Dong Wang. &amp;quot;CN-Celeb: multi-genre speaker recognition&amp;quot;, Speech Communication, 2022.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;[18] Cai, Y., Li, J. &amp;amp; Wang, D. Fast and generalizable micromagnetic simulation with deep neural nets. Nat Mach Intell 6, 1330–1343 (2024).&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;*&lt;/ins&gt;[18] Cai, Y., Li, J. &amp;amp; Wang, D. Fast and generalizable micromagnetic simulation with deep neural nets. Nat Mach Intell 6, 1330–1343 (2024).&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	<entry>
		<id>http://index.cslt.org/mediawiki/index.php?title=%E8%AE%BA%E6%96%87%E5%88%97%E8%A1%A8&amp;diff=43935&amp;oldid=prev</id>
		<title>Cslt：以“[1] Haoran Sun, Chen Chen, Lantian Li, Dong Wang, CycleFlow: Purify Information Factors by Cycle Loss, Odyssey 2022 (Best Student Paper) [2] Haoran Sun, Dong Wang, L...”为内容创建页面</title>
		<link rel="alternate" type="text/html" href="http://index.cslt.org/mediawiki/index.php?title=%E8%AE%BA%E6%96%87%E5%88%97%E8%A1%A8&amp;diff=43935&amp;oldid=prev"/>
				<updated>2026-02-13T15:33:44Z</updated>
		
		<summary type="html">&lt;p&gt;以“[1] Haoran Sun, Chen Chen, Lantian Li, Dong Wang, CycleFlow: Purify Information Factors by Cycle Loss, Odyssey 2022 (Best Student Paper) [2] Haoran Sun, Dong Wang, L...”为内容创建页面&lt;/p&gt;
&lt;p&gt;&lt;b&gt;新页面&lt;/b&gt;&lt;/p&gt;&lt;div&gt;[1] Haoran Sun, Chen Chen, Lantian Li, Dong Wang, CycleFlow: Purify Information Factors by Cycle Loss, Odyssey 2022 (Best Student Paper)&lt;br /&gt;
[2] Haoran Sun, Dong Wang, Lantian Li, Chen Chen, Thomas Fang Zheng, Random Cycle Loss and Its Application to Voice Conversion, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.45, no.8, August, 2023.&lt;br /&gt;
[3] Chen Chen, Dong Wang, Thomas Fang Zheng, CN-CVS: A Mandarin Audio-Visual Dataset for Large Vocabulary Continuous Visual to Speech Synthesis, ICASSP 2023.&lt;br /&gt;
[4] Chen Chen, Xiaolou Li, Zehua Liu, Lantian Li, Dong Wang, Quantitative Analysis of Audio-Visual Tasks: An Information-Theoretic Perspective, ISCSLP 2024.&lt;br /&gt;
[5] Liu Z, Li X, Chen C, et al. CNVSRC 2024: The Second Chinese Continuous Visual Speech Recognition Challenge[J]. Interspeech 2025&lt;br /&gt;
[6] Chen Chen, Zehua Liu, Xiaolou Li, Lantian Li, Dong Wang, CNVSRC 2023: The First Chinese Continuous Visual Speech Recognition Challenge, Interspeech 2024&lt;br /&gt;
[7] Ying Shi, Lantian Li, Shi Yin, Dong Wang, Jiqing Han, Serialized Output Training by Learned Dominance, Interspeech 2024&lt;br /&gt;
[8] Yuan J, Shi Y, Wang D, et al. MT-HuBERT: Self-Supervised Mix-Training for Few-Shot Keyword Spotting in Mixed Speech[J]. ICASSP 2026.&lt;br /&gt;
[9] Junming Yuan, Ying Shi, Lantian Li, Dong Wang, Askar Hamdulla, Few-Shot Keyword Spotting from Mixed Speech, Interspeech 2024&lt;br /&gt;
[10] Shi, Y., Wang, D.*, Li, L., Han, J., Yin, S. (2023) Spot Keywords From Very Noisy and Mixed Speech. Proc. INTERSPEECH 2023, 1488-1492&lt;br /&gt;
[11] Lin W, Chen J, Wang T, et al. Neural Scoring: A Refreshed End-to-End Approach for Speaker Verification in Complex Conditions[J]. IEEE Signal Processing Letters, 2025.&lt;br /&gt;
[12] Ying Shi, Lantian Li, Dong Wang, Jiqing Han, Keyword Guided Target Speech Recognition, IEEE Signal Processing Letters, 2024&lt;br /&gt;
[13] Zehua Liu, Xiaolou Li, Li Guo, Lantian Li, Dong Wang, Leveraging Large Language Models in Visual Speech Recognition: Model Scaling, Context-Aware Decoding, and Iterative Polishing, APSIPA 2025.&lt;br /&gt;
[14] 陈琛，语音转换任务中的信息解耦优化方法研究，硕士论文，清华大学，2024.&lt;br /&gt;
[15] Liu Z, Li X, Chen C, et al. AlignVSR: Audio-visual cross-modal alignment for visual speech recognition[C]//Proceedings of the 2025 11th International Conference on Communication and Information Processing. 2025: 161-165.&lt;br /&gt;
[16] Cai Y, Li L, Abel A, et al. Maximum Gaussianality training for deep speaker vector normalization[J]. Pattern Recognition, 2024, 145: 109977&lt;br /&gt;
[17] Lantian Li, Ruiqi Liu, Jiawen Kang, Yue Fa, Hao Cui, Yunqi Cai, Ravichander Vipperla, Thomas Fang Zheng and Dong Wang. &amp;quot;CN-Celeb: multi-genre speaker recognition&amp;quot;, Speech Communication, 2022.&lt;br /&gt;
[18] Cai, Y., Li, J. &amp;amp; Wang, D. Fast and generalizable micromagnetic simulation with deep neural nets. Nat Mach Intell 6, 1330–1343 (2024).&lt;/div&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	</feed>