<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://www.cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://www.cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-07-25</id>
		<title>2014-07-25 - 版本历史</title>
		<link rel="self" type="application/atom+xml" href="http://www.cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-07-25"/>
		<link rel="alternate" type="text/html" href="http://www.cslt.org/mediawiki/index.php?title=2014-07-25&amp;action=history"/>
		<updated>2026-04-03T21:25:45Z</updated>
		<subtitle>本wiki的该页面的版本历史</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://www.cslt.org/mediawiki/index.php?title=2014-07-25&amp;diff=10525&amp;oldid=prev</id>
		<title>Cslt：以内容“==Resoruce Building==  == Leftover questions==  * Investigating LOUDS FST.  * CLG embedded decoder plus online compiler. * DNN-GMM co-training  == AM development ==  ==...”创建新页面</title>
		<link rel="alternate" type="text/html" href="http://www.cslt.org/mediawiki/index.php?title=2014-07-25&amp;diff=10525&amp;oldid=prev"/>
				<updated>2014-07-25T02:13:03Z</updated>
		
		<summary type="html">&lt;p&gt;以内容“==Resoruce Building==  == Leftover questions==  * Investigating LOUDS FST.  * CLG embedded decoder plus online compiler. * DNN-GMM co-training  == AM development ==  ==...”创建新页面&lt;/p&gt;
&lt;p&gt;&lt;b&gt;新页面&lt;/b&gt;&lt;/p&gt;&lt;div&gt;==Resoruce Building==&lt;br /&gt;
&lt;br /&gt;
== Leftover questions==&lt;br /&gt;
&lt;br /&gt;
* Investigating LOUDS FST. &lt;br /&gt;
* CLG embedded decoder plus online compiler.&lt;br /&gt;
* DNN-GMM co-training&lt;br /&gt;
&lt;br /&gt;
== AM development ==&lt;br /&gt;
&lt;br /&gt;
=== Sparse DNN ===&lt;br /&gt;
* WJS sparse DNN shows a slightly better than non-sparse cases when the network is in a large scale&lt;br /&gt;
* Pre-training does work for DNN training (for both 4/5/6 layers)&lt;br /&gt;
&lt;br /&gt;
===Noise training===&lt;br /&gt;
:* Journal paper writing on going&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Multilingual ASR===&lt;br /&gt;
&lt;br /&gt;
* With multlingual training, performance is largely retained with most of known test sets; &lt;br /&gt;
* However for unknown accents, performance is not stable&lt;br /&gt;
&lt;br /&gt;
==Drop out &amp;amp; convolutional network==&lt;br /&gt;
&lt;br /&gt;
* Zhiyong will study drop out&lt;br /&gt;
* Zhiyong &amp;amp; Mengyuan will study convolutional network&lt;br /&gt;
&lt;br /&gt;
===Denoising &amp;amp; Farfield ASR===&lt;br /&gt;
&lt;br /&gt;
* Use an reverberation tool to generate a new set of datasets&lt;br /&gt;
&lt;br /&gt;
*  xEnt  results(eval 92):&lt;br /&gt;
               before adaptation    after adaptation&lt;br /&gt;
    clean:      -                          -&lt;br /&gt;
    near:       19.25                    12.94&lt;br /&gt;
    far:        59.38                    40.46&lt;br /&gt;
&lt;br /&gt;
*  Lasso-based reverberation cancellation got initial clean data&lt;br /&gt;
&lt;br /&gt;
===VAD===&lt;br /&gt;
&lt;br /&gt;
* Waiting for engineering work&lt;br /&gt;
&lt;br /&gt;
===Scoring===&lt;br /&gt;
&lt;br /&gt;
* Refine the acoustic model with AMIDA database. problem solved by involving both wsj and AMIDA.&lt;br /&gt;
&lt;br /&gt;
===Embedded decoder===&lt;br /&gt;
&lt;br /&gt;
* Chatting LM release&lt;br /&gt;
* Train two smaller network: 500x4+600, 400x4+500: on going&lt;br /&gt;
* Need to upload the new client code onto git&lt;br /&gt;
* Build a new graph with MPE3 am and chatting LM.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==LM development==&lt;br /&gt;
&lt;br /&gt;
===Domain specific LM===&lt;br /&gt;
&lt;br /&gt;
h2. Domain specific LM construction&lt;br /&gt;
&lt;br /&gt;
h3. TAG LM&lt;br /&gt;
&lt;br /&gt;
* TAG still problematic with all-to-number tag&lt;br /&gt;
* check the randomness of the number tag.&lt;br /&gt;
&lt;br /&gt;
h3. Chatting LM&lt;br /&gt;
* Building chatting lexicon&lt;br /&gt;
* First version released (80k lexicon)&lt;br /&gt;
&lt;br /&gt;
==Word2Vector==&lt;br /&gt;
&lt;br /&gt;
===W2V based doc classification===&lt;br /&gt;
&lt;br /&gt;
* Initial results variable Bayesian GMM obtained. Performance is not as good as the conventional GMM.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Semantic word tree==&lt;br /&gt;
&lt;br /&gt;
:* Version v2.0 released (filter with query log)&lt;br /&gt;
:* Please deliver to /nfs/disk/perm/data/corpora/semanticTree (Xingchao)&lt;br /&gt;
:* Version v3.0 under going. Further refinement with Baidu Baike hierarchy&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==NN LM==&lt;br /&gt;
&lt;br /&gt;
* Character-based NNLM (6700 chars, 7gram), 500M data training done.&lt;br /&gt;
:* Inconsistent pattern in WER were found on Tenent test sets&lt;br /&gt;
:* probably need to use another test set to do investigation. &lt;br /&gt;
* Investigate MS RNN LM training&lt;br /&gt;
&lt;br /&gt;
==Speaker ID==&lt;br /&gt;
&lt;br /&gt;
* reading materials &lt;br /&gt;
* prepare to run sre08&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Translation==&lt;br /&gt;
* collecting more data (Xinhua parallel text, bible, name entity)  for the second version&lt;br /&gt;
* work into text alignment&lt;br /&gt;
* Will release v2.0 today&lt;/div&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	</feed>