﻿<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:trackback="http://madskills.com/public/xml/rss/module/trackback/" xmlns:wfw="http://wellformedweb.org/CommentAPI/" xmlns:slash="http://purl.org/rss/1.0/modules/slash/"><channel><title>C++博客-sunrise-随笔分类-自然语言处理</title><link>http://www.cppblog.com/sunrise/category/18950.html</link><description>每天不断学习，才能不断提升自己。

欢迎交流 QQ：703979707

我的铺子：http://www.u148.net/u/lwx</description><language>zh-cn</language><lastBuildDate>Thu, 20 Sep 2012 11:03:41 GMT</lastBuildDate><pubDate>Thu, 20 Sep 2012 11:03:41 GMT</pubDate><ttl>60</ttl><item><title>NLP数据收集</title><link>http://www.cppblog.com/sunrise/archive/2012/09/20/191408.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Thu, 20 Sep 2012 09:29:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/09/20/191408.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/191408.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/09/20/191408.html#Feedback</comments><slash:comments>0</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/191408.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/191408.html</trackback:ping><description><![CDATA[<p>目前网上可供下载的<a title="数据" href="http://www.haogongju.net/tag/%E6%95%B0%E6%8D%AE" target="_blank">数据</a>众多，但是内容庞杂，把其中比较有用的数据找了出来。<br /><br />wiki系：<br />wikipedia大家都不陌生，它的下载地址是：<a href="http://dumps.wikimedia.org/" rel="nofollow" target="_blank">http://dumps.wikimedia.org/</a>&nbsp;, 这里有详细介绍：<a href="http://en.wikipedia.org/wiki/Wikipedia:Database_download" rel="nofollow" target="_blank">http://en.wikipedia.org/wiki/Wikipedia:Database_download</a><br />但是wikipedia只是Wikimedia基金会的一个子<a title="项目" href="http://www.haogongju.net/tag/%E9%A1%B9%E7%9B%AE" target="_blank">项目</a>，wikimedia下面还有多个其他的重要项目，包括：<br />wiktionary &nbsp; &nbsp;一个<a title="语义化" href="http://www.haogongju.net/tag/%E8%AF%AD%E4%B9%89%E5%8C%96" target="_blank">语义化</a>的关联词典，形式上类似于word<a title="net" href="http://www.haogongju.net/tag/net" target="_blank">net</a><br />wikiquote &nbsp; &nbsp;<a title="收录" href="http://www.haogongju.net/tag/%E6%94%B6%E5%BD%95" target="_blank">收录</a>各种名人名言<br />Wikibooks &nbsp; &nbsp;免费的教科书和手册<br />Wikinews &nbsp; &nbsp;大量的新闻故事<br />Wikiversity &nbsp; &nbsp;免费的教育材料<br />Wikisource &nbsp; &nbsp;免费的文本内容<br />上述的这些内容，都可以通过<a href="http://dumps.wikimedia.org/" rel="nofollow" target="_blank">http://dumps.wikimedia.org/</a>&nbsp;下载到。<br />还有一些小型的wiki项目，比如：<br /><a href="http://simple.wikipedia.org/" rel="nofollow" target="_blank">http://simple.wikipedia.org</a>&nbsp;&nbsp; &nbsp;使用Basic English写的wiki，给儿童和初学者看<br /><a href="http://simple.wiktionary.org/" rel="nofollow" target="_blank">http://simple.wiktionary.org</a>&nbsp;&nbsp; &nbsp;使用Basic English写的wiktionary<br /><br />wikipedia的数据处理有很多方式，我比较推崇这两个：<br />jwpl: &nbsp; &nbsp;<a href="http://code.google.com/p/jwpl/" rel="nofollow" target="_blank">http://code.google.com/p/jwpl/</a><br />wikipedia-miner: &nbsp;&nbsp;<a href="http://wikipedia-miner.cms.waikato.ac.nz/wiki/" rel="nofollow" target="_blank">http://wikipedia-miner.cms.waikato.ac.nz/wiki/</a><br /><br />下面我介绍下另一个商业化的wiki<a title="网站" href="http://www.haogongju.net/tag/%E7%BD%91%E7%AB%99" target="_blank">网站</a>:<a href="http://www.wikia.com/" rel="nofollow" target="_blank">http://www.wikia.com</a>&nbsp;&nbsp;这个网站上<a title="用户" href="http://www.haogongju.net/tag/%E7%94%A8%E6%88%B7" target="_blank">用户</a>可以创建单独的维基网站，下面是排名前250位wikia网站：<br /><a href="http://wikis.wikia.com/wiki/List_of_Wikia_wikis" rel="nofollow" target="_blank">http://wikis.wikia.com/wiki/List_of_Wikia_wikis</a><br />wikia上的<a title="资源" href="http://www.haogongju.net/tag/%E8%B5%84%E6%BA%90" target="_blank">资源</a>也可供下载：<a href="http://community.wikia.com/wiki/Help:Database_download" rel="nofollow" target="_blank">http://community.wikia.com/wiki/Help:Database_download</a><br /><br />Freebase:<br />freebase是啥就不解释了，下面给出数据的下载地址：<br /><a href="http://wiki.freebase.com/wiki/Data_dumps" rel="nofollow" target="_blank">http://wiki.freebase.com/wiki/Data_dumps</a>&nbsp;&nbsp; &nbsp;freebase自身的数据<br /><a href="http://wiki.freebase.com/wiki/WEX" rel="nofollow" target="_blank">http://wiki.freebase.com/wiki/WEX</a>&nbsp;&nbsp; &nbsp;freebase从wikipedia中提取的数据<br /><br />YAGO2:<br /><a href="http://www.mpi-inf.mpg.de/yago-naga/yago/" rel="nofollow" target="_blank">http://www.mpi-inf.mpg.de/yago-naga/yago/</a><br /><br />dbpedia:<br /><a href="http://www.dbpedia.org/" rel="nofollow" target="_blank">http://www.dbpedia.org</a><br /><br />如果要找<a title="Link" href="http://www.haogongju.net/tag/Link" target="_blank">Link</a>edData，可以来这里：<a href="http://www.thedatahub.org/" rel="nofollow" target="_blank">http://www.thedatahub.org</a>&nbsp;&nbsp; &nbsp;这里收集了很多Linked Data<br /><a href="http://linkeddata.org/" rel="nofollow" target="_blank">http://linkeddata.org/</a>&nbsp;&nbsp; &nbsp;这里有一张图，给出了各种linkeddata的关系和影响力。<br /><br />如果要找各种网上的api，可以来这里：<a href="http://www.programmableweb.com/" rel="nofollow" target="_blank">http://www.programmableweb.com</a>&nbsp;&nbsp; &nbsp;<br />现在外国政府纷纷对外公开数据，下面是几个政府的开放数据集：<br /><a href="http://data.gov.au/" rel="nofollow" target="_blank">http://data.gov.au</a>&nbsp;&nbsp; &nbsp;澳大利亚<br /><a href="http://data.dc.gov/" rel="nofollow" target="_blank">http://data.dc.gov</a>&nbsp;&nbsp; &nbsp;美国哥伦比亚州的<br /><a href="http://www.data.gov/" rel="nofollow" target="_blank">http://www.data.gov</a>&nbsp;&nbsp; &nbsp;美国<br /><a href="http://data.gov.uk/" rel="nofollow" target="_blank">http://data.gov.uk</a>&nbsp;&nbsp; &nbsp;英国<br /><a href="http://databases.lapl.org/" rel="nofollow" target="_blank">http://databases.lapl.org/</a>&nbsp;&nbsp; &nbsp;洛杉矶地区的开放数据集，知道硅谷为啥这么牛了吧<br /><a href="http://www.gov.hk/en/theme/psi/welcome" rel="nofollow" target="_blank">http://www.gov.hk/en/theme/psi/welcome</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp;香港政府也公开了很多数据<br />对比一下，外国政府做了这么多实事，人民大会堂里的那些酒囊饭袋们都在干什么？<br /><br /><a href="http://lexsrv3.nlm.nih.gov/LexSysGroup/Projects/lexAccess/current/web/download.html" rel="nofollow" target="_blank">http://lexsrv3.nlm.nih.gov/LexSysGroup/Projects/lexAccess/current/web/download.html</a>&nbsp;&nbsp; &nbsp;美国国家卫生署发布的词表<br /><a href="http://www.census.gov/genealogy/www/data/2000surnames/index.html" rel="nofollow" target="_blank">http://www.census.gov/genealogy/www/data/2000surnames/index.html</a>&nbsp;&nbsp; &nbsp;美国统计局的姓名数据<br /><a href="https://www.cia.gov/library/publications/download/" rel="nofollow" target="_blank">https://www.cia.gov/library/publications/download/</a>&nbsp;&nbsp; &nbsp;美国中央情报局发布的factbook，介绍了世界各国情况<br />连卫生署，统计局和中情局这种单位都为美国的信息建设做出了这么多的贡献，我们应该知道自己跟美帝的差距有多大了吧。<br /><br />叙词表：<br /><a href="http://www.nlm.nih.gov/mesh/filelist.html" rel="nofollow" target="_blank">http://www.nlm.nih.gov/mesh/filelist.html</a>&nbsp;&nbsp; &nbsp;mesh,关于医学的受控词表<br /><a href="http://id.loc.gov/download/" rel="nofollow" target="_blank">http://id.loc.gov/download/</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;美国国会图书馆发布的叙词表<br /><br />一些三元组数据：<br /><a href="http://www.cs.utexas.edu/users/pclark/dart/" rel="nofollow" target="_blank">http://www.cs.utexas.edu/users/pclark/dart/</a>&nbsp;&nbsp; &nbsp;<a title="采集" href="http://www.haogongju.net/tag/%E9%87%87%E9%9B%86" target="_blank">采集</a>自BNC（英国国家语料库）和Reuters，2300万条<br /><a href="http://reverb.cs.washington.edu/" rel="nofollow" target="_blank">http://reverb.cs.washington.edu/</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp;华盛顿大学的项目，1500万条<br /><a href="http://www.cs.washington.edu/research/sherlock-hornclauses/" rel="nofollow" target="_blank">http://www.cs.washington.edu/research/sherlock-hornclauses/</a>&nbsp;&nbsp; &nbsp;大约有200-300万条数据<br /><a href="http://www.cs.rochester.edu/research/knext" rel="nofollow" target="_blank">http://www.cs.rochester.edu/research/knext</a>&nbsp;&nbsp; &nbsp;有535万条数据，来自BNC和布朗语料库<br /><a href="http://rtw.ml.cmu.edu/rtw/resources" rel="nofollow" target="_blank">http://rtw.ml.cmu.edu/rtw/resources</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp;readthe<a title="web" href="http://www.haogongju.net/tag/web" target="_blank">web</a>项目，数据量较小<br /><br />机读词典：<br /><a href="http://wordnet.princeton.edu/" rel="nofollow" target="_blank">http://wordnet.princeton.edu/</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;英语的wordnet<br /><a href="http://nlpwww.nict.go.jp/wn-ja/index.en.html" rel="nofollow" target="_blank">http://nlpwww.nict.go.jp/wn-ja/index.en.html</a>&nbsp;&nbsp; &nbsp;日语的wordnet<br /><a href="http://alpage.inria.fr/~sagot/wolf-en.html" rel="nofollow" target="_blank">http://alpage.inria.fr/~sagot/wolf-en.html</a>&nbsp;&nbsp; &nbsp;法语的wordnet<br /><a href="http://wordnet.ru/" rel="nofollow" target="_blank">http://wordnet.ru/</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;俄罗斯的wordnet<br /><a href="http://cl.haifa.ac.il/projects/mwn/index.shtml" rel="nofollow" target="_blank">http://cl.haifa.ac.il/projects/mwn/index.shtml</a>&nbsp;&nbsp; &nbsp;希伯来语的wordnet<br /><a href="http://wordnet.dk/dannet/menu?item=2" rel="nofollow" target="_blank">http://wordnet.dk/dannet/menu?item=2</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp;丹麦语的wordnet<br /><a href="http://grial.uab.es/sensem/download?idioma=en" rel="nofollow" target="_blank">http://grial.uab.es/sensem/download?idioma=en</a>&nbsp;&nbsp; &nbsp;西班牙语的wordnet<br /><a href="http://www.ling.helsinki.fi/en/lt/research/finnwordnet/download.shtml" rel="nofollow" target="_blank">http://www.ling.helsinki.fi/en/lt/research/finnwordnet/download.shtml</a>&nbsp;&nbsp; &nbsp;芬兰语的wordnet<br />这些不同版本的wordnet都是免费下载的。可恨中国泱泱五千年的文明古国，文献典故浩如烟海，竟连一份免费且公开的机读词典都没有。这是汉语的耻辱，中国的耻辱，也是中华民族的耻辱。特别是中科院计算所和自动化所的人们，你们觉得呢？（顺祝hownet生意兴隆，越卖越好）<br /><br /><a href="http://dico.fj.free.fr/dico.php" rel="nofollow" target="_blank">http://dico.fj.free.fr/dico.php</a>&nbsp;&nbsp; &nbsp; &nbsp; &nbsp;日法词典<br /><a href="http://www.csse.monash.edu.au/~jwb/edict.html" rel="nofollow" target="_blank">http://www.csse.monash.edu.au/~jwb/edict.html</a>&nbsp;&nbsp; &nbsp;日英词典<br /><a href="http://cc-cedict.org/wiki/start" rel="nofollow" target="_blank">http://cc-cedict.org/wiki/start</a>&nbsp;&nbsp; &nbsp; 中文到英文的词典，终于出来中文的了，可惜是外国人搞出来的。<br /><a href="https://framenet.icsi.berkeley.edu/" rel="nofollow" target="_blank">https://framenet.icsi.berkeley.edu</a>&nbsp;&nbsp; &nbsp;基于框架语义学的东东，恐怕不能算词典，不过没地儿放了。<br /><br />语料库：<br /><a href="http://opus.lingfil.uu.se/" rel="nofollow" target="_blank">http://opus.lingfil.uu.se/</a>&nbsp;&nbsp; &nbsp;开放的平行语料库<br /><a href="http://opus.lingfil.uu.se/OpenSubtitles_v2.php" rel="nofollow" target="_blank">http://opus.lingfil.uu.se/OpenSubtitles_v2.php</a>&nbsp;&nbsp; &nbsp;大量<a title="电影" href="http://www.haogongju.net/tag/%E7%94%B5%E5%BD%B1" target="_blank">电影</a>字幕的下载地址<br /><a href="http://www.statmt.org/europarl" rel="nofollow" target="_blank">http://www.statmt.org/europarl</a>&nbsp;&nbsp; &nbsp;欧洲议会的平行语料库<br /><a href="http://www.anc.org/OANC/" rel="nofollow" target="_blank">http://www.anc.org/OANC/</a>&nbsp;&nbsp; &nbsp;开放的美国国家语料库<br /><br /><a href="http://snap.stanford.edu/data/" rel="nofollow" target="_blank">http://snap.stanford.edu/data/</a>&nbsp;&nbsp; &nbsp;斯坦福大学的SNAP项目，抓了很多数据，不过时间较早，只有研究价值</p><img src ="http://www.cppblog.com/sunrise/aggbug/191408.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-09-20 17:29 <a href="http://www.cppblog.com/sunrise/archive/2012/09/20/191408.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item><item><title>搭建wiki镜像小结</title><link>http://www.cppblog.com/sunrise/archive/2012/08/21/187819.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Tue, 21 Aug 2012 01:22:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/08/21/187819.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/187819.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/08/21/187819.html#Feedback</comments><slash:comments>0</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/187819.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/187819.html</trackback:ping><description><![CDATA[<p style="margin-bottom: 0cm">&nbsp; &nbsp; &nbsp;一.Apache，Php5，Mysql不可少，然后下载mediawiki软件。   </p> <p style="margin-bottom: 0cm">     之前没有接触过这些软件，so每一个都需要装....</p> <p style="margin-bottom: 0cm">     (1)apache配置</p> <p style="margin-bottom: 0cm">     　在Debian下， 安装完成后， 软件包为我们提供的配置文件位于/etc/apache2目录下：</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　tony@tonybox:/etc/apache2$ ls -l</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　total 72</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　-rw-r--r-- 1 root root 12482 2006-01-16 18:15 apache2.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　drwxr-xr-x 2 root root 4096 2006-06-30 13:56 conf.d</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　-rw-r--r-- 1 root root 748 2006-01-16 18:05 envvars</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　-rw-r--r-- 1 root root 268 2006-06-30 13:56 httpd.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　-rw-r--r-- 1 root root 12441 2006-01-16 18:15 magic</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　drwxr-xr-x 2 root root 4096 2006-06-30 13:56 mods-available</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　drwxr-xr-x 2 root root 4096 2006-06-30 13:56 mods-enabled</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　-rw-r--r-- 1 root root 10 2006-06-30 13:56 ports.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　-rw-r--r-- 1 root root 2266 2006-01-16 18:15 README</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　drwxr-xr-x 2 root root 4096 2006-06-30 13:56 sites-available</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　drwxr-xr-x 2 root root 4096 2006-06-30 13:56 sites-enabled</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　drwxr-xr-x 2 root root 4096 2006-01-16 18:15&nbsp;</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　其中</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　apache2.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　为apache2服务器的主配置文件， 查看此配置文件， 你会发现以下内容</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　# Include module configuration:</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　Include /etc/apache2/mods-enabled/*.load</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　Include /etc/apache2/mods-enabled/*.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　# Include all the user configurations:</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　Include /etc/apache2/httpd.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　# Include ports listing</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　Include /etc/apache2/ports.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　# Include generic snippets of statements</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　Include /etc/apache2/conf.d/[^.#]*</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　有此可见， apache2 根据配置功能的不同， 对配置文件进行了分割， 这样更利于管理</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　conf.d</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　下为配置文件的附加片断，默认情况下， 仅提供了 charset 片断，</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　tony@tonybox:/etc/apache2/conf.d$ cat charset</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　AddDefaultCharset UTF-8</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　如有需要我们可以将默认编码修改为 GB2312, 即文件的内容为： AddDefaultCharset GB2312</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　httpd.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　是个空文件</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　magic</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　文件中包含的是有关mod_mime_magic模块的数据， 一般不需要修改它。</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　ports.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　则为服务器监听IP和端口设置的配置文件，</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　tony@tonybox:/etc/apache2$ cat ports.conf</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　Listen 80</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　mods-available</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　目录下是一些。conf和。load 文件， 为系统中可以使用的加载各种模块的配置文件， 而mods-enabled目录下则是指向这些配置文件的符号连接， 从配置文件apache2.conf 中可以看出， 系统通过mods-enabled目录来加载模块， 也就是说， 系统仅通过在此目录下创建了符号连接的mods-available 目录下的配置文件来加载模块。同时系统还提供了两个命令 a2enmod 和 a2dismod用于维护这些符号连接。这两个命令由 apache2-common 包提供。命令各式也非常简单： a2enmod [module] 或 a2dismod [module]</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　sites-available</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　目录下为配置好的站点的配置文件， sites-enabled 目录下则是指向这些配置文件的符号连接， 系统通过这些符号连接来起用站点 sites-enabled目录下的符号连接附有一个数字前缀， 如000-default, 这个数字用于决定启动顺序， 数字越小， 启动优先级越高。 系统提供了两个命令 a2ensite 和 a2dissite 用于维护这些符号连接。这两个命令由 apache2-common 包提供。</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　/var/<u><a href="http://baike.baidu.com/view/1453.htm" target="_blank">www</a></u></p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　默认情况下将要发布的网页文件应该置于/var/www目录下，这一默认值可以同过主配置文件中的DocumnetRoot 选项修改。</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; "> 　　二.mediawiki直接解压到apache里面(就是解压在var/www路径下),解压后重名为wiki；</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">        三. 然后进主页localhost/wiki，对MediaWiki进行安装。去创建数据库wikidb。里面有41个表。在导入数据之间，要先清除page,revision,text三个表。</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">               delete from page;&nbsp;</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">               delete from revision;&nbsp;</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">               delete from text;&nbsp;</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">         四.<a href="http://dumps.wikimedia.org/backup-index.html在这里可以下载任何语言wiki的数据库xml文件">http://dumps.wikimedia.org/backup-index.html</a><a href="http://dumps.wikimedia.org/backup-index.html在这里可以下载任何语言wiki的数据库xml文件">在这里可以下载任何语言wiki</a><a href="http://dumps.wikimedia.org/backup-index.html在这里可以下载任何语言wiki的数据库xml文件">的数据库xml</a><a href="http://dumps.wikimedia.org/backup-index.html在这里可以下载任何语言wiki的数据库xml文件">文件</a>。下载的文件类似于enwiki-20061130-pages-articles.xml.bz2（英文版的），wiki差不多每两个月更新一次数据。</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">         五.安装mediawiki。去下载mediawiki的源代码，如果其官方网站被封的话可以去www.allwiki.com这个中文网站上去下载。下载后解压到你的apache能找到的一个目录下，将其config目录权限设置为777，然后在浏览器里访问其 config/index.php，进行一些配置后，会在config目录下生成一个LocalSettings.php的文件，将这个文件拷贝到它的上一级目录。最后别忘了将config的目录再改回原来的权限。</p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">          六.把文件导入数据库：&nbsp;<br />          命令：&nbsp;<br />                java -Xmx600M -server -jar mwdumper.jar --format=sql:1.5&nbsp;<br />                enwiki-20061130-pages-articles.xml.bz2 | mysql -u wikiuser -p wikidb&nbsp;<br /><br /></p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">       参见：<a href="http://fuhao-987.iteye.com/blog/1044933">http://fuhao-987.iteye.com/blog/1044933</a></p> <p align="LEFT" style="margin-bottom: 0cm; line-height: 0.64cm; ">                  <a href="http://jgs80.blog.163.com/blog/static/3566265320076177435762/">http://jgs80.blog.163.com/blog/static/3566265320076177435762/</a></p><img src ="http://www.cppblog.com/sunrise/aggbug/187819.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-08-21 09:22 <a href="http://www.cppblog.com/sunrise/archive/2012/08/21/187819.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item><item><title>Penn Treebank Tags</title><link>http://www.cppblog.com/sunrise/archive/2012/07/31/185743.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Tue, 31 Jul 2012 05:31:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/07/31/185743.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/185743.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/07/31/185743.html#Feedback</comments><slash:comments>0</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/185743.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/185743.html</trackback:ping><description><![CDATA[<p style="margin-top: 0px; margin-right: 0px; margin-left: 0px; padding: 0px; border: 0px; list-style: none; word-wrap: normal; word-break: normal; color: #323e32; font-family: simsun; background-color: #9caec1; "></p><p>&nbsp;</p><p>Note: This information comes from "Bracketing Guidelines for Treebank II Style Penn Treebank Project" - part of the documentation that comes with the Penn Treebank.</p>  <p>&nbsp;</p>  <p>Contents:</p>  <p>Bracket Labels</p>  <p>Clause Level</p>  <p>Phrase Level</p>  <p>Word Level</p>  <p>Function Tags</p>  <p>Form/function discrepancies</p>  <p>Grammatical role</p>  <p>Adverbials</p>  <p>Miscellaneous</p>  <p>Index of All Tags</p>  <p>Bracket Labels</p>  <p>Clause Level</p>  <p>S - simple declarative clause, i.e. one that is not introduced by a (possible empty) subordinating conjunction or a wh-word and that does not exhibit subject-verb inversion.</p>  <p>SBAR - Clause introduced by a (possibly empty) subordinating conjunction.</p>  <p>SBARQ - Direct question introduced by a wh-word or a wh-phrase. Indirect questions and relative clauses should be bracketed as SBAR, not SBARQ.</p>  <p>SINV - Inverted declarative sentence, i.e. one in which the subject follows the tensed verb or modal.</p>  <p>SQ - Inverted yes/no question, or main clause of a wh-question, following the wh-phrase in SBARQ.</p>  <p>&nbsp;</p>  <p>Phrase Level</p>  <p>ADJP - Adjective Phrase.</p>  <p>ADVP - Adverb Phrase.</p>  <p>CONJP - Conjunction Phrase.</p>  <p>FRAG - Fragment.</p>  <p>INTJ - Interjection. Corresponds approximately to the part-of-speech tag UH.</p>  <p>LST - List marker. Includes surrounding punctuation.</p>  <p>NAC - Not a Constituent; used to show the scope of certain prenominal modifiers within an NP.</p>  <p>NP - Noun Phrase.</p>  <p>NX - Used within certain complex NPs to mark the head of the NP. Corresponds very roughly to N-bar level but used quite differently.</p>  <p>PP - Prepositional Phrase.</p>  <p>PRN - Parenthetical.</p>  <p>PRT - Particle. Category for words that should be tagged RP.</p>  <p>QP - Quantifier Phrase (i.e. complex measure/amount phrase); used within NP.</p>  <p>RRC - Reduced Relative Clause.</p>  <p>UCP - Unlike Coordinated Phrase.</p>  <p>VP - Vereb Phrase.</p>  <p>WHADJP - Wh-adjective Phrase. Adjectival phrase containing a wh-adverb, as in how hot.</p>  <p>WHAVP - Wh-adverb Phrase. Introduces a clause with an NP gap. May be null (containing the 0 complementizer) or lexical, containing a wh-adverb such as how or why.</p>  <p>WHNP - Wh-noun Phrase. Introduces a clause with an NP gap. May be null (containing the 0 complementizer) or lexical, containing some wh-word, e.g. who, which book, whose daughter, none of which, or how many leopards.</p>  <p>WHPP - Wh-prepositional Phrase. Prepositional phrase containing a wh-noun phrase (such as of which or by whose authority) that either introduces a PP gap or is contained by a WHNP.</p>  <p>X - Unknown, uncertain, or unbracketable. X is often used for bracketing typos and in bracketing the...the-constructions.</p>  <p>Word level</p>  <p>CC - Coordinating conjunction</p>  <p>CD - Cardinal number</p>  <p>DT - Determiner</p>  <p>EX - Existential there</p>  <p>FW - Foreign word</p>  <p>IN - Preposition or subordinating conjunction</p>  <p>JJ - Adjective</p>  <p>JJR - Adjective, comparative</p>  <p>JJS - Adjective, superlative</p>  <p>LS - List item marker</p>  <p>MD - Modal</p>  <p>NN - Noun, singular or mass</p>  <p>NNS - Noun, plural</p>  <p>NNP - Proper noun, singular</p>  <p>NNPS - Proper noun, plural</p>  <p>PDT - Predeterminer</p>  <p>POS - Possessive ending</p>  <p>PRP - Personal pronoun</p>  <p>PRP$ - Possessive pronoun (prolog version PRP-S)</p>  <p>RB - Adverb</p>  <p>RBR - Adverb, comparative</p>  <p>RBS - Adverb, superlative</p>  <p>RP - Particle</p>  <p>SYM - Symbol</p>  <p>TO - to</p>  <p>UH - Interjection</p>  <p>VB - Verb, base form</p>  <p>VBD - Verb, past tense</p>  <p>VBG - Verb, gerund or present participle</p>  <p>VBN - Verb, past participle</p>  <p>VBP - Verb, non-3rd person singular present</p>  <p>VBZ - Verb, 3rd person singular present</p>  <p>WDT - Wh-determiner</p>  <p>WP - Wh-pronoun</p>  <p>WP$ - Possessive wh-pronoun (prolog version WP-S)</p>  <p>WRB - Wh-adverb</p>  <p>Function tags</p>  <p>Form/function discrepancies</p>  <p>-ADV (adverbial) - marks a constituent other than ADVP or PP when it is used adverbially (e.g. NPs or free ("headless" relatives). However, constituents that themselves are modifying an ADVP generally do not get -ADV. If a more specific tag is available (for example, -TMP) then it is used alone and -ADV is implied. See the Adverbials section.</p>  <p>-NOM (nominal) - marks free ("headless") relatives and gerunds when they act nominally.</p>  <p>Grammatical role</p>  <p>-DTV (dative) - marks the dative object in the unshifted form of the double object construction. If the preposition introducing the "dative" object is for, it is considered benefactive (-BNF). -DTV (and -BNF) is only used after verbs that can undergo dative shift.</p>  <p>-LGS (logical subject) - is used to mark the logical subject in passives. It attaches to the NP object of by and not to the PP node itself.</p>  <p>-PRD (predicate) - marks any predicate that is not VP. In the do so construction, the so is annotated as a predicate.</p>  <p>-PUT - marks the locative complement of put.</p>  <p>-SBJ (surface subject) - marks the structural surface subject of both matrix and embedded clauses, including those with null subjects.</p>  <p>-TPC ("topicalized") - marks elements that appear before the subject in a declarative sentence, but in two cases only:</p>  <p>if the front element is associated with a *T* in the position of the gap.</p>  <p>if the fronted element is left-dislocated (i.e. it is associated with a resumptive pronoun in the position of the gap).</p>  <p>-VOC (vocative) - marks nouns of address, regardless of their position in the sentence. It is not coindexed to the subject and not get -TPC when it is sentence-initial.</p>  <p>Adverbials</p>  <p>Adverbials are generally VP adjuncts.</p>  <p>-BNF (benefactive) - marks the beneficiary of an action (attaches to NP or PP).</p>  <p>This tag is used only when (1) the verb can undergo dative shift and (2) the prepositional variant (with the same meaning) uses for. The prepositional objects of dative-shifting verbs with other prepositions than for (such as to or of) are annotated -DTV.</p>  <p>&nbsp;</p>  <p>-DIR (direction) - marks adverbials that answer the questions "from where?" and "to where?" It implies motion, which can be metaphorical as in "...rose 5 pts. to 57-1/2" or "increased 70% to 5.8 billion yen" -DIR is most often used with verbs of motion/transit and financial verbs.</p>  <p>-EXT (extent) - marks adverbial phrases that describe the spatial extent of an activity. -EXT was incorporated primarily for cases of movement in financial space, but is also used in analogous situations elsewhere. Obligatory complements do not receive -EXT. Words such as fully and completely are absolutes and do not receive -EXT.</p>  <p>-LOC (locative) - marks adverbials that indicate place/setting of the event. -LOC may also indicate metaphorical location. There is likely to be some varation in the use of -LOC due to differing annotator interpretations. In cases where the annotator is faced with a choice between -LOC or -TMP, the default is -LOC. In cases involving SBAR, SBAR should not receive -LOC. -LOC has some uses that are not adverbial, such as with place names that are adjoined to other NPs and NAC-LOC premodifiers of NPs. The special tag -PUT is used for the locative argument of put.</p>  <p>-MNR (manner) - marks adverbials that indicate manner, including instrument phrases.</p>  <p>-PRP (purpose or reason) - marks purpose or reason clauses and PPs.</p>  <p>-TMP (temporal) - marks temporal or aspectual adverbials that answer the questions when, how often, or how long. It has some uses that are not strictly adverbial, auch as with dates that modify other NPs at S- or VP-level. In cases of apposition involving SBAR, the SBAR should not be labeled -TMP. Only in "financialspeak," and only when the dominating PP is a PP-DIR, may temporal modifiers be put at PP object level. Note that -TMP is not used in possessive phrases.</p>  <p>&nbsp;</p>  <p>Miscellaneous</p>  <p>-CLR (closely related) - marks constituents that occupy some middle ground between arguments and adjunct of the verb phrase. These roughly correspond to "predication adjuncts", prepositional ditransitives, and some "phrasel verbs". Although constituents marked with -CLR are not strictly speaking complements, they are treated as complements whenever it makes a bracketing difference. The precise meaning of -CLR depends somewhat on the category of the phrase.</p>  <p>on S or SBAR - These categories are usually arguments, so the -CLR tag indicates that the clause is more adverbial than normal clausal arguments. The most common case is the infinitival semi-complement of use, but there are a variety of other cases.</p>  <p>on PP, ADVP, SBAR-PRP, etc - On categories that are ordinarily interpreted as (adjunct) adverbials, -CLR indicates a somewhat closer relationship to the verb. For example:</p>  <p>Prepositional Ditransitives</p>  <p>In order to ensure consistency, the Treebank recognizes only a limited class of verbs that take more than one complement (-DTV and -PUT and Small Clauses) Verbs that fall outside these classes (including most of the prepositional ditransitive verbs in class [D2]) are often associated with -CLR.</p>  <p>Phrasal verbs</p>  <p>Phrasal verbs are also annotated with -CLR or a combination of -PRT and PP-CLR. Words that are considered borderline between particle and adverb are often bracketed with ADVP-CLR.</p>  <p>Predication Adjuncts</p>  <p>Many of Quirk&amp;apos;s predication adjuncts are annotated with -CLR.</p>  <p>on NP - To the extent that -CLR is used on NPs, it indicates that the NP is part of some kind of "fixed phrase" or expression, such as take care of. Variation is more likely for NPs than for other uses of -CLR.</p>  <p>-CLF (cleft) - marks it-clefts ("true clefts") and may be added to the labels S, SINV, or SQ.</p>  <p>-HLN (headline) - marks headlines and datelines. Note that headlines and datelines always constitute a unit of text that is structurally independent from the following sentence.</p>  <p>-TTL (title) - is attached to the top node of a title when this title appears inside running text. -TTL implies -NOM. The internal structure of the title is bracketed as usual.</p>  <p>Index of All Tags</p>  <p>ADJP</p>  <p>-ADV</p>  <p>ADVP</p>  <p>-BNF</p>  <p>CC</p>  <p>CD</p>  <p>-CLF</p>  <p>-CLR</p>  <p>CONJP</p>  <p>-DIR</p>  <p>DT</p>  <p>-DTV</p>  <p>EX</p>  <p>-EXT</p>  <p>FRAG</p>  <p>FW</p>  <p>-HLN</p>  <p>IN</p>  <p>INTJ</p>  <p>JJ</p>  <p>JJR</p>  <p>JJS</p>  <p>-LGS</p>  <p>-LOC</p>  <p>LS</p>  <p>LST</p>  <p>MD</p>  <p>-MNR</p>  <p>NAC</p>  <p>NN</p>  <p>NNS</p>  <p>NNP</p>  <p>NNPS</p>  <p>-NOM</p>  <p>NP</p>  <p>NX</p>  <p>PDT</p>  <p>POS</p>  <p>PP</p>  <p>-PRD</p>  <p>PRN</p>  <p>PRP</p>  <p>-PRP</p>  <p>PRP$ or PRP-S</p>  <p>PRT</p>  <p>-PUT</p>  <p>QP</p>  <p>RB</p>  <p>RBR</p>  <p>RBS</p>  <p>RP</p>  <p>RRC</p>  <p>S</p>  <p>SBAR</p>  <p>SBARQ</p>  <p>-SBJ</p>  <p>SINV</p>  <p>SQ</p>  <p>SYM</p>  <p>-TMP</p>  <p>TO</p>  <p>-TPC</p>  <p>-TTL</p>  <p>UCP</p>  <p>UH</p>  <p>VB</p>  <p>VBD</p>  <p>VBG</p>  <p>VBN</p>  <p>VBP</p>  <p>VBZ</p>  <p>-VOC</p>  <p>VP</p>  <p>WDT</p>  <p>WHADJP</p>  <p>WHADVP</p>  <p>WHNP</p>  <p>WHPP</p>  <p>WP</p>  <p>WP$ or WP-S</p>  <p>WRB</p>  <p>X</p><p>&nbsp;</p><p>&nbsp;</p><img src ="http://www.cppblog.com/sunrise/aggbug/185743.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-07-31 13:31 <a href="http://www.cppblog.com/sunrise/archive/2012/07/31/185743.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item><item><title>召回率和准确率</title><link>http://www.cppblog.com/sunrise/archive/2012/07/23/184693.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Mon, 23 Jul 2012 01:41:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/07/23/184693.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/184693.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/07/23/184693.html#Feedback</comments><slash:comments>0</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/184693.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/184693.html</trackback:ping><description><![CDATA[转自：<a href="http://uwei.blogbus.com/logs/11424864.html" target="_blank">http://uwei.blogbus.com/logs/11424864.html</a><br />外行人做互联网，很多概念不懂。就拿最基础的&#8220;召回率&#8221;和&#8220;准确率&#8221;这种概念，看看网上资料知道大概，自己用的时候，脑子里绕着弯儿能想明白，可碰到别人活用的时候，脑子里还是没法一下子反应过来，还是要绕弯想一下。特地找了些资料，将这两个概念整理一下，希望能更熟练。 
<p>召回率和准确率是搜索引擎（或其它检索系统）的设计中很重要的两个概念和指标。<br />召回率：Recall，又称&#8220;查全率&#8221;； <br />准确率：Precision，又称&#8220;精度&#8221;、&#8220;正确率&#8221;。<br />在一个大规模数据集合中检索文档时，可把集合中的所有文档分成四类：</p>
<table border="0" cellspacing="0" cellpadding="0" width="258" align="left" height="79">
<tbody>
<tr>
<td height="27" width="96">&nbsp;</td>
<td width="74">
<div align="center">相关</div></td>
<td width="74">
<div align="center">不相关</div></td></tr>
<tr>
<td height="25">
<div align="center">检索到</div></td>
<td bgcolor="#6633cc">
<div align="center"><strong>A</strong></div></td>
<td bgcolor="#ffff00">
<div align="center"><strong>B</strong></div></td></tr>
<tr>
<td height="23">
<div align="center">未检索到</div></td>
<td bgcolor="#66ff66">
<div align="center"><strong>C</strong></div></td>
<td bgcolor="#66ccff">
<div align="center"><strong>D</strong></div></td></tr></tbody></table>
<p>&nbsp;</p>
<p>&nbsp;</p>
<div align="left">
<p>&nbsp;</p>
<p>A：检索到的，相关的&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; （搜到的也想要的）<br />B：检索到的，但是不相关的&nbsp;&nbsp;&nbsp;&nbsp; &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;（搜到的但没用的）<br />C：未检索到的，但却是相关的&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; （没搜到，然而实际上想要的）<br />D：未检索到的，也不相关的&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; （没搜到也没用的）</p>
<p>通常我们希望：数据库中相关的文档，被检索到的越多越好，这是追求&#8220;查全率&#8221;，即A/(A+C)，越大越好。<br />同时我们还希望：检索到的文档中，相关的越多越好，不相关的越少越好，这是追求&#8220;准确率&#8221;，即A/(A+B)，越大越好。<br />&nbsp; <br />归纳如下：<br />召回率：检索到的相关文档 比 库中所有的相关文档<br />准确率：检索到的相关文档 比 所有被检索到的文档<br />&nbsp; <br />&#8220;召回率&#8221;与&#8220;准确率&#8221;虽然没有必然的关系（从上面公式中可以看到），然而在大规模数据集合中，这两个指标却是相互制约的。<br />由于&#8220;检索策略&#8221;并不完美，希望更多相关的文档被检索到时，放宽&#8220;检索策略&#8221;时，往往也会伴随出现一些不相关的结果，从而使准确率受到影响。<br />而希望去除检索结果中的不相关文档时，务必要将&#8220;检索策略&#8221;定的更加严格，这样也会使有一些相关的文档不再能被检索到，从而使召回率受到影响。</p></div>
<p>凡是设计到大规模数据集合的检索和选取，都涉及到&#8220;召回率&#8221;和&#8220;准确率&#8221;这两个指标。而由于两个指标相互制约，我们通常也会根据需要为&#8220;检索策略&#8221;选择一个合适的度，不能太严格也不能太松，寻求在召回率和准确率中间的一个平衡点。这个平衡点由具体需求决定。</p>
<p>其实，准确率（precision，精度）比较好理解。往往难以迅速反应的是&#8220;召回率&#8221;。我想这与字面意思也有关系，从&#8220;召回&#8221;的字面意思不能直接看到其意义。<br />我觉得&#8220;召回率&#8221;这个词翻译的不够好。&#8220;召回&#8221;在中文的意思是：把xx调回来。比如sony电池有问题，厂家召回。<br />既然说翻译的不好，我们回头看&#8220;召回率&#8221;对应的英文&#8220;recall&#8221;，recall除了有上面说到的&#8220;order sth to return&#8221;的意思之外，还有&#8220;remember&#8221;的意思。</p>
<p>Recall：the ability to remember sth. that you have learned or sth. that has happened in the past.</p>
<p>这里，recall应该是这个意思，这样就更容易理解&#8220;召回率&#8221;的意思了。<br />当我们问检索系统某一件事的所有细节时（输入检索query），Recall就是指：检索系统能&#8220;回忆&#8221;起那些事的多少细节，通俗来讲就是&#8220;回忆的能力&#8221;。能回忆起来的细节数 除以 系统知道这件事的所有细节，就是&#8220;记忆率&#8221;，也就是recall&#8212;&#8212;召回率。<br />&nbsp; <br />这样想，要容易的多了。</p><img src ="http://www.cppblog.com/sunrise/aggbug/184693.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-07-23 09:41 <a href="http://www.cppblog.com/sunrise/archive/2012/07/23/184693.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item><item><title>《数学之美》－－马尔可夫链</title><link>http://www.cppblog.com/sunrise/archive/2012/07/12/182950.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Thu, 12 Jul 2012 00:54:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/07/12/182950.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/182950.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/07/12/182950.html#Feedback</comments><slash:comments>4</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/182950.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/182950.html</trackback:ping><description><![CDATA[&nbsp; &nbsp; 最近在读的一本书《数学之美》，由于自己对马尔可夫链缺乏相关的知识背景，故学习了一下。对于N久没有看过概率论的人来说，重拾起来也花费了一点时间。比如：P(A|B)是指在B的条件下A的概率，诸如此类，都需要重新复习一下，正所谓温故而知新。知道这个了，也就不难理解马尔可夫链的性质，即<span style="font-family: sans-serif; ">每一步可以移动到任何一个相邻的点，在这里移动到每一个点的概率都是相同的。<br /></span><p>&nbsp;&nbsp; &nbsp;关于马尔可夫链的定义：&nbsp;<a href="http://zh.wikipedia.org/wiki/%E9%A6%AC%E5%8F%AF%E5%A4%AB%E9%8F%88">http://zh.wikipedia.org/wiki/%E9%A6%AC%E5%8F%AF%E5%A4%AB%E9%8F%88<br />&nbsp;</a>&nbsp; &nbsp;&nbsp;<br />&nbsp; &nbsp; 隐含马尔可夫模型是上述马尔可夫链的一个扩展：任何一个时刻t的状态St是不可见的。隐含马尔可夫模型在每一个时刻t会输出一个符号，而且这个符合和st相关，而且仅和st相关，这个被称为独立输出假设。关于隐含马尔可夫模型的成功应用可以参见吴军的《数学之美》第5章的内容。<br />&nbsp; &nbsp;额，快到上班时间了，小总结一下。继续码农中......</p><img src ="http://www.cppblog.com/sunrise/aggbug/182950.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-07-12 08:54 <a href="http://www.cppblog.com/sunrise/archive/2012/07/12/182950.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item><item><title>统计自然语言处理--互信息</title><link>http://www.cppblog.com/sunrise/archive/2012/06/01/177044.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Fri, 01 Jun 2012 05:06:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/06/01/177044.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/177044.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/06/01/177044.html#Feedback</comments><slash:comments>2</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/177044.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/177044.html</trackback:ping><description><![CDATA[<div style="layout-grid:  15.6pt none" class="Section0">
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">&nbsp;&nbsp;&nbsp;今天六一，<font face="Times New Roman">C</font><font face="宋体">小加不在身边，混球啊。任务需要在看曼宁的《统计自然语言处理基础》。然后用到互信息，每次我觉得好高深的名字，做下去的时候就发现没有那么难。</font></span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">搭配</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">搭配由有限的复合构词法所描述。</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">识别搭配对的方法有三种：<font face="Times New Roman">1.</font><font face="宋体">使用频率信息的搭配识别。</font><font face="Times New Roman">2.</font><font face="宋体">基于含义和主词搭配词之间的距离识别。</font><font face="Times New Roman">3.</font><font face="宋体">基于假设测试和互信息的识别。</font></span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">1.</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">频率</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">将语料过滤后得到的动词，名词，之间进行两两配对，统计每个词语在一个句子，或在一个段落中出现的次数，即为频率。</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">2.</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">均值和方差&nbsp;</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">&nbsp;&nbsp;&nbsp;由于两个词之间的距离是可以变化的，计算两个词之间的偏移量的均值和方差。</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">均值就是简单的平均偏移量。</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">方差衡量的是单独的偏移量偏离均值的距离：</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span></span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><img border="0" alt="" src="http://www.cppblog.com/images/cppblog_com/sunrise/QQ截图20120601103629.png" width="232" height="140" /><br />&nbsp;</span><span><img alt="" src="file:///C:/DOCUME~1/ADMINI~1/LOCALS~1/Temp/ksohtml/wps_clip_image-31877.png" width="18" height="25" /></span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">是同现<font face="Times New Roman">i</font><font face="宋体">的偏移量，</font></span><span><img alt="" src="file:///C:/DOCUME~1/ADMINI~1/LOCALS~1/Temp/ksohtml/wps_clip_image-14102.png" width="15" height="30" /></span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">表示的是样本偏移量的均值。&nbsp;</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;我们可以通过使用这个信息来发现搭配。具体的方法是通过寻找带有低偏差的词对。一个低的偏差值意味着这两个词通常大致相同距离出现。零偏差意味着这两个词总是以相同的距离出现。</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">&nbsp;&nbsp;&nbsp;方差是关于一个相对于其他词分布峰值情况的度量。</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">关于互信息</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: 'Times New Roman'; font-size: 10.5pt; mso-spacerun: 'yes'">互信息的计算公式是这样的：</span><span style="font-family: 'Times New Roman'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: 'Times New Roman'; font-size: 10.5pt; mso-spacerun: 'yes'">MI(a,b)&nbsp;=&nbsp;log(&nbsp;p(ab)&nbsp;/&nbsp;(p(a)*p(b))&nbsp;)</span><span style="font-family: 'Times New Roman'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: 'Times New Roman'; font-size: 10.5pt; mso-spacerun: 'yes'">其中<font face="Times New Roman">log</font><font face="宋体">的底数是</font><font face="Times New Roman">2</font><font face="宋体">，</font><font face="Times New Roman">p(x)</font><font face="宋体">表示</font><font face="Times New Roman">x</font><font face="宋体">出现的概率。</font></span><span style="font-family: 'Times New Roman'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p>
<p style="margin-top: 0pt; margin-bottom: 0pt" class="p0"><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'">好吧，好水，好简单。。着手写代码了。</span><span style="font-family: '宋体'; font-size: 10.5pt; mso-spacerun: 'yes'"><o:p></o:p></span></p></div><!--EndFragment-->  <img src ="http://www.cppblog.com/sunrise/aggbug/177044.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-06-01 13:06 <a href="http://www.cppblog.com/sunrise/archive/2012/06/01/177044.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item><item><title>基于HowNet语义相似度的ＦＡＱ的研究</title><link>http://www.cppblog.com/sunrise/archive/2012/04/12/171089.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Thu, 12 Apr 2012 08:19:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/04/12/171089.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/171089.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/04/12/171089.html#Feedback</comments><slash:comments>1</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/171089.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/171089.html</trackback:ping><description><![CDATA[<p>&nbsp;&nbsp;&nbsp; 继建立同义词库后的新任务，读文献，然后找出问题的解决方案。几篇文献都是研究句子与句子的相似度计算，我们的关键是词语与句子的相似度计算。据说FAQ是自然语言处理领域研究的热点。看了几篇论文，感觉都是大同小异。<br />&nbsp;&nbsp; 因为是第一次接触这些东西，所以有很多陌生的词汇，就自己动手查了查。<br />&nbsp;&nbsp; 关于HowNet,见<a href="http://www.keenage.com/zhiwang/c_zhiwang.html">http://www.keenage.com/zhiwang/c_zhiwang.html</a><br />&nbsp;&nbsp; FAQ自动问答系统的核心问题是如何快速地将客户所提问题与FAQ数据库的问题比较，进而确定与其最相似的问题，如果有，则将对应的答案作为结果回复给客户。<br />&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; &nbsp; <img border="0" alt="" src="http://www.cppblog.com/images/cppblog_com/sunrise/QQ截图20120412143643.jpg" width="559" longdesc="FAQ系统结构" height="215" /><br /><br />&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;FAQ系统结构图<br />&nbsp;&nbsp;&nbsp;&nbsp; &nbsp;相似度流程的计算就是先计算义原相似度，然后是概念相似度，接着词语相似度，最后就是句子相似度。<br />&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<a href="/Files/sunrise/相似度.doc">/Files/sunrise/相似度.doc</a>这里公式不能显示就相似度计算就插在附件中了。<br />&nbsp;&nbsp;&nbsp;&nbsp; FAQ差不多就进行到这里了。程序小白的小白文章，小白将继续小白下去。</p>
<p>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<br />&nbsp; <br /></p> <img src ="http://www.cppblog.com/sunrise/aggbug/171089.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-04-12 16:19 <a href="http://www.cppblog.com/sunrise/archive/2012/04/12/171089.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item><item><title>自然语言处理相关书籍及其他资源</title><link>http://www.cppblog.com/sunrise/archive/2012/03/28/169243.html</link><dc:creator>SunRise_at</dc:creator><author>SunRise_at</author><pubDate>Wed, 28 Mar 2012 01:35:00 GMT</pubDate><guid>http://www.cppblog.com/sunrise/archive/2012/03/28/169243.html</guid><wfw:comment>http://www.cppblog.com/sunrise/comments/169243.html</wfw:comment><comments>http://www.cppblog.com/sunrise/archive/2012/03/28/169243.html#Feedback</comments><slash:comments>1</slash:comments><wfw:commentRss>http://www.cppblog.com/sunrise/comments/commentRss/169243.html</wfw:commentRss><trackback:ping>http://www.cppblog.com/sunrise/services/trackbacks/169243.html</trackback:ping><description><![CDATA[<p style="margin-top: 0px; margin-right: 0px; margin-bottom: 0px; margin-left: 0px; padding-top: 0px; padding-right: 0px; padding-bottom: 0px; padding-left: 0px; font-family: Arial; line-height: 26px; text-align: left; ">特别推荐：<br />1、<a href="http://ishare.iask.sina.com.cn/f/22444899.html" target="_blank" style="color: #ca0000; text-decoration: none; ">HMM学习最佳范例</a>全文文档<br />2、<a href="http://ishare.iask.sina.com.cn/f/22444904.html" target="_blank" style="color: #ca0000; text-decoration: none; ">无约束最优化</a>全文文档</p><p style="margin-top: 0px; margin-right: 0px; margin-bottom: 0px; margin-left: 0px; padding-top: 0px; padding-right: 0px; padding-bottom: 0px; padding-left: 0px; font-family: Arial; line-height: 26px; text-align: left; ">一、书籍：<br />1、<a href="http://ishare.iask.sina.com.cn/f/22444344.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《自然语言处理综论》英文版第二版</a><br />2、<a href="http://ishare.iask.sina.com.cn/f/22445062.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《统计自然语言处理基础》英文版</a><br />3、<a href="http://ishare.iask.sina.com.cn/f/22442315.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《用Python进行自然语言处理》，NLTK配套书</a><br />4、<a href="http://ishare.iask.sina.com.cn/f/22444405.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《Learning Python第三版》</a>，Python入门经典书籍，详细而不厌其烦<br />5、<a href="http://www.xun6.com/file/5b2fa5c16/Pattern+Recognition+in+Speech+and+Language+Processing.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《自然语言处理中的模式识别》</a><br />6、<a href="http://www.xun6.com/file/fe894b416/Mc+Lachlan%2C+Krishnan+-+the+EM+algorithm+and+extensions.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《EM算法及其扩展》</a><br />7、<a href="http://www.xun6.com/file/7d914ba16/The+Elements+of+Statistical+Learning.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《统计学习基础》</a><br />8、《<a href="http://www.xun6.com/file/49f979f18/Natural_Language_Understanding.chm.html" target="_blank" style="color: #ca0000; text-decoration: none; ">自然语言理解</a>》英文版（似乎只有前9章）<br />9、<a href="http://www.xun6.com/file/2fe8ed418/Fundamentals_of_Speech_Recognition.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">《Fundamentals of Speech Recognition》</a>，质量不太好，不过第6章关于HMM的部分比较详细，作者之一便是Lawrence Rabiner；<br />10、概率统计经典入门书：《概率论及其应用》（英文版，威廉*费勒著）<br />　　<a href="http://www.xun6.com/file/1d285e118/An+Introduction+to+Probability+Theory+and+Its+Applications%2C+Volume+1.djvu.html" target="_blank" style="color: #ca0000; text-decoration: none; ">第一卷</a>　　<a href="http://www.xun6.com/file/7237f1618/An+Introduction+to+Probability+Theory+and+Its+Applications%2C+Volume+2.djvu.html" target="_blank" style="color: #ca0000; text-decoration: none; ">第二卷</a>　　<a href="http://www.xun6.com/file/f034c6418/DjVuLibre%2BDjView-3.5.22%2B4.5-Setup.exe.html" target="_blank" style="color: #ca0000; text-decoration: none; ">DjVuLibre阅读器</a>（阅读前两卷书需要）<br />11、一本利用Perl和Prolog进行自然语言处理的介绍书籍：《<a href="http://www.xun6.com/file/0d80185e5/An+Introduction+to+Language+Processing+with+Perl+and+Prolog.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">An Introduction to Language Processing with Perl and Prolog</a>》<br />12、国外机器学习书籍之：<br />　1) &#8220;<a href="http://www.xun6.com/file/9e6fc1914/OReilly_Programming_Collective_Intelligence.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Programming Collective Intelligence</a>&#8220;，中文译名《集体智慧编程》，机器学习&amp;数据挖掘领域&#8221;近年出的入门好书，培养兴趣是最重要的一环，一上来看大部头很容易被吓走的&#8221;<br />　2) &#8220;<a href="http://www.xun6.com/file/8d6a6eb14/McGrawHill.Tom.Mitchell-Machine.Learning.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Machine Learning</a>&#8220;,机器学习领域无可争议的经典书籍，下载完毕将后缀改为pdf即可。豆瓣评论 by王宁）：老书，牛人。现在看来内容并不算深，很多章节有点到为止的感觉，但是很适合新手（当然，不能&#8221;新&#8221;到连算法和概率都不知道）入门。比如决策树部分就很精彩，并且这几年没有特别大的进展，所以并不过时。另外，这本书算是对97年前数十年机器学习工作的大综述，参考文献列表极有价值。国内有翻译和影印版，不知道绝版否。<br />　3) &#8220;<a href="http://www.xun6.com/file/7bf0464f5/Introduction+to+Machine+Learning.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Introduction to Machine Learning</a>&#8221;<br />13、国外数据挖掘书籍之：<br />　1) &#8220;<a href="http://www.xun6.com/file/da7274114/Data.Mining.Concepts.and.Techniques.2nd.Ed.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Data.Mining.Concepts.and.Techniques.2nd</a>&#8220;，数据挖掘经典书籍 作者 : Jiawei Han/Micheline Kamber 出版社 : Morgan Kaufmann 评语 : 华裔科学家写的书，相当深入浅出。<br />　2)&nbsp;<a href="http://www.xun6.com/file/b3fbba9f5/Data_Mining.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Data Mining:Practical Machine Learning Tools and Techniques</a><br />　3)&nbsp;<a href="http://www.xun6.com/file/cc9883d12/Beautiful+Data.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Beautiful Data: The Stories Behind Elegant Data Solutions</a>（ Toby Segaran, Jeff Hammerbacher）<br />14、国外模式识别书籍之：<br />　1）&#8220;<a href="http://www.xun6.com/file/52ea8a539/Pattern_Recognition.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Pattern Recognition</a>&#8221;<br />　2）&#8220;<a href="http://www.xun6.com/file/61bf4ca39/Pattern.Recognition.Technologies.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Pattern Recongnition Technologies and Applications</a>&#8221;<br />　3）&#8220;<a href="http://www.xun6.com/file/d73ad5716/An_Introduction_to_Pattern_Recognition_-_Michael_Alder.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">An Introduction to Pattern Recognition</a>&#8221;<br />　4）&#8220;<a href="http://www.xun6.com/file/ec9acbb16/Introduction_to_Statistical_Pattern_Recognition.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Introduction to Statistical Pattern Recognition</a>&#8221;<br />　5）&#8220;<a href="http://www.xun6.com/file/e59b45816/Statistical_Pattern_Recognition__2nd_Edition.zip.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Statistical Pattern Recognition 2nd Edition</a>&#8221;<br />　6）&#8220;<a href="http://www.xun6.com/file/d3bf01816/Supervised_and_Unsupervised_Pattern_Recognition.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Supervised and Unsupervised Pattern Recognition</a>&#8221;<br />　7）&#8220;<a href="http://www.xun6.com/file/847d54516/Support+Vector+Machines+for+Pattern+Classification.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Support Vector Machines for Pattern Classification</a>&#8221;<br />15、国外人工智能书籍之：<br />　1）<a href="http://www.xun6.com/file/a7b4eb814/Artificial+Intelligence+A+Modern+Approach-+Second+Edition.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Artificial Intelligence: A Modern Approach</a>&nbsp;(2nd Edition) 人工智能领域无争议的经典。<br />　2）&#8220;<a href="http://www.xun6.com/file/9d3b5ca16/Paradigms+of+Artificial+Intelligence+Programming+Case+Studies+in+Common+LISP.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Paradigms of Artificial Intelligence Programming: Case Studies in Common LISP</a>&#8221;<br />16、其他相关书籍：<br />　1）<a href="http://www.xun6.com/file/c8dcedf12/Programming+the+Semantic+Web.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Programming the Semantic Web</a>，Toby Segaran , Colin Evans, Jamie Taylor<br />　2）<a href="http://www.xun6.com/file/b43b36b12/Oreilly.Learning.Python.4th.Edition.Sep.2009.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Learning.Python第四版</a>，英文</p><p style="margin-top: 0px; margin-right: 0px; margin-bottom: 0px; margin-left: 0px; padding-top: 0px; padding-right: 0px; padding-bottom: 0px; padding-left: 0px; font-family: Arial; line-height: 26px; text-align: left; ">二、课件：<br />1、哈工大刘挺老师的&#8220;<a href="http://ishare.iask.sina.com.cn/f/22509389.html" target="_blank" style="color: #ca0000; text-decoration: none; ">统计自然语言处理</a>&#8221;课件；<br />2、哈工大刘秉权老师的&#8220;<a href="http://www.xun6.com/file/6b34e1a15/%E8%87%AA%E7%84%B6%E8%AF%AD%E8%A8%80%E5%A4%84%E7%90%86-%E5%93%88%E5%B7%A5%E5%A4%A7%E5%88%98%E7%A7%89%E6%9D%83.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">自然语言处理</a>&#8221;课件；<br />3、中科院计算所刘群老师的&#8220;<a href="http://www.xun6.com/file/ea552a315/%E8%AE%A1%E7%AE%97%E8%AF%AD%E8%A8%80%E5%AD%A6%E8%AE%B2%E4%B9%89-%E5%88%98%E7%BE%A4.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">计算语言学讲义</a>&#8220;课件；<br />4、中科院自动化所宗成庆老师的&#8220;<a href="http://ishare.iask.sina.com.cn/f/22509399.html" target="_blank" style="color: #ca0000; text-decoration: none; ">自然语言理解</a>&#8221;课件；<br />5、北大常宝宝老师的&#8220;<a href="http://ishare.iask.sina.com.cn/f/22509403.html" target="_blank" style="color: #ca0000; text-decoration: none; ">计算语言学</a>&#8221;课件；<br />6、北大詹卫东老师的&#8220;<a href="http://ishare.iask.sina.com.cn/f/22509596.html" target="_blank" style="color: #ca0000; text-decoration: none; ">中文信息处理基础</a>&#8221;的课件及相关代码；<br />7、MIT Regina Barzilay教授的&#8220;<a href="http://www.xun6.com/file/7a1d66015/MIT%E8%87%AA%E7%84%B6%E8%AF%AD%E8%A8%80%E5%A4%84%E7%90%86-Regina+Barzilay.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">自然语言处理</a>&#8221;课件，52nlp上翻译了前5章；<br />8、MIT大牛Michael Collins的&#8220;<a href="http://www.xun6.com/file/d15696e15/MIT-Machine-learning-for-nlp-Michael+Collins.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Machine Learning Approaches for Natural Language Processing</a>(面向自然语言处理的机器学习方法)&#8221;课件；<br />9、Michael Collins的&#8220;<a href="http://www.xun6.com/file/072cfd815/MIT-Machine-learning-Michael+Collins.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Machine Learning&nbsp;</a>（机器学习）&#8221;课件；<br />10、SMT牛人Philipp Koehn &#8220;<a href="http://www.xun6.com/file/e45969316/Advanced+Natural+Language+Processing-Philipp+Koehn.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Advanced Natural Language Processing</a>（高级自然语言处理）&#8221;课件；<br />11、Philipp Koehn &#8220;<a href="http://www.xun6.com/file/7dbe55316/Empirical+Methods+in+Natural+Language+Processing-Philipp+Koehn.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Empirical Methods in Natural Language Processing</a>&#8221;课件；<br />12、Philipp Koehn&#8220;<a href="http://www.xun6.com/file/6e5766d16/Machine+Translation-Philipp+Koehn.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">Machine Translation</a>（机器翻译）&#8221;课件；</p><p style="margin-top: 0px; margin-right: 0px; margin-bottom: 0px; margin-left: 0px; padding-top: 0px; padding-right: 0px; padding-bottom: 0px; padding-left: 0px; font-family: Arial; line-height: 26px; text-align: left; ">三、语言资源和开源工具：<br />1、Brown语料库：<br />　a)&nbsp;<a href="http://www.xun6.com/file/fe1271a16/brown_tei.zip.html" target="_blank" style="color: #ca0000; text-decoration: none; ">XML格式的brown语料库</a>，带词性标注；<br />　b)&nbsp;<a href="http://www.xun6.com/file/7b4ab4816/brown.zip.html" target="_blank" style="color: #ca0000; text-decoration: none; ">普通文本格式的brown语料库</a>，带词性标注；<br />　c) 合并并去除空行、行首空格，用于词性标注训练：<a href="http://www.xun6.com/file/cdf8c0b16/browntest.zip.html" target="_blank" style="color: #ca0000; text-decoration: none; ">browntest.zip</a><br />2、<a href="http://nltk.googlecode.com/svn/trunk/nltk_data/index.xml" target="_blank" style="color: #ca0000; text-decoration: none; ">NLTK官方提供的语料库资源列表</a><br />3、<a href="http://opennlp.sourceforge.net/projects.html" target="_blank" style="color: #ca0000; text-decoration: none; ">OpenNLP上的开源自然语言处理工具列表</a><br />4、斯坦福大学自然语言处理组维护的&#8220;<a href="http://nlp.stanford.edu/links/statnlp.html" target="_blank" style="color: #ca0000; text-decoration: none; ">统计自然语言处理及基于语料库的计算语言学资源列表</a>&#8221;<br />5、<a href="http://projects.ldc.upenn.edu/Chinese/" target="_blank" style="color: #ca0000; text-decoration: none; ">LDC上免费的中文信息处理资源</a><br />6、中文分词相关工具：<br />　1）Java版本的MMSEG：<a href="http://www.xun6.com/file/43e341b16/mmseg-v0.3.zip.html" target="_blank" style="color: #ca0000; text-decoration: none; ">mmseg-v0.3.zip</a>，作者为solol，详情可参见：《<a href="http://www.52nlp.cn/%E4%B8%AD%E6%96%87%E5%88%86%E8%AF%8D%E5%85%A5%E9%97%A8%E4%B9%8B%E7%AF%87%E5%A4%96" target="_blank" style="color: #ca0000; text-decoration: none; ">中文分词入门之篇外</a>》<br />　2）张华平老师的ICTCLAS2010，该版本非商用免费一年，下载地址：<br /><a href="http://cid-51de2738d3ea0fdd.skydrive.live.com/self.aspx/.Public/ICTCLAS2010-packet-release.rar" target="_blank" style="color: #ca0000; text-decoration: none; ">http://cid-51de2738d3ea0fdd.skydrive.live.com/self.aspx/.Public/ICTCLAS2010-packet-release.rar</a><br />7、热心读者&#8220;<a href="http://www.cnblogs.com/finallyliuyu/" target="_blank" style="color: #ca0000; text-decoration: none; ">finallyliuyu</a>&#8221;提供的一批新闻语料库，包括腾讯，新浪，网易，凤凰等，目前放在CSDN上：<a href="http://finallyliuyu.download.csdn.net/" target="_blank" style="color: #ca0000; text-decoration: none; ">http://finallyliuyu.download.csdn.net/</a><br />　　另外finalllyliuyu在2010年9月又提供了一批文本文类语料，详情见：<a href="http://www.cnblogs.com/finallyliuyu/archive/2010/09/11/1824091.html" target="_blank" style="color: #ca0000; text-decoration: none; ">献给热衷于自然语言处理的业余爱好者的中文新闻分类语料库之二</a></p><p style="margin-top: 0px; margin-right: 0px; margin-bottom: 0px; margin-left: 0px; padding-top: 0px; padding-right: 0px; padding-bottom: 0px; padding-left: 0px; font-family: Arial; line-height: 26px; text-align: left; ">四、文献：<br />1、ACL-IJCNLP 2009论文全集：<br />　a)&nbsp;<a href="http://www.xun6.com/file/e5150b618/ACLIJCNLP-2009-1.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">大会论文Full Paper第一卷</a><br />　b)&nbsp;<a href="http://www.xun6.com/file/0c783df18/ACLIJCNLP-2009-2.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">大会论文Full Paper第二卷</a><br />　c)&nbsp;<a href="http://www.xun6.com/file/53b33b618/Short-2009.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">大会论文Short Paper合集</a><br />　d)&nbsp;<a href="http://www.xun6.com/file/8e6e49f18/EMNLP-2009.pdf.html" target="_blank" style="color: #ca0000; text-decoration: none; ">ACL09之EMNLP-2009合集</a><br />　e)&nbsp;<a href="http://www.xun6.com/file/5f1aa7518/ACL09_others.rar.html" target="_blank" style="color: #ca0000; text-decoration: none; ">ACL09 所有workshop论文合集</a></p><img src ="http://www.cppblog.com/sunrise/aggbug/169243.html" width = "1" height = "1" /><br><br><div align=right><a style="text-decoration:none;" href="http://www.cppblog.com/sunrise/" target="_blank">SunRise_at</a> 2012-03-28 09:35 <a href="http://www.cppblog.com/sunrise/archive/2012/03/28/169243.html#Feedback" target="_blank" style="text-decoration:none;">发表评论</a></div>]]></description></item></channel></rss>