|   | 1 | [[PageOutline]] | 
                  
                          |   | 2 | = Nutch1.5 + Solr3.6.1 = | 
                  
                          |   | 3 |  | 
                  
                          |   | 4 | == 下載 == | 
                  
                          |   | 5 |  * [http://apache.stu.edu.tw/nutch/1.5/apache-nutch-1.5-bin.tar.gz Nutch1.5] | 
                  
                          |   | 6 |  * [http://ftp.twaren.net/Unix/Web/apache/lucene/solr/3.6.1/apache-solr-3.6.1.tgz Solr3.6.1] | 
                  
                          |   | 7 |  | 
                  
                          |   | 8 | == Steps == | 
                  
                          |   | 9 | === 0. 前置環境設定 === | 
                  
                          |   | 10 | ==== 安裝JAVA,確認環境變數 ==== | 
                  
                          |   | 11 | {{{ | 
                  
                          |   | 12 | $ vim ~/.bashrc | 
                  
                          |   | 13 | }}} | 
                  
                          |   | 14 | 加入下列參數(or其他版本的Java路徑) | 
                  
                          |   | 15 | {{{ | 
                  
                          |   | 16 | export JAVA_HOME=/usr/lib/jvm/java-6-sun/ | 
                  
                          |   | 17 | }}} | 
                  
                          |   | 18 |  | 
                  
                          |   | 19 | === 1. Nutch設定 === | 
                  
                          |   | 20 | ==== 解壓縮nutch安裝包 ==== | 
                  
                          |   | 21 | {{{ | 
                  
                          |   | 22 | $ tar zxvf apache-nutch-1.5-bin.tar.gz | 
                  
                          |   | 23 | }}} | 
                  
                          |   | 24 |  * 解壓縮的資料路徑,以下開始以_[$NUTCH_HOME]_表示 | 
                  
                          |   | 25 |  | 
                  
                          |   | 26 | ==== 確認是否可以執行 ==== | 
                  
                          |   | 27 |  * 執行以下指令 | 
                  
                          |   | 28 | {{{ | 
                  
                          |   | 29 | $ [$NUTCH_HOME]/bin/nutch | 
                  
                          |   | 30 | }}} | 
                  
                          |   | 31 |  | 
                  
                          |   | 32 |  * 執行結果 | 
                  
                          |   | 33 | {{{ | 
                  
                          |   | 34 | Usage: nutch COMMAND | 
                  
                          |   | 35 | where COMMAND is one of: | 
                  
                          |   | 36 |   crawl             one-step crawler for intranets | 
                  
                          |   | 37 |   readdb            read / dump crawl db | 
                  
                          |   | 38 |   mergedb           merge crawldb-s, with optional filtering | 
                  
                          |   | 39 |   readlinkdb        read / dump link db | 
                  
                          |   | 40 |   inject            inject new urls into the database | 
                  
                          |   | 41 |   generate          generate new segments to fetch from crawl db | 
                  
                          |   | 42 |   freegen           generate new segments to fetch from text files | 
                  
                          |   | 43 |   fetch             fetch a segment's pages | 
                  
                          |   | 44 |   parse             parse a segment's pages | 
                  
                          |   | 45 |   readseg           read / dump segment data | 
                  
                          |   | 46 |   mergesegs         merge several segments, with optional filtering and slicing | 
                  
                          |   | 47 |   updatedb          update crawl db from segments after fetching | 
                  
                          |   | 48 |   invertlinks       create a linkdb from parsed segments | 
                  
                          |   | 49 |   mergelinkdb       merge linkdb-s, with optional filtering | 
                  
                          |   | 50 |   solrindex         run the solr indexer on parsed segments and linkdb | 
                  
                          |   | 51 |   solrdedup         remove duplicates from solr | 
                  
                          |   | 52 |   solrclean         remove HTTP 301 and 404 documents from solr | 
                  
                          |   | 53 |   parsechecker      check the parser for a given url | 
                  
                          |   | 54 |   indexchecker      check the indexing filters for a given url | 
                  
                          |   | 55 |   domainstats       calculate domain statistics from crawldb | 
                  
                          |   | 56 |   webgraph          generate a web graph from existing segments | 
                  
                          |   | 57 |   linkrank          run a link analysis program on the generated web graph | 
                  
                          |   | 58 |   scoreupdater      updates the crawldb with linkrank scores | 
                  
                          |   | 59 |   nodedumper        dumps the web graph's node scores | 
                  
                          |   | 60 |   plugin            load a plugin and run one of its classes main() | 
                  
                          |   | 61 |   junit             runs the given JUnit test | 
                  
                          |   | 62 |  or | 
                  
                          |   | 63 |   CLASSNAME         run the class named CLASSNAME | 
                  
                          |   | 64 | Most commands print help when invoked w/o parameters. | 
                  
                          |   | 65 | }}} | 
                  
                          |   | 66 |  | 
                  
                          |   | 67 |  * 若出現以上片段,則執行環境OK! | 
                  
                          |   | 68 |  | 
                  
                          |   | 69 | ==== 設定爬取機器人名稱 ==== | 
                  
                          |   | 70 | {{{ | 
                  
                          |   | 71 | $ vim [$NUTCH_HOME]/conf/nutch-site.xml | 
                  
                          |   | 72 | }}} | 
                  
                          |   | 73 |  * 加入以下資訊: | 
                  
                          |   | 74 | {{{ | 
                  
                          |   | 75 | #!text | 
                  
                          |   | 76 | <property> | 
                  
                          |   | 77 |  <name>http.agent.name</name> | 
                  
                          |   | 78 |  <value>My Nutch Spider</value> | 
                  
                          |   | 79 | </property> | 
                  
                          |   | 80 | }}} | 
                  
                          |   | 81 |  | 
                  
                          |   | 82 | ==== 設定欲爬取的網址 ==== | 
                  
                          |   | 83 |  * 建立網址資料(以爬取http://www.nchc.) | 
                  
                          |   | 84 | {{{ | 
                  
                          |   | 85 | $ mkdir -p [$NUTCH_HOME]/urls | 
                  
                          |   | 86 | $ echo "http://www.nchc.org.tw/tw/" >> [$NUTCH_HOME]/urls/seed.txt | 
                  
                          |   | 87 | }}} | 
                  
                          |   | 88 | ==== 設定filter ==== | 
                  
                          |   | 89 | {{{ | 
                  
                          |   | 90 | $ vim [$NUTCH_HOME]/conf/regex-urlfilter.txt | 
                  
                          |   | 91 | }}} | 
                  
                          |   | 92 |  * 用下列文字取代原始設定 | 
                  
                          |   | 93 | {{{ | 
                  
                          |   | 94 | #!text | 
                  
                          |   | 95 | # accept anything else | 
                  
                          |   | 96 | +. | 
                  
                          |   | 97 | }}} | 
                  
                          |   | 98 |  | 
                  
                          |   | 99 | ==== 透過指令執行爬取任務 ==== | 
                  
                          |   | 100 |  * 深度3層,每層最多抓取五個文件 | 
                  
                          |   | 101 | {{{ | 
                  
                          |   | 102 | $ [$NUTCH_HOME]/bin/nutch crawl urls -dir crawl -depth 3 -topN 5 | 
                  
                          |   | 103 | solrUrl is not set, indexing will be skipped... | 
                  
                          |   | 104 | crawl started in: crawl | 
                  
                          |   | 105 | rootUrlDir = urls | 
                  
                          |   | 106 | threads = 10 | 
                  
                          |   | 107 | depth = 3 | 
                  
                          |   | 108 | solrUrl=null | 
                  
                          |   | 109 | topN = 5 | 
                  
                          |   | 110 | Injector: starting at 2012-09-11 16:25:29 | 
                  
                          |   | 111 | Injector: crawlDb: crawl/crawldb | 
                  
                          |   | 112 | Injector: urlDir: urls | 
                  
                          |   | 113 | Injector: Converting injected urls to crawl db entries. | 
                  
                          |   | 114 | ...(略) | 
                  
                          |   | 115 | }}} | 
                  
                          |   | 116 |  | 
                  
                          |   | 117 |  * 出現以下訊息,表示已經抓取完成 | 
                  
                          |   | 118 | {{{ | 
                  
                          |   | 119 |  | 
                  
                          |   | 120 | }}} |