亚洲欧美第一页_禁久久精品乱码_粉嫩av一区二区三区免费野_久草精品视频

? 歡迎來到蟲蟲下載站! | ?? 資源下載 ?? 資源專輯 ?? 關于我們
? 蟲蟲下載站

?? crawlserver.html

?? 網絡爬蟲開源代碼
?? HTML
?? 第 1 頁 / 共 3 頁
字號:
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"><html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en" lang="en"><head><meta http-equiv="content-type" content="text/html; charset=UTF-8" /><title>CrawlServer xref</title><link type="text/css" rel="stylesheet" href="../../../../stylesheet.css" /></head><body><div id="overview"><a href="../../../../../apidocs/org/archive/crawler/datamodel/CrawlServer.html">View Javadoc</a></div><pre><a name="1" href="#1">1</a>   <em class="comment">/*<em class="comment"> Copyright (C) 2003 Internet Archive.</em></em><a name="2" href="#2">2</a>   <em class="comment"> *</em><a name="3" href="#3">3</a>   <em class="comment"> * This file is part of the Heritrix web crawler (crawler.archive.org).</em><a name="4" href="#4">4</a>   <em class="comment"> *</em><a name="5" href="#5">5</a>   <em class="comment"> * Heritrix is free software; you can redistribute it and/or modify</em><a name="6" href="#6">6</a>   <em class="comment"> * it under the terms of the GNU Lesser Public License as published by</em><a name="7" href="#7">7</a>   <em class="comment"> * the Free Software Foundation; either version 2.1 of the License, or</em><a name="8" href="#8">8</a>   <em class="comment"> * any later version.</em><a name="9" href="#9">9</a>   <em class="comment"> *</em><a name="10" href="#10">10</a>  <em class="comment"> * Heritrix is distributed in the hope that it will be useful,</em><a name="11" href="#11">11</a>  <em class="comment"> * but WITHOUT ANY WARRANTY; without even the implied warranty of</em><a name="12" href="#12">12</a>  <em class="comment"> * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the</em><a name="13" href="#13">13</a>  <em class="comment"> * GNU Lesser Public License for more details.</em><a name="14" href="#14">14</a>  <em class="comment"> *</em><a name="15" href="#15">15</a>  <em class="comment"> * You should have received a copy of the GNU Lesser Public License</em><a name="16" href="#16">16</a>  <em class="comment"> * along with Heritrix; if not, write to the Free Software</em><a name="17" href="#17">17</a>  <em class="comment"> * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA</em><a name="18" href="#18">18</a>  <em class="comment"> *</em><a name="19" href="#19">19</a>  <em class="comment"> * CrawlServer.java</em><a name="20" href="#20">20</a>  <em class="comment"> * Created on Apr 17, 2003</em><a name="21" href="#21">21</a>  <em class="comment"> *</em><a name="22" href="#22">22</a>  <em class="comment"> * $Header$</em><a name="23" href="#23">23</a>  <em class="comment"> */</em><a name="24" href="#24">24</a>  <strong>package</strong> <a href="../../../../org/archive/crawler/datamodel/package-summary.html">org.archive.crawler.datamodel</a>;<a name="25" href="#25">25</a>  <a name="26" href="#26">26</a>  <strong>import</strong> java.io.BufferedReader;<a name="27" href="#27">27</a>  <strong>import</strong> java.io.IOException;<a name="28" href="#28">28</a>  <strong>import</strong> java.io.InputStreamReader;<a name="29" href="#29">29</a>  <strong>import</strong> java.io.ObjectInputStream;<a name="30" href="#30">30</a>  <strong>import</strong> java.io.Serializable;<a name="31" href="#31">31</a>  <strong>import</strong> java.io.StringReader;<a name="32" href="#32">32</a>  <strong>import</strong> java.util.HashSet;<a name="33" href="#33">33</a>  <strong>import</strong> java.util.Set;<a name="34" href="#34">34</a>  <strong>import</strong> java.util.zip.Checksum;<a name="35" href="#35">35</a>  <a name="36" href="#36">36</a>  <strong>import</strong> org.apache.commons.httpclient.URIException;<a name="37" href="#37">37</a>  <strong>import</strong> org.archive.crawler.datamodel.credential.CredentialAvatar;<a name="38" href="#38">38</a>  <strong>import</strong> org.archive.crawler.framework.Checkpointer;<a name="39" href="#39">39</a>  <strong>import</strong> org.archive.crawler.framework.ToeThread;<a name="40" href="#40">40</a>  <strong>import</strong> org.archive.crawler.settings.CrawlerSettings;<a name="41" href="#41">41</a>  <strong>import</strong> org.archive.crawler.settings.SettingsHandler;<a name="42" href="#42">42</a>  <strong>import</strong> org.archive.io.ReplayInputStream;<a name="43" href="#43">43</a>  <strong>import</strong> org.archive.net.UURIFactory;<a name="44" href="#44">44</a>  <a name="45" href="#45">45</a>  <em>/**<em>*</em></em><a name="46" href="#46">46</a>  <em> * Represents a single remote "server".</em><a name="47" href="#47">47</a>  <em> *</em><a name="48" href="#48">48</a>  <em> * A server is a service on a host. There might be more than one service on a</em><a name="49" href="#49">49</a>  <em> * host differentiated by a port number.</em><a name="50" href="#50">50</a>  <em> *</em><a name="51" href="#51">51</a>  <em> * @author gojomo</em><a name="52" href="#52">52</a>  <em> */</em><a name="53" href="#53">53</a>  <strong>public</strong> <strong>class</strong> <a href="../../../../org/archive/crawler/datamodel/CrawlServer.html">CrawlServer</a> implements Serializable, CrawlSubstats.HasCrawlSubstats {<a name="54" href="#54">54</a>  <a name="55" href="#55">55</a>      <strong>private</strong> <strong>static</strong> <strong>final</strong> <strong>long</strong> serialVersionUID = -989714570750970369L;<a name="56" href="#56">56</a>  <a name="57" href="#57">57</a>      <strong>public</strong> <strong>static</strong> <strong>final</strong> <strong>long</strong> ROBOTS_NOT_FETCHED = -1;<a name="58" href="#58">58</a>      <em>/**<em>* only check if robots-fetch is perhaps superfluous </em></em><a name="59" href="#59">59</a>  <em>     * after this many tries */</em><a name="60" href="#60">60</a>      <strong>public</strong> <strong>static</strong> <strong>final</strong> <strong>long</strong> MIN_ROBOTS_RETRIES = 2;<a name="61" href="#61">61</a>  <a name="62" href="#62">62</a>      <strong>private</strong> <strong>final</strong> String server; <em class="comment">// actually, host+port in the https case</em><a name="63" href="#63">63</a>      <strong>private</strong> <strong>int</strong> port;<a name="64" href="#64">64</a>      <strong>private</strong> <strong>transient</strong> <a href="../../../../org/archive/crawler/settings/SettingsHandler.html">SettingsHandler</a> settingsHandler;<a name="65" href="#65">65</a>      <strong>private</strong> <a href="../../../../org/archive/crawler/datamodel/RobotsExclusionPolicy.html">RobotsExclusionPolicy</a> robots;<a name="66" href="#66">66</a>      <strong>long</strong> robotsFetched = ROBOTS_NOT_FETCHED;<a name="67" href="#67">67</a>      <strong>boolean</strong> validRobots = false;<a name="68" href="#68">68</a>      Checksum robotstxtChecksum;<a name="69" href="#69">69</a>      <a href="../../../../org/archive/crawler/datamodel/CrawlSubstats.html">CrawlSubstats</a> substats = <strong>new</strong> <a href="../../../../org/archive/crawler/datamodel/CrawlSubstats.html">CrawlSubstats</a>();<a name="70" href="#70">70</a>      <a name="71" href="#71">71</a>      <em class="comment">// how many consecutive connection errors have been encountered;</em><a name="72" href="#72">72</a>      <em class="comment">// used to drive exponentially increasing retry timeout or decision</em><a name="73" href="#73">73</a>      <em class="comment">// to 'freeze' entire class (queue) of URIs</em><a name="74" href="#74">74</a>      <strong>protected</strong> <strong>int</strong> consecutiveConnectionErrors = 0;<a name="75" href="#75">75</a>  <a name="76" href="#76">76</a>      <em>/**<em>*</em></em><a name="77" href="#77">77</a>  <em>     * Set of credential avatars.</em><a name="78" href="#78">78</a>  <em>     */</em><a name="79" href="#79">79</a>      <strong>private</strong> <strong>transient</strong> Set&lt;CredentialAvatar> avatars =  <strong>null</strong>;<a name="80" href="#80">80</a>  <a name="81" href="#81">81</a>      <em>/**<em>*</em></em><a name="82" href="#82">82</a>  <em>     * Creates a new CrawlServer object.</em><a name="83" href="#83">83</a>  <em>     *</em><a name="84" href="#84">84</a>  <em>     * @param h the host string for the server.</em><a name="85" href="#85">85</a>  <em>     */</em><a name="86" href="#86">86</a>      <strong>public</strong> <a href="../../../../org/archive/crawler/datamodel/CrawlServer.html">CrawlServer</a>(String h) {<a name="87" href="#87">87</a>          <em class="comment">// TODO: possibly check for illegal host string</em><a name="88" href="#88">88</a>          server = h;<a name="89" href="#89">89</a>          <strong>int</strong> colonIndex = server.lastIndexOf(<span class="string">":"</span>);<a name="90" href="#90">90</a>          <strong>if</strong> (colonIndex &lt; 0) {<a name="91" href="#91">91</a>              port = -1;<a name="92" href="#92">92</a>          } <strong>else</strong> {<a name="93" href="#93">93</a>              <strong>try</strong> {<a name="94" href="#94">94</a>                  port = Integer.parseInt(server.substring(colonIndex + 1));<a name="95" href="#95">95</a>              } <strong>catch</strong> (NumberFormatException e) {<a name="96" href="#96">96</a>                  port = -1;<a name="97" href="#97">97</a>              }<a name="98" href="#98">98</a>          }<a name="99" href="#99">99</a>      }<a name="100" href="#100">100</a> <a name="101" href="#101">101</a>     <em>/**<em>* Get the robots exclusion policy for this server.</em></em><a name="102" href="#102">102</a> <em>     *</em><a name="103" href="#103">103</a> <em>     * @return the robots exclusion policy for this server.</em><a name="104" href="#104">104</a> <em>     */</em><a name="105" href="#105">105</a>     <strong>public</strong> <a href="../../../../org/archive/crawler/datamodel/RobotsExclusionPolicy.html">RobotsExclusionPolicy</a> getRobots() {<a name="106" href="#106">106</a>         <strong>return</strong> robots;<a name="107" href="#107">107</a>     }<a name="108" href="#108">108</a> <a name="109" href="#109">109</a>     <em>/**<em>* Set the robots exclusion policy for this server.</em></em><a name="110" href="#110">110</a> <em>     *</em><a name="111" href="#111">111</a> <em>     * @param policy the policy to set.</em><a name="112" href="#112">112</a> <em>     */</em><a name="113" href="#113">113</a>     <strong>public</strong> <strong>void</strong> setRobots(<a href="../../../../org/archive/crawler/datamodel/RobotsExclusionPolicy.html">RobotsExclusionPolicy</a> policy) {<a name="114" href="#114">114</a>         robots = policy;<a name="115" href="#115">115</a>     }<a name="116" href="#116">116</a> <a name="117" href="#117">117</a>     <strong>public</strong> String toString() {<a name="118" href="#118">118</a>         <strong>return</strong> <span class="string">"CrawlServer("</span>+server+<span class="string">")"</span>;<a name="119" href="#119">119</a>     }<a name="120" href="#120">120</a> <a name="121" href="#121">121</a>     <em>/**<em>* Update the robots exclusion policy.</em></em>

?? 快捷鍵說明

復制代碼 Ctrl + C
搜索代碼 Ctrl + F
全屏模式 F11
切換主題 Ctrl + Shift + D
顯示快捷鍵 ?
增大字號 Ctrl + =
減小字號 Ctrl + -
亚洲欧美第一页_禁久久精品乱码_粉嫩av一区二区三区免费野_久草精品视频
91精品免费观看| 亚洲免费观看在线视频| 中文字幕成人网| 五月综合激情网| 成人av电影在线播放| 69堂国产成人免费视频| 亚洲色图欧美偷拍| 国产精品伊人色| 日韩欧美在线综合网| 一区av在线播放| av一区二区三区| 亚洲精品一区二区在线观看| 亚洲国产综合色| 免费日韩伦理电影| 精品视频免费看| 亚洲视频一区二区在线观看| 国产激情视频一区二区三区欧美| 欧美一区二区观看视频| 午夜精品久久久久影视| 99国产精品久| 国产精品女同一区二区三区| 久久99最新地址| 精品久久国产字幕高潮| 日本欧美大码aⅴ在线播放| 欧美色老头old∨ideo| 亚洲综合小说图片| 91国产免费观看| 亚洲精品高清视频在线观看| 99久久99久久精品免费观看| 中文字幕巨乱亚洲| 成人一区二区三区| 国产精品不卡一区二区三区| 国产精品一区二区久激情瑜伽| 精品久久久久久久久久久久包黑料| 麻豆精品视频在线观看免费| 日韩视频免费观看高清完整版在线观看 | 日韩成人伦理电影在线观看| 欧美亚洲图片小说| 图片区日韩欧美亚洲| 7777精品伊人久久久大香线蕉的 | 91精品国产综合久久久蜜臀粉嫩| 亚洲国产精品精华液网站| 91福利精品第一导航| 五月天婷婷综合| 日韩情涩欧美日韩视频| 国产在线精品国自产拍免费| 国产日韩成人精品| 97精品久久久午夜一区二区三区| 最新欧美精品一区二区三区| 91福利精品视频| 久久精品国产亚洲高清剧情介绍| 精品久久国产97色综合| 成人av中文字幕| 性做久久久久久免费观看欧美| 91精品在线观看入口| 国产成人午夜精品5599| 成人久久视频在线观看| 奇米影视在线99精品| 欧美三级中文字幕在线观看| 视频一区欧美精品| 久久九九久久九九| 91福利视频网站| 激情深爱一区二区| 亚洲欧洲日韩av| 欧美一区二区私人影院日本| 国产一区二区在线视频| 亚洲品质自拍视频网站| 欧美一级高清大全免费观看| 成人一区二区视频| 日韩中文字幕亚洲一区二区va在线 | 91精品国产乱| 高清在线不卡av| 日韩激情一二三区| 亚洲国产高清在线观看视频| 欧美久久高跟鞋激| 成人性视频免费网站| 亚洲国产精品久久艾草纯爱| 国产欧美日韩精品在线| 欧美日韩国产另类不卡| 成人av第一页| 男人操女人的视频在线观看欧美| 国产精品免费视频网站| 欧美一区二区视频免费观看| 99久久婷婷国产| 国内精品久久久久影院薰衣草| 亚洲欧美韩国综合色| 国产亚洲欧美日韩日本| 911国产精品| 日本久久一区二区三区| 成人激情午夜影院| 国产一区二区网址| 日韩成人精品视频| 亚洲男人电影天堂| 国产精品―色哟哟| 久久午夜羞羞影院免费观看| 91精品国产高清一区二区三区蜜臀 | 91一区一区三区| 成人一区二区三区中文字幕| 韩日欧美一区二区三区| 婷婷国产v国产偷v亚洲高清| 亚洲一区二区在线免费看| 国产精品久久看| 国产欧美日韩综合精品一区二区| 日韩欧美国产三级| 91精品国产91久久综合桃花 | 91尤物视频在线观看| 国产成人精品aa毛片| 国产中文字幕精品| 久久精品国产99国产精品| 日精品一区二区| 日韩精品一级中文字幕精品视频免费观看| 亚洲欧美视频在线观看| 18涩涩午夜精品.www| 国产精品久久久久久亚洲伦 | 色香蕉成人二区免费| av亚洲产国偷v产偷v自拍| 成人一区在线看| 91亚洲大成网污www| 91视频在线观看免费| 一本一道综合狠狠老| 色综合久久综合| 欧洲色大大久久| 欧美日韩在线播| 欧美一区二区三区四区久久| 日韩一区和二区| 久久精品网站免费观看| 国产精品国产三级国产aⅴ原创| 中文字幕电影一区| 亚洲欧美国产毛片在线| 亚洲国产va精品久久久不卡综合| 午夜电影网一区| 激情成人午夜视频| 粉嫩aⅴ一区二区三区四区| av电影天堂一区二区在线| 欧美综合在线视频| 欧美一区二区三区免费观看视频| 日韩欧美一级二级三级久久久| 精品人在线二区三区| 国产三级欧美三级日产三级99| 国产精品第13页| 午夜精品影院在线观看| 国产乱码精品一区二区三区av| 成a人片国产精品| 欧美亚洲国产怡红院影院| 日韩一区二区在线观看视频| 国产欧美1区2区3区| 亚洲黄色av一区| 久久国产生活片100| 成人中文字幕在线| 欧美另类z0zxhd电影| 久久久99久久| 五月婷婷激情综合| 国产盗摄一区二区| 欧美日韩一区二区三区高清| 精品动漫一区二区三区在线观看| 国产精品久久久久久久午夜片| 天天av天天翘天天综合网 | 欧美日韩在线一区二区| 国产亚洲成aⅴ人片在线观看| 亚洲成a天堂v人片| 高清不卡一区二区在线| 555夜色666亚洲国产免| 国产精品伦一区| 久久成人麻豆午夜电影| 色94色欧美sute亚洲13| 国产人伦精品一区二区| 三级一区在线视频先锋| gogogo免费视频观看亚洲一| 欧美大片在线观看一区| 亚洲一区二区三区四区在线免费观看| 久久精品国产精品青草| 欧美日韩国产片| 亚洲欧美乱综合| av在线播放不卡| 国产日韩欧美精品综合| 美女免费视频一区| 欧美美女黄视频| 亚洲精品中文字幕在线观看| 国产成人精品一区二区三区四区 | 成人精品国产一区二区4080| 欧美成人精品福利| 天涯成人国产亚洲精品一区av| 色诱视频网站一区| 亚洲欧洲精品一区二区三区| 国产精品一区二区黑丝| 欧美大白屁股肥臀xxxxxx| 亚洲成人自拍网| 91电影在线观看| 亚洲精品综合在线| 91猫先生在线| 亚洲精品一二三四区| 91麻豆自制传媒国产之光| 中文字幕va一区二区三区| 成人美女在线观看| 国产精品嫩草影院av蜜臀| 国产成人午夜片在线观看高清观看| 久久先锋影音av鲁色资源| 国产精品综合在线视频| 国产亚洲自拍一区| 懂色av一区二区在线播放|