亚洲欧美第一页_禁久久精品乱码_粉嫩av一区二区三区免费野_久草精品视频

? 歡迎來(lái)到蟲(chóng)蟲(chóng)下載站! | ?? 資源下載 ?? 資源專輯 ?? 關(guān)于我們
? 蟲(chóng)蟲(chóng)下載站

?? jobconf.java

?? Hadoop是一個(gè)用于運(yùn)行應(yīng)用程序在大型集群的廉價(jià)硬件設(shè)備上的框架。Hadoop為應(yīng)用程序透明的提供了一組穩(wěn)定/可靠的接口和數(shù)據(jù)運(yùn)動(dòng)。在 Hadoop中實(shí)現(xiàn)了Google的MapReduce算法
?? JAVA
字號(hào):
/** * Copyright 2005 The Apache Software Foundation * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * *     http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */package org.apache.hadoop.mapred;import java.io.IOException;import java.io.File;import java.util.StringTokenizer;import java.util.ArrayList;import java.util.Collections;import java.util.Enumeration;import java.net.URL;import java.net.URLDecoder;import org.apache.hadoop.fs.FileUtil;import org.apache.hadoop.conf.Configuration;import org.apache.hadoop.io.Writable;import org.apache.hadoop.io.WritableComparable;import org.apache.hadoop.io.WritableComparator;import org.apache.hadoop.io.LongWritable;import org.apache.hadoop.io.UTF8;import org.apache.hadoop.mapred.lib.IdentityMapper;import org.apache.hadoop.mapred.lib.IdentityReducer;import org.apache.hadoop.mapred.lib.HashPartitioner;/** A map/reduce job configuration.  This names the {@link Mapper}, combiner * (if any), {@link Partitioner}, {@link Reducer}, {@link InputFormat}, and * {@link OutputFormat} implementations to be used.  It also indicates the set * of input files, and where the output files should be written. */public class JobConf extends Configuration {  public JobConf() {    super();  }      /**   * Construct a map/reduce job configuration.   *    * @param conf   *          a Configuration whose settings will be inherited.   */  public JobConf(Configuration conf) {    super(conf);    addDefaultResource("mapred-default.xml");  }  /** Construct a map/reduce job configuration.   *    * @param conf a Configuration whose settings will be inherited.   * @param aClass a class whose containing jar is used as the job's jar.   */  public JobConf(Configuration conf, Class aClass) {    this(conf);    String jar = findContainingJar(aClass);    if (jar != null) {      setJar(jar);    }  }  /** Construct a map/reduce configuration.   *   * @param config a Configuration-format XML job description file   */  public JobConf(String config) {    this(new File(config));  }  /** Construct a map/reduce configuration.   *   * @param config a Configuration-format XML job description file   */  public JobConf(File config) {    super();    addDefaultResource("mapred-default.xml");    addDefaultResource(config);  }  public String getJar() { return get("mapred.jar"); }  public void setJar(String jar) { set("mapred.jar", jar); }  public File getSystemDir() {    return new File(get("mapred.system.dir", "/tmp/hadoop/mapred/system"))      .getAbsoluteFile();  }  public String[] getLocalDirs() throws IOException {    return getStrings("mapred.local.dir");  }  public void deleteLocalFiles() throws IOException {    String[] localDirs = getLocalDirs();    for (int i = 0; i < localDirs.length; i++) {      FileUtil.fullyDelete(new File(localDirs[i]), this);    }  }  public void deleteLocalFiles(String subdir) throws IOException {    String[] localDirs = getLocalDirs();    for (int i = 0; i < localDirs.length; i++) {      FileUtil.fullyDelete(new File(localDirs[i], subdir), this);    }  }  /** Constructs a local file name.  Files are distributed among configured   * local directories.*/  public File getLocalFile(String subdir, String name) throws IOException {    return getFile("mapred.local.dir", name + File.separator + subdir);  }  public void setInputDir(File dir) { set("mapred.input.dir", dir); }  public void addInputDir(File dir) {    String dirs = get("mapred.input.dir");    set("mapred.input.dir", dirs == null ? dir.toString() : dirs + "," + dir);  }  public File[] getInputDirs() {    String dirs = get("mapred.input.dir", "");    ArrayList list = Collections.list(new StringTokenizer(dirs, ","));    File[] result = new File[list.size()];    for (int i = 0; i < list.size(); i++) {      result[i] = new File((String)list.get(i));    }    return result;  }  /**   * Get the reported username for this job.   * @return the username   */  public String getUser() {    return get("user.name");  }    /**   * Set the reported username for this job.   * @param user the username   */  public void setUser(String user) {    set("user.name", user);  }    /**   * Set the current working directory for the default file system   * @param dir the new current working directory   */  public void setWorkingDirectory(String dir) {    set("mapred.working.dir", dir);  }    /**   * Get the current working directory for the default file system.   * @return the directory name   */  public String getWorkingDirectory() {    return get("mapred.working.dir");   }    public File getOutputDir() {     String name = get("mapred.output.dir");    return name == null ? null: new File(name);  }  public void setOutputDir(File dir) { set("mapred.output.dir", dir); }  public InputFormat getInputFormat() {    return (InputFormat)newInstance(getClass("mapred.input.format.class",                                             TextInputFormat.class,                                             InputFormat.class));  }  public void setInputFormat(Class theClass) {    setClass("mapred.input.format.class", theClass, InputFormat.class);  }  public OutputFormat getOutputFormat() {    return (OutputFormat)newInstance(getClass("mapred.output.format.class",                                              TextOutputFormat.class,                                              OutputFormat.class));  }  public void setOutputFormat(Class theClass) {    setClass("mapred.output.format.class", theClass, OutputFormat.class);  }    public Class getInputKeyClass() {    return getClass("mapred.input.key.class",                    LongWritable.class, WritableComparable.class);  }  public void setInputKeyClass(Class theClass) {    setClass("mapred.input.key.class", theClass, WritableComparable.class);  }  public Class getInputValueClass() {    return getClass("mapred.input.value.class", UTF8.class, Writable.class);  }  public void setInputValueClass(Class theClass) {    setClass("mapred.input.value.class", theClass, Writable.class);  }  public Class getOutputKeyClass() {    return getClass("mapred.output.key.class",                    LongWritable.class, WritableComparable.class);  }  public void setOutputKeyClass(Class theClass) {    setClass("mapred.output.key.class", theClass, WritableComparable.class);  }  public WritableComparator getOutputKeyComparator() {    Class theClass = getClass("mapred.output.key.comparator.class", null,                              WritableComparator.class);    if (theClass != null)      return (WritableComparator)newInstance(theClass);    return WritableComparator.get(getOutputKeyClass());  }  public void setOutputKeyComparatorClass(Class theClass) {    setClass("mapred.output.key.comparator.class",             theClass, WritableComparator.class);  }  public Class getOutputValueClass() {    return getClass("mapred.output.value.class", UTF8.class, Writable.class);  }  public void setOutputValueClass(Class theClass) {    setClass("mapred.output.value.class", theClass, Writable.class);  }  public Class getMapperClass() {    return getClass("mapred.mapper.class", IdentityMapper.class, Mapper.class);  }  public void setMapperClass(Class theClass) {    setClass("mapred.mapper.class", theClass, Mapper.class);  }  public Class getMapRunnerClass() {    return getClass("mapred.map.runner.class",                    MapRunner.class, MapRunnable.class);  }  public void setMapRunnerClass(Class theClass) {    setClass("mapred.map.runner.class", theClass, MapRunnable.class);  }  public Class getPartitionerClass() {    return getClass("mapred.partitioner.class",                    HashPartitioner.class, Partitioner.class);  }  public void setPartitionerClass(Class theClass) {    setClass("mapred.partitioner.class", theClass, Partitioner.class);  }  public Class getReducerClass() {    return getClass("mapred.reducer.class",                    IdentityReducer.class, Reducer.class);  }  public void setReducerClass(Class theClass) {    setClass("mapred.reducer.class", theClass, Reducer.class);  }  public Class getCombinerClass() {    return getClass("mapred.combiner.class", null, Reducer.class);  }  public void setCombinerClass(Class theClass) {    setClass("mapred.combiner.class", theClass, Reducer.class);  }    /**   * Should speculative execution be used for this job?   * @return Defaults to true   */  public boolean getSpeculativeExecution() {     return getBoolean("mapred.speculative.execution", true);  }    /**   * Turn on or off speculative execution for this job.   * In general, it should be turned off for map jobs that have side effects.   */  public void setSpeculativeExecution(boolean new_val) {    setBoolean("mapred.speculative.execution", new_val);  }    public int getNumMapTasks() { return getInt("mapred.map.tasks", 1); }  public void setNumMapTasks(int n) { setInt("mapred.map.tasks", n); }  public int getNumReduceTasks() { return getInt("mapred.reduce.tasks", 1); }  public void setNumReduceTasks(int n) { setInt("mapred.reduce.tasks", n); }  /**   * Get the user-specified job name. This is only used to identify the    * job to the user.   * @return the job's name, defaulting to ""   */  public String getJobName() {    return get("mapred.job.name", "");  }    /**   * Set the user-specified job name.   * @param name the job's new name   */  public void setJobName(String name) {    set("mapred.job.name", name);  }    public Object newInstance(Class theClass) {    Object result;    try {      result = theClass.newInstance();    } catch (Exception e) {      throw new RuntimeException(e);    }    if (result instanceof JobConfigurable)      ((JobConfigurable)result).configure(this);    return result;  }  /** Find a jar that contains a class of the same name, if any.   * It will return a jar file, even if that is not the first thing   * on the class path that has a class with the same name.   * @author Owen O'Malley   * @param my_class the class to find   * @return a jar file that contains the class, or null   * @throws IOException   */  private static String findContainingJar(Class my_class) {    ClassLoader loader = my_class.getClassLoader();    String class_file = my_class.getName().replaceAll("\\.", "/") + ".class";    try {      for(Enumeration itr = loader.getResources(class_file);          itr.hasMoreElements();) {        URL url = (URL) itr.nextElement();        if ("jar".equals(url.getProtocol())) {          String toReturn = url.getPath();          if (toReturn.startsWith("file:")) {            toReturn = toReturn.substring("file:".length());          }          toReturn = URLDecoder.decode(toReturn, "UTF-8");          return toReturn.replaceAll("!.*$", "");        }      }    } catch (IOException e) {      throw new RuntimeException(e);    }    return null;  }}

?? 快捷鍵說(shuō)明

復(fù)制代碼 Ctrl + C
搜索代碼 Ctrl + F
全屏模式 F11
切換主題 Ctrl + Shift + D
顯示快捷鍵 ?
增大字號(hào) Ctrl + =
減小字號(hào) Ctrl + -
亚洲欧美第一页_禁久久精品乱码_粉嫩av一区二区三区免费野_久草精品视频
久久国产人妖系列| 日韩二区三区四区| 久久 天天综合| 日韩欧美激情一区| 亚洲成人午夜电影| 欧美三级韩国三级日本一级| 亚洲欧美日韩国产综合| 色av一区二区| 亚洲第一电影网| 欧美一区二区三区婷婷月色| 裸体在线国模精品偷拍| 精品久久久久一区| 国产电影精品久久禁18| 亚洲欧洲av另类| 在线观看免费成人| 免费视频最近日韩| 精品av综合导航| 91蜜桃免费观看视频| 亚洲国产美国国产综合一区二区| 91精品婷婷国产综合久久性色 | 日韩视频一区二区三区| 久久精品久久久精品美女| 中日韩av电影| 欧美亚洲国产一区在线观看网站| 美国三级日本三级久久99| 亚洲国产精品精华液ab| 色视频一区二区| 国产精品一二三| 亚洲一区在线观看视频| 国产日产精品1区| 欧美一区中文字幕| 欧洲色大大久久| 91丨九色丨蝌蚪丨老版| 麻豆免费看一区二区三区| 亚洲丝袜美腿综合| 国产精品麻豆欧美日韩ww| 7799精品视频| 欧美一区二区三区男人的天堂| 91丨九色丨蝌蚪富婆spa| 国产成人av一区二区| 九色|91porny| 极品瑜伽女神91| 国产乱码精品1区2区3区| 麻豆精品在线播放| 久久精品免费看| 99久久er热在这里只有精品15| 国产欧美日韩视频在线观看| 欧美大度的电影原声| 日韩一区二区影院| 欧美成人a视频| 久久精品人人爽人人爽| 久久精品无码一区二区三区| 国产午夜一区二区三区| 亚洲国产精品成人综合色在线婷婷| 国产亚洲成年网址在线观看| 国产精品短视频| 玉米视频成人免费看| 亚洲成av人片在www色猫咪| 另类小说图片综合网| 国产精品18久久久久| 91在线你懂得| 日韩久久精品一区| 国产精品麻豆欧美日韩ww| 性欧美大战久久久久久久久| 日韩电影免费一区| 国产99久久久国产精品潘金| 色88888久久久久久影院按摩| 欧美一区二区三区系列电影| 成人av中文字幕| 国产成人一区在线| 精品国产乱码久久久久久夜甘婷婷| 91.com视频| ww久久中文字幕| 夜夜嗨av一区二区三区四季av| 久久超级碰视频| 91成人看片片| 国产无人区一区二区三区| 亚洲精品国产成人久久av盗摄 | 欧美高清视频一二三区| 日本一区二区三区国色天香 | 国产欧美精品一区aⅴ影院| 99麻豆久久久国产精品免费优播| 欧美日产在线观看| 亚洲欧美日韩一区二区三区在线观看| 久久精品噜噜噜成人88aⅴ| 欧美四级电影网| 综合久久给合久久狠狠狠97色| 免费高清在线一区| 欧美精品在线视频| 亚洲人成伊人成综合网小说| 成人福利在线看| 欧美激情一区在线| 精品一区二区免费看| 日韩三级视频在线看| 日本sm残虐另类| 日韩欧美国产高清| 国产乱人伦偷精品视频免下载| 成人动漫av在线| 亚洲少妇最新在线视频| 成人一区在线看| 国产精品久久久久影视| 97精品电影院| 亚洲成av人片一区二区梦乃 | 91国偷自产一区二区三区成为亚洲经典 | 99视频一区二区三区| 国产精品人成在线观看免费 | 粉嫩av亚洲一区二区图片| 中文字幕高清不卡| 色综合天天在线| 亚洲一区二区三区四区五区中文| 欧美色爱综合网| 九九国产精品视频| 国产精品嫩草99a| 欧美午夜精品一区| 麻豆91在线播放免费| 国产精品久久免费看| 91丨porny丨首页| 蜜臀精品久久久久久蜜臀| 国产视频亚洲色图| 欧美色综合影院| 高清视频一区二区| 亚洲精品美腿丝袜| 久久综合久久鬼色| av亚洲精华国产精华| 日本伊人午夜精品| 亚洲美女偷拍久久| 国产亚洲一区二区三区| 欧美日韩亚洲综合在线 | 91麻豆视频网站| 国产精品一品二品| 免费人成精品欧美精品| 亚洲人成伊人成综合网小说| 精品福利视频一区二区三区| 欧美午夜在线观看| 91美女蜜桃在线| 91污在线观看| 成人av在线资源| 成人一级视频在线观看| 国产精品亚洲午夜一区二区三区| 日韩精品视频网| 日韩av电影免费观看高清完整版| 一级做a爱片久久| 一区二区欧美在线观看| 一区二区三区四区在线播放| 亚洲国产成人自拍| 韩国午夜理伦三级不卡影院| 国产精品久久精品日日| 国产精品美女久久久久久2018| 国产午夜精品福利| 亚洲日本丝袜连裤袜办公室| 亚洲免费在线观看视频| 一区二区三区在线不卡| 亚洲国产aⅴ天堂久久| 亚洲国产欧美在线| 日韩vs国产vs欧美| 黄色精品一二区| 91免费看片在线观看| 欧美亚洲综合另类| 91精品国产综合久久精品| 精品国产91亚洲一区二区三区婷婷| 香蕉成人啪国产精品视频综合网| 亚洲成国产人片在线观看| 免费亚洲电影在线| 国产成人综合视频| 日本久久一区二区| 欧美一级视频精品观看| 中文字幕欧美区| 奇米色一区二区| 99热这里都是精品| 日韩欧美国产综合在线一区二区三区| 久久久电影一区二区三区| 亚洲mv在线观看| 成人午夜激情片| 欧美一级电影网站| 亚洲天堂精品视频| 精品一区二区免费视频| 欧美性猛片xxxx免费看久爱| 26uuu成人网一区二区三区| 亚洲电影第三页| 色一情一乱一乱一91av| 中日韩免费视频中文字幕| 久久国产夜色精品鲁鲁99| 在线观看亚洲成人| 亚洲欧洲国产专区| 国产xxx精品视频大全| 欧美va亚洲va在线观看蝴蝶网| 亚洲免费在线视频| 99re热视频这里只精品| 国产欧美日韩中文久久| 国产精品 日产精品 欧美精品| 日韩免费电影网站| 精品一区二区三区免费视频| 欧美一区二区三区色| 日韩精品一区第一页| 欧美精品一卡两卡| 天天爽夜夜爽夜夜爽精品视频| 91搞黄在线观看| 日本一不卡视频| 久久婷婷一区二区三区| 99久久综合色|