这个问题一般是在hadoop2.x版本里会出现,Hadoop的datanode需要访问namenode的jobhistory server,如果没有修改,则默认为0.0.0.0:10020,则可以修改mapred-site.xml文件:

<property>
    <name>mapreduce.jobhistory.address</name>

<!-- 配置实际的Master主机名和端口-->
    <value>0.0.0.0:10020</value>
</property>

<property>
    <name>mapreduce.jobhistory.webapp.address</name>

<!-- 配置实际的Master主机名和端口-->
    <value>0.0.0.0:19888</value>
</property>

最后不要忘记了启动jobhistory

 
$HADOOP_HOME/sbin/mr-jobhistory-daemon.sh start historyserver
 
 
********************************************************************
a:C:\words.txt  文件内容:  
hello alamp s
hello qq
 
hello xx
 
hello aa
 
hello swk
 
hello zbj
 
hello blm
 
blm   xixi
zbj  hehe
swk   haha
b. java
 
@Test
     public void testUpload() throws IllegalArgumentException, IOException {
          FSDataOutputStream out = fs.create(new Path(
                   "hdfs://itcast01:9000/upload2"));
          FileInputStream in = new FileInputStream(new File("c:/words.txt"));
          IOUtils.copyBytes(in, out, 2048, true);
     }
package cn.itcast.hadoop.mr;
 
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
 
 
public class WordCount {
 
    public static void main(String[] args) throws Exception {
        Configuration conf = new Configuration();
        conf.setInt("mapreduce.client.submit.file.replication", 20);
        Job job = Job.getInstance(conf);
 
        //notice
        job.setJarByClass(WordCount.class);
 
        //set mapper`s property
        job.setMapperClass(WCMapper.class);
        job.setMapOutputKeyClass(Text.class);
        job.setMapOutputValueClass(LongWritable.class);
        FileInputFormat.setInputPaths(job, new Path("/upload2/"));
 
        //set reducer`s property
        job.setReducerClass(WCReducer.class);
        job.setOutputKeyClass(Text.class);
        job.setOutputValueClass(LongWritable.class);
        FileOutputFormat.setOutputPath(job, new Path("/usr/mapperReduce"));
 
        //submit
        job.waitForCompletion(true);
    }
 
}
c.打jar包
                            
 
                           
      
[root@itcast01 usr]# hadoop jar mr.jar
17/05/20 16:44:09 INFO client.RMProxy: Connecting to ResourceManager at itcast01/192.168.233.128:8032
17/05/20 16:44:10 WARN mapreduce.JobSubmitter: Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this.
17/05/20 16:44:11 INFO input.FileInputFormat: Total input paths to process : 1
17/05/20 16:44:11 INFO mapreduce.JobSubmitter: number of splits:1
17/05/20 16:44:11 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1495288787912_0006
17/05/20 16:44:12 INFO impl.YarnClientImpl: Submitted application application_1495288787912_0006
17/05/20 16:44:12 INFO mapreduce.Job: The url to track the job: http://itcast01:8088/proxy/application_1495288787912_0006/
17/05/20 16:44:12 INFO mapreduce.Job: Running job: job_1495288787912_0006
17/05/20 16:44:55 INFO mapreduce.Job: Job job_1495288787912_0006 running in uber mode : false
17/05/20 16:44:55 INFO mapreduce.Job:  map 0% reduce 0%
17/05/20 16:46:24 INFO mapreduce.Job:  map 100% reduce 0%
17/05/20 16:47:02 INFO mapreduce.Job:  map 100% reduce 100%
17/05/20 16:47:03 INFO mapreduce.Job: Job job_1495288787912_0006 completed successfully
17/05/20 16:47:06 INFO mapreduce.Job: Counters: 49
        File System Counters
                FILE: Number of bytes read=435
                FILE: Number of bytes written=186471
                FILE: Number of read operations=0
                FILE: Number of large read operations=0
                FILE: Number of write operations=0
                HDFS: Number of bytes read=210
                HDFS: Number of bytes written=78
                HDFS: Number of read operations=6
                HDFS: Number of large read operations=0
                HDFS: Number of write operations=2
        Job Counters
                Launched map tasks=1
                Launched reduce tasks=1
                Data-local map tasks=1
                Total time spent by all maps in occupied slots (ms)=65092
                Total time spent by all reduces in occupied slots (ms)=32649
                Total time spent by all map tasks (ms)=65092
                Total time spent by all reduce tasks (ms)=32649
                Total vcore-seconds taken by all map tasks=65092
                Total vcore-seconds taken by all reduce tasks=32649
                Total megabyte-seconds taken by all map tasks=66654208
                Total megabyte-seconds taken by all reduce tasks=33432576
        Map-Reduce Framework
                Map input records=17
                Map output records=32
                Map output bytes=365
                Map output materialized bytes=435
                Input split bytes=93
                Combine input records=0
                Combine output records=0
                Reduce input groups=13
                Reduce shuffle bytes=435
                Reduce input records=32
                Reduce output records=13
                Spilled Records=64
                Shuffled Maps =1
                Failed Shuffles=0
                Merged Map outputs=1
                GC time elapsed (ms)=290
                CPU time spent (ms)=5530
                Physical memory (bytes) snapshot=284258304
                Virtual memory (bytes) snapshot=1685770240
                Total committed heap usage (bytes)=136515584
        Shuffle Errors
                BAD_ID=0
                CONNECTION=0
                IO_ERROR=0
                WRONG_LENGTH=0
                WRONG_MAP=0
                WRONG_REDUCE=0
        File Input Format Counters
                Bytes Read=117
        File Output Format Counters
                Bytes Written=78
 
[root@itcast01 usr]# hadoop fs -ls /
Found 9 items
-rw-r--r--   1 root supergroup  153512879 2017-05-17 05:34 /jdk
-rw-r--r--   1 root supergroup         62 2017-05-20 13:54 /score_in
drwx------   - root supergroup          0 2017-05-17 06:15 /tmp
-rw-r--r--   3 root supergroup         32 2017-05-17 14:47 /upload
-rw-r--r--   3 root supergroup        117 2017-05-20 16:35 /upload2
drwxr-xr-x   - root supergroup          0 2017-05-20 16:44 /usr
drwxr-xr-x   - root supergroup          0 2017-05-17 06:18 /wcout
-rw-r--r--   1 root supergroup         70 2017-05-17 06:12 /words
drwxr-xr-x   - root supergroup          0 2017-05-20 15:23 /xx
[root@itcast01 usr]# hadoop fs -cat /usr/upload2/part-r-00000
cat: `/usr/upload2/part-r-00000': No such file or directory
 
 
[root@itcast01 usr]# hadoop fs -ls /usr/
Found 4 items
drwxr-xr-x   - root supergroup          0 2017-05-17 14:54 /usr/local
drwxr-xr-x   - root supergroup          0 2017-05-20 16:47 /usr/mapperReduce
drwxr-xr-x   - root supergroup          0 2017-05-20 16:08 /usr/swk
drwxr-xr-x   - root supergroup          0 2017-05-17 11:25 /usr/test
[root@itcast01 usr]# hadoop fs -ls /usr/mapperReduce
Found 2 items
-rw-r--r--   1 root supergroup          0 2017-05-20 16:47 /usr/mapperReduce/_SUCCESS
-rw-r--r--   1 root supergroup         78 2017-05-20 16:46 /usr/mapperReduce/part-r-00000
[root@itcast01 usr]# hadoop fs -cat /usr/mapperReduce/part-r-00000
        11
aa      1
alamp   1
blm     2
haha    1
hehe    1
hello   7
qq      1
s       1
swk     2
xixi    1
xx      1
zbj     2
[root@itcast01 usr]#

hadoop 遇到java.net.ConnectException: to 0.0.0.0:10020 failed on connection的更多相关文章

  1. hadoop报错:java.io.IOException(java.net.ConnectException: Call From xxx/xxx to xxx:10020 failed on connection exception: java.net.ConnectException: 拒绝连接

    任务一直报错 现象比较奇怪,部分任务可以正常跑,部分问题报错 报错信息如下: Ended Job = job_1527476268558_132947 with exception 'java.io. ...

  2. Hadoop格式化 From hu-hadoop1/192.168.11.11 to hu-hadoop2:8485 failed on connection exception: java.net.

    192.168.11.12:8485: Call From hu-hadoop1/192.168.11.11 to hu-hadoop2:8485 failed on connection excep ...

  3. Hadoop: failed on connection exception: java.net.ConnectException: Connection refuse

    ssh那些都已经搞了,跑一个书上的例子出现了Connection Refused异常,如下: 12/04/09 01:00:54 INFO ipc.Client: Retrying connect t ...

  4. Call to localhost/127.0.0.1:9000 failed on connection exception:java.net.ConnectException的解决方案

    Call to localhost/127.0.0.1:9000 failed on connection exception:java.net.ConnectException的解决方案 作者:凯鲁 ...

  5. android异常: java.net.ConnectException: localhost/127.0.0.1:8080 - Connection refused

    android手机做下载文件时,报了如下异常: java.net.ConnectException: localhost/127.0.0.1:8080 - Connection refused 模拟器 ...

  6. [转]android访问网络:java.net.ConnectException: localhost/127.0.0.1:8888 - Connection refused

    这对刚学会向tomcat模拟的本地服务器发送请求的同学非常重要! 转自:http://wing123.iteye.com/blog/1873763 描述:在做注册功能的时候,向本地服务器:127.0. ...

  7. 【android】java.net.ConnectException: localhost/127.0.0.1:8080 - Connection refused

    调试中通过android simulator模拟器链接localhost或者127.0.0.1,因为我在电脑上面建立了apache,我的代码大概就是URL url = new URL(urlStrin ...

  8. Debian部署RMI异常:java.rmi.ConnectException: Connection refused to host: 127.0.1.1;

    现象:在windows上部署RMI很顺利,但移到debian上部署后,客户端报异常: java.rmi.ConnectException: Connection refused to host: 12 ...

  9. 调用远程主机上的 RMI 服务时抛出 java.rmi.ConnectException: Connection refused to host: 127.0.0.1 异常原因及解决方案

    最近使用 jmx 遇到一个问题,client/server 同在一台机器上,jmx client能够成功连接 server,如果把 server 移植到另一台机器上192.168.134.128,抛出 ...

随机推荐

  1. android&sqlsever

    http://blog.csdn.net/chaoyu168/article/details/51910601

  2. acid. cap

    BASE是下面三个术语的缩写: 基本可用(Basically Available) 软状态(Soft state) 最终一致(Eventually consistent) 目前最快的KV数据库,10W ...

  3. Page9:结构分解以及系统内部稳定和BIBO稳定概念及其性质[Linear System Theory]

    内容包含系统能控性结构分解.系统能观测性结构分解以及系统结构规范分解原理,线性系统的内部稳定.BIBO稳定概念及其性质

  4. [DPI][TCP] linux API的接口如何控制urgent包的收发

    做DPI,写协议栈的时候,处理到了urgent数据包.突然好奇应用层是如何控制发出urgent包的呢?而接收端又是如何知道,接受到了urgent包的呢? man 7 tcp,中有如下一段: TCP s ...

  5. 《HTTP - 理解 Content-Type》

    一:引言 在此之前先看一个小例子:(html 上传文件,服务端为PHP) <?php var_dump($_FILES);?> <!DOCTYPE html> <html ...

  6. 20165336 2017-2018-2 《Java程序设计》第8周学习总结

    20165336 2017-2018-2 <Java程序设计>第8周学习总结 教材学习内容总结 第十二章 1.程序:一段静态的代码.进程:程序的一次动态执行过程,它对应了从代码加载.执行至 ...

  7. 微信小程序首支视频广告片发布

    自2017年1月9日上线以来,微信小程序瞬间引爆网络,又迅速归于平静.对这个“无需安装.触手可及.用完即走.无需卸载”的小程序,微信从一开始就对它寄予了无限期望.在公布一系列规则更改后,微信小程序终于 ...

  8. 小程序支持打开APP了 还有小程序的标题栏也可以自定义

    就在刚刚,小程序上线两个新能力——小程序支持打开APP了,小程序的标题栏区域开放自定义.用户可以在小程序里更方便地获取到APP的服务了——APP链接分享到微信,打开小程序页面后,用户从该小程序页面里, ...

  9. 【SVM、决策树、adaboost、LR对比】

    一.SVM 1.应用场景: 文本和图像分类. 2.优点: 分类效果好:有效处理高维空间的数据:无局部最小值问题:不易过拟合(模型中含有L2正则项): 3.缺点: 样本数据量较大需要较长训练时间:噪声不 ...

  10. 静态库lib和动态库dll相关总结

    1.静态链接库LIB和动态链接库DLL的区别 若采用静态链接库,lib 中的指令都全部被直接包含在最终生成的 EXE 文件中了.而动态动态链接库则不必被包含在最终 EXE 文件中,EXE 文件执行时可 ...