错误日志如下:

-- ::, DEBUG [org.apache.hadoop.metrics2.lib.MutableMetricsFactory] - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, value=[Rate of successful kerberos logins and latency (milliseconds)], valueName=Time)
-- ::, DEBUG [org.apache.hadoop.metrics2.lib.MutableMetricsFactory] - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, value=[Rate of failed kerberos logins and latency (milliseconds)], valueName=Time)
-- ::, DEBUG [org.apache.hadoop.metrics2.impl.MetricsSystemImpl] - UgiMetrics, User and group related metrics
-- ::, DEBUG [org.apache.hadoop.security.authentication.util.KerberosName] - Kerberos krb5 configuration not found, setting default realm to empty
-- ::, DEBUG [org.apache.hadoop.security.Groups] - Creating new Groups object
-- ::, DEBUG [org.apache.hadoop.util.NativeCodeLoader] - Trying to load the custom-built native-hadoop library...
-- ::, DEBUG [org.apache.hadoop.util.NativeCodeLoader] - Failed to load native-hadoop with error: java.lang.UnsatisfiedLinkError: no hadoop in java.library.path
-- ::, DEBUG [org.apache.hadoop.util.NativeCodeLoader] - java.library.path=D:\Software\Java\jdk1..0_66\bin;C:\Windows\Sun\Java\bin;C:\Windows\system32;C:\Windows;D:/Software/Java/jdk1..0_66/bin/../jre/bin/server;D:/Software/Java/jdk1..0_66/bin/../jre/bin;D:/Software/Java/jdk1..0_66/bin/../jre/lib/amd64;D:\Software\Anaconda;D:\Software\Anaconda\Library\mingw-w64\bin;D:\Software\Anaconda\Library\usr\bin;D:\Software\Anaconda\Library\bin;D:\Software\Anaconda\Scripts;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.\;D:\Software\MySQL Server\bin;D:\Software\CodeBlocks\share\CodeBlocks\MinGW\bin;D:\Software\hadoop2.6.0\hadoop-2.6.\bin;D:\Software\CTEX\UserData\miktex\bin;D:\Software\CTEX\MiKTeX\miktex\bin;D:\Software\CTEX\CTeX\ctex\bin;D:\Software\CTEX\CTeX\cct\bin;D:\Software\CTEX\CTeX\ty\bin;D:\Software\CTEX\Ghostscript\gs9.\bin;D:\Software\CTEX\GSview\gsview;D:\Software\CTEX\WinEdt;D:\Software\Java\jdk1..0_66\bin;D:\Software\Java\jdk1..0_66\jre\bin;C:\Users\Zimo\AppData\Local\Microsoft\WindowsApps;HADOOP_HOME\bin;D:\Software\搜狐影音\5.2.3.15;D:\Software\Java\eclipse;;.
-- ::, WARN [org.apache.hadoop.util.NativeCodeLoader] - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
-- ::, DEBUG [org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback] - Falling back to shell based
-- ::, DEBUG [org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback] - Group mapping impl=org.apache.hadoop.security.ShellBasedUnixGroupsMapping
-- ::, DEBUG [org.apache.hadoop.security.Groups] - Group mapping impl=org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback; cacheTimeout=
-- ::, DEBUG [org.apache.hadoop.security.UserGroupInformation] - hadoop login
-- ::, DEBUG [org.apache.hadoop.security.UserGroupInformation] - hadoop login commit
-- ::, DEBUG [org.apache.hadoop.security.UserGroupInformation] - using local user:NTUserPrincipal: Zimo
-- ::, DEBUG [org.apache.hadoop.security.UserGroupInformation] - UGI loginUser:Zimo (auth:SIMPLE)
-- ::, DEBUG [org.apache.hadoop.hdfs.BlockReaderLocal] - dfs.client.use.legacy.blockreader.local = false
-- ::, DEBUG [org.apache.hadoop.hdfs.BlockReaderLocal] - dfs.client.read.shortcircuit = false
-- ::, DEBUG [org.apache.hadoop.hdfs.BlockReaderLocal] - dfs.client.domain.socket.data.traffic = false
-- ::, DEBUG [org.apache.hadoop.hdfs.BlockReaderLocal] - dfs.domain.socket.path =
-- ::, DEBUG [org.apache.hadoop.metrics2.impl.MetricsSystemImpl] - StartupProgress, NameNode startup progress
-- ::, DEBUG [org.apache.hadoop.io.retry.RetryUtils] - multipleLinearRandomRetry = null
-- ::, DEBUG [org.apache.hadoop.ipc.Server] - rpcKind=RPC_PROTOCOL_BUFFER, rpcRequestWrapperClass=class org.apache.hadoop.ipc.ProtobufRpcEngine$RpcRequestWrapper, rpcInvoker=org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker@543788f3
-- ::, DEBUG [org.apache.hadoop.hdfs.BlockReaderLocal] - Both short-circuit local reads and UNIX domain socket are disabled.
-- ::, ERROR [org.apache.hadoop.util.Shell] - Failed to locate the winutils binary in the hadoop binary path
java.io.IOException: Could not locate executable D:\Software\hadoop2.6.0\hadoop-2.6.0\bin\winutils.exe in the Hadoop binaries.
at org.apache.hadoop.util.Shell.getQualifiedBinPath(Shell.java:)
at org.apache.hadoop.util.Shell.getWinUtilsPath(Shell.java:)
at org.apache.hadoop.util.Shell.<clinit>(Shell.java:)
at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:)
at org.apache.hadoop.conf.Configuration.getTrimmedStrings(Configuration.java:)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:)
at org.apache.hadoop.fs.FileSystem.access$(FileSystem.java:)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:)
at com.hadoop.phoneStatistics.ExcelPhoneStatistics.run(ExcelPhoneStatistics.java:)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:)
at com.hadoop.phoneStatistics.ExcelPhoneStatistics.main(ExcelPhoneStatistics.java:)
-- ::, DEBUG [org.apache.hadoop.ipc.Client] - The ping interval is ms.
-- ::, DEBUG [org.apache.hadoop.ipc.Client] - Connecting to master/192.168.86.131:

 解决方法:

  1.未配置环境变量-->配置环境变量HADOOP_HOME,然后重启电脑。或者代码中设置System.setProperty("hadoop.home.dir", "hadoop安装路径")。

  2.还有一种可能HADOOP_HOME的bin目录下根本没有winutils.exe,下载地址https://github.com/srccodes/hadoop-common-2.2.0-bin然后解压到该目录。

以上就是博主为大家介绍的这一板块的主要内容,这都是博主自己的学习过程,希望能给大家带来一定的指导作用,有用的还望大家点个支持,如果对你没用也望包涵,有错误烦请指出。如有期待可关注博主以第一时间获取更新哦,谢谢! 

版权声明:本文为博主原创文章,未经博主允许不得转载。

ERROR [org.apache.hadoop.util.Shell] - Failed to locate the winutils binary in the hadoop binary path的更多相关文章

  1. windows本地调试安装hadoop(idea) : ERROR util.Shell: Failed to locate the winutils binary in the hadoop binary path

    1,本地安装hadoop https://mirrors.tuna.tsinghua.edu.cn/apache/hadoop/common/ 下载hadoop对应版本 (我本意是想下载hadoop ...

  2. idea 提示:ERROR util.Shell: Failed to locate the winutils binary in the hadoop binary path java.io.IOException解决方法

    Windows系统中的IDEA链接Linux里面的Hadoop的api时出现的问题 提示:ERROR util.Shell: Failed to locate the winutils binary ...

  3. Windows本地运行调试Spark或Hadoop程序失败:ERROR util.Shell: Failed to locate the winutils binary in the hadoop binary path

    报错内容 ERROR util.Shell: Failed to locate the winutils binary in the hadoop binary path java.io.IOExce ...

  4. ERROR Shell: Failed to locate the winutils binary in the hadoop binary path

    文章发自:http://www.cnblogs.com/hark0623/p/4170172.html  转发请注明 14/12/17 19:18:53 ERROR Shell: Failed to ...

  5. Hadoop:开发机运行spark程序,抛出异常:ERROR Shell: Failed to locate the winutils binary in the hadoop binary path

    问题: windows开发机运行spark程序,抛出异常:ERROR Shell: Failed to locate the winutils binary in the hadoop binary ...

  6. Spark- ERROR Shell: Failed to locate the winutils binary in the hadoop binary path java.io.IOException: Could not locate executable null\bin\winutils.exe in the Hadoop binaries.

    运行 mport org.apache.log4j.{Level, Logger} import org.apache.spark.rdd.RDD import org.apache.spark.{S ...

  7. 【解决】org.apache.hadoop.util.Shell$ExitCodeException: /bin/bash: line 0: fg: no job control

    [环境信息] Hadoop版本:2.4.0 客户端OS:Windows Server 2008 R2 服务器端OS:CentOS 6.4 [问题现象] 在通过Windows客户端向Linux服务器提交 ...

  8. java.lang.NullPointerException at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:482)

    1:问题出现的原因,部署好的hadoop-2.6.4进行window10操作hadoop api出现的错误,具体错误是我向hdfs上传文件,还好点,之前解决过,这里不叙述,这里说一下从hdfs下载文件 ...

  9. Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeException

    使用MapReduce编写的中文分词程序出现了 Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeExcept ...

随机推荐

  1. 利用src.rpm包修改源码后重新制作rpm包

    1. 下载 .src.rpm包.例:https://repos.fedorapeople.org/repos/openstack/EOL/openstack-grizzly/epel-6/openst ...

  2. mount总结

    挂载分区 mount基本语法 mount [参数] /dev/sdb1(需要挂载的分区) /sdb1(挂载目录) 参数是可选的,也可以不带参数,参数的使用方法(-o ro,sync,atime). 参 ...

  3. Ruby模块(module)

    Ruby模块(module) 2013-04-03 16:47:09|  分类: Ruby |  标签:ruby  require  load  extend  include  |字号 订阅     ...

  4. centos6 启动流程

    具体过程:1)加载BIOS的硬件信息,执行BIOS内置程序.2)读取MBR(Master Boot Record)中Boot Loader中的引导信息.3)加载内核Kernel boot到内存中.4) ...

  5. 关于使用struts2跳转后css和js失效的解决方式

    根据观察,主要是由于通过action跳转后的url会根据命名空间,自动跳转到命名空间子目录,使得当前引用的css和js查找不到,从而失效,根据这个原因,可使用四种办法解决: 1.使用struts2.x ...

  6. Ganglia 安装 for centos6.5

    环境: centos6.5x64 ,cdh5.1.0 三个节点: c1:gmeta gmond(gmeta类似于namenode,gmond 相当于datanode)主节点 c2: gmond (相当 ...

  7. springboot 使用war包部署

    ps://www.cnblogs.com/jiaoyiping/p/4251718.html

  8. 关于UsedRange方法选中了空区域的解决方案

    使用worksheet.usedrange属性去制作数据透视表的时候会出现blank项,debug时发现它选中了空的区域. 解决方案: 属性usedrange包含着带格式的.空白的单元格(即使设置过单 ...

  9. C++开源库(一) ----log4cpp详解

    我们在写程序的时候通常会希望将一些信息记录下来,方便我们进行日后的一些信息跟踪,错误排查等等.比如:我们在进行数据库操作的时候,我们通常希望知道现在是程序的哪一部分进行了数据库的操作,所以我们会记录下 ...

  10. Codeforces Round #507 (Div. 2)C(贪心,逆向遍历,构造)

    #include<bits/stdc++.h>using namespace std;int order[300007];long long t[300007],ans[300007];i ...