报错信息:

INFO [2016-11-03 17:05:08,481] ({pool-2-thread-2} SchedulerFactory.java[jobFinished]:137) - Job remoteInterpretJob_1478163908285 finished by scheduler org.apache.zeppelin.jdbc.JDBCInterpreter1260005516
INFO [2016-11-03 17:12:54,001] ({pool-1-thread-2} InterpreterGroup.java[close]:141) - Close interpreter group 2C3DW1MMZ:shared_process
INFO [2016-11-03 17:12:54,001] ({pool-1-thread-2} InterpreterGroup.java[destroy]:205) - Destroy interpreter group 2C3DW1MMZ:shared_process
INFO [2016-11-03 17:13:10,488] ({Thread-0} RemoteInterpreterServer.java[run]:81) - Starting remote interpreter server on port 34181
INFO [2016-11-03 17:13:10,848] ({pool-1-thread-2} RemoteInterpreterServer.java[createInterpreter]:169) - Instantiate interpreter org.apache.zeppelin.jdbc.JDBCInterpreter
INFO [2016-11-03 17:13:10,942] ({pool-2-thread-2} SchedulerFactory.java[jobStarted]:131) - Job remoteInterpretJob_1478164390937 started by scheduler org.apache.zeppelin.jdbc.JDBCInterpreter968846253
INFO [2016-11-03 17:13:10,950] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: phoenix, value: user
INFO [2016-11-03 17:13:10,951] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: hive, value: url
INFO [2016-11-03 17:13:10,951] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: default, value: driver
INFO [2016-11-03 17:13:10,951] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: phoenix, value: driver
INFO [2016-11-03 17:13:10,951] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: hive, value: user
INFO [2016-11-03 17:13:10,951] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: psql, value: password
INFO [2016-11-03 17:13:10,952] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: psql, value: user
INFO [2016-11-03 17:13:10,952] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: psql, value: url
INFO [2016-11-03 17:13:10,952] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: default, value: user
INFO [2016-11-03 17:13:10,952] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: phoenix, value: hbase.client.retries.number
INFO [2016-11-03 17:13:10,952] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: phoenix, value: url
INFO [2016-11-03 17:13:10,953] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: tajo, value: url
INFO [2016-11-03 17:13:10,953] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: tajo, value: driver
INFO [2016-11-03 17:13:10,956] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: psql, value: driver
INFO [2016-11-03 17:13:10,956] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: default, value: password
INFO [2016-11-03 17:13:10,956] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: zeppelin, value: interpreter.localRepo
INFO [2016-11-03 17:13:10,957] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: hive, value: password
INFO [2016-11-03 17:13:10,957] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: zeppelin, value: jdbc.concurrent.use
INFO [2016-11-03 17:13:10,957] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: hive, value: driver
INFO [2016-11-03 17:13:10,957] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: common, value: max_count
INFO [2016-11-03 17:13:10,957] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: phoenix, value: password
INFO [2016-11-03 17:13:10,958] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: zeppelin, value: jdbc.concurrent.max_connection
INFO [2016-11-03 17:13:10,958] ({pool-2-thread-2} JDBCInterpreter.java[open]:142) - key: default, value: url
ERROR [2016-11-03 17:13:10,958] ({pool-2-thread-2} JDBCInterpreter.java[open]:159) - zeppelin will be ignored. driver.zeppelin and zeppelin.url is mandatory.
INFO [2016-11-03 17:13:10,958] ({pool-2-thread-2} JDBCInterpreter.java[getConnection]:219) - org.apache.hive.jdbc.HiveDriver
INFO [2016-11-03 17:13:11,086] ({pool-2-thread-2} Utils.java[parseURL]:285) - Supplied authorities: 192.18.10.100:10000
INFO [2016-11-03 17:13:11,088] ({pool-2-thread-2} Utils.java[parseURL]:372) - Resolved authority: 192.18.10.100:10000
ERROR [2016-11-03 17:13:11,121] ({pool-2-thread-2} Job.java[run]:189) - Job failed
java.lang.InternalError
at sun.security.ec.SunEC.initialize(Native Method)
at sun.security.ec.SunEC.access$000(SunEC.java:49)
at sun.security.ec.SunEC$1.run(SunEC.java:61)
at sun.security.ec.SunEC$1.run(SunEC.java:58)
at java.security.AccessController.doPrivileged(Native Method)
at sun.security.ec.SunEC.<clinit>(SunEC.java:58)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at java.lang.Class.newInstance(Class.java:383)
at sun.security.jca.ProviderConfig$2.run(ProviderConfig.java:221)
at sun.security.jca.ProviderConfig$2.run(ProviderConfig.java:206)
at java.security.AccessController.doPrivileged(Native Method)
at sun.security.jca.ProviderConfig.doLoadProvider(ProviderConfig.java:206)
at sun.security.jca.ProviderConfig.getProvider(ProviderConfig.java:187)
at sun.security.jca.ProviderList.loadAll(ProviderList.java:282)
at sun.security.jca.ProviderList.removeInvalid(ProviderList.java:299)
at sun.security.jca.Providers.getFullProviderList(Providers.java:173)
at java.security.Security.insertProviderAt(Security.java:360)
at java.security.Security.addProvider(Security.java:407)
at org.apache.hive.service.auth.PlainSaslHelper.<clinit>(PlainSaslHelper.java:52)
at org.apache.hive.jdbc.HiveConnection.createBinaryTransport(HiveConnection.java:396)
at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:187)
at org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:163)
at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:105)
at java.sql.DriverManager.getConnection(DriverManager.java:571)
at java.sql.DriverManager.getConnection(DriverManager.java:187)
at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:222)
at org.apache.zeppelin.jdbc.JDBCInterpreter.open(JDBCInterpreter.java:176)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:341)
at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
at org.apache.zeppelin.scheduler.ParallelScheduler$JobRunner.run(ParallelScheduler.java:162)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
INFO [2016-11-03 17:13:11,134] ({pool-2-thread-2} SchedulerFactory.java[jobFinished]:137) - Job remoteInterpretJob_1478164390937 finished by scheduler org.apache.zeppelin.jdbc.JDBCInterpreter968846253

解决办法:将openjdk换成oraclejdk1.7。

Zeppelin使用报错的更多相关文章

  1. adb驱动安装和使用报错笔记

    adb驱动安装 adb驱动下载地址:https://adb.clockworkmod.com/ 安装时候选择一个容易记住的路径,这个很重要,因为adb驱动没有自动配置环境变量,所以实验时候将adb安装 ...

  2. animate is not a function(zepto 使用报错)[转]

    animate is not a function(zepto 使用报错) 1.为什么使用zepto写animate报错? 因为zepto默认构建包含: Core, Ajax, Event, Form ...

  3. Windows下Git使用报错:warning:LF will be replaced by CRLF in ××××.××

    Windows下Git使用报错: warning:LF will be replaced by CRLF in ××××.××(文件名) The file will have its original ...

  4. yum源使用报错

    CentOS系统yum源使用报错:Error: Cannot retrieve repository metadata (repomd.xml) for repository: rpmforge. 服 ...

  5. 2019-9-9:渗透测试,docker下载dvwa,使用报错型sql注入dvwa

    docker下载dvwa镜像,报错型注入dvwa,low级 一,安装并配置docker 1,更新源,apt-get update && apt-get upgrade &&am ...

  6. .net core中Grpc使用报错:The remote certificate is invalid according to the validation procedure.

    因为Grpc采用HTTP/2作为通信协议,默认采用LTS/SSL加密方式传输,比如使用.net core启动一个服务端(被调用方)时: public static IHostBuilder Creat ...

  7. VirtualBox使用报错

    VirtualBox使用报错 1.启动报错:Failed to instantiate CLSID_VirtualBox... 报错内容: Failed to instantiate CLSID_Vi ...

  8. antd-mobile使用报错

    在第一次使用时,按照官网的进行配置,完了报错找不到antd-mobile下面的css 解决方法来源于 :https://github.com/ant-design/ant-design-mobile/ ...

  9. selenium使用报错“selenium.common.exceptions.WebDriverException: Message: 'geckodriver' executable needs to be in PATH.”

    安装了python3,使用pip安装了selenium,但是在使用时,报了“selenium.common.exceptions.WebDriverException: Message: 'gecko ...

随机推荐

  1. IOS开发之路二十一(UIWebView加载本地html)

    挺简单不多说的直接代码: // // ViewController.h // JSAndJson // // Created by WildCat on 13-9-8. // Copyright (c ...

  2. 读取xml并将节点保存到Excal

    using NPOI.HPSF; using NPOI.HSSF.UserModel; using NPOI.SS.UserModel; using System; using System.Coll ...

  3. fiddle2 代理HTTPS请求无效?解决方法。

    fiddle2: 捕获的https请求结尾跟着443,是因为没有开启HTTPS捕获. 解决方案,开启HTTPS捕获:         然后你就看到能正常捕获HTTPS请求了:    

  4. resin-pro-4.0.34 服務器在windows环境下的配置

    resin-pro-4.0.34 服務器在windows环境下的配置(轉載请注明作者:icelong) 到caucho網站上http://www.caucho.com/download/下載resin ...

  5. kivy create a package for Android

    Now that you've successfully coded an app. Now you want to deploy it to Android. So now we would nee ...

  6. MapXtreme+Asp.net 动态轨迹

    MapXtreme+Asp.net 动态轨迹(请求大神指点)   功能简介:在MapXtreme+Asp.net的环境下实现轨迹回放功能,经过两天的努力基本实现此功能.但还有部分问题需要解决,求大神们 ...

  7. MySQL 的 phpmyadmin上传大小限制(转)以及 MySQL server has gone away 的解决办法

    phpmyadmin上传大小限制 原帖地址:http://www.hmidc.com/home/news/?13914.html 时间:2011-6-21 11:17:57  作者:红帽之家  来源: ...

  8. ThinkPHP中连接mysql数据库的四种实用和通用的连接方法

    ThinkPHP内置了抽象数据库访问层,把不同的数据库操作封装起来,我们只需要使用公共的Db类进行操作,而无需针对不同的数据库写不同的代码和底层实现,Db类会自动调用相应的数据库适配器来处理.目前的数 ...

  9. UVA 216 - Getting in Line

    216 - Getting in Line Computer networking requires that the computers in the network be linked. This ...

  10. 在OpenStack虚拟机实例中创建swap分区的一种方法

    测试组里一个同学负责MapR的搭建,MapR文档中建议每个节点上至少有24GB的swap分区,不知道MapR为啥会有这种反人类的建议……swap无非就是一块顺序读写的磁盘空间,莫非省着内存不用,用sw ...