메뉴 건너뛰기

Cloudera, BigData, Semantic IoT, Hadoop, NoSQL

Cloudera CDH/CDP 및 Hadoop EcoSystem, Semantic IoT등의 개발/운영 기술을 정리합니다. gooper@gooper.com로 문의 주세요.


hive-site.xml에서 아래 3개의 <value>에서 ms를 뺀다.


 <property>

    <name>hive.hmshandler.retry.interval</name>

    <value>2000ms</value>

    <description>

      Expects a time value with unit (d/day, h/hour, m/min, s/sec, ms/msec, us/usec, ns/nsec), which is msec if not specified.

      The time between HMSHandler retry attempts on failure.

    </description>

  </property>


  <property>

    <name>hive.llap.am.liveness.connection.sleep.between.retries.ms</name>

    <value>2000ms</value>

    <description>

      Expects a time value with unit (d/day, h/hour, m/min, s/sec, ms/msec, us/usec, ns/nsec), which is msec if not specified.

      Sleep duration while waiting to retry connection failures to the AM from the daemon for

      the general keep-alive thread (milliseconds).

    </description>

  </property>



  <property>

    <name>hive.llap.task.communicator.connection.sleep.between.retries.ms</name>

    <value>2000ms</value>

    <description>

      Expects a time value with unit (d/day, h/hour, m/min, s/sec, ms/msec, us/usec, ns/nsec), which is msec if not specified.

      Sleep duration (in milliseconds) to wait before retrying on error when obtaining a

      connection to LLAP daemon from Tez AM.

    </description>

  </property>




---------------------------------오류내용-----------------------------------------

16/06/09 13:57:38 INFO SparkDeploySchedulerBackend: SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.0

SET spark.sql.hive.version=0.13.1

16/06/09 13:57:39 ERROR log: Got exception: java.lang.NumberFormatException For input string: "2000ms"

java.lang.NumberFormatException: For input string: "2000ms"

        at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)

        at java.lang.Integer.parseInt(Integer.java:580)

        at java.lang.Integer.parseInt(Integer.java:615)

        at org.apache.hadoop.conf.Configuration.getInt(Configuration.java:1134)

        at org.apache.hadoop.hive.conf.HiveConf.getIntVar(HiveConf.java:1211)

        at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:87)

        at com.sun.proxy.$Proxy6.get_all_databases(Unknown Source)

        at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getAllDatabases(HiveMetaStoreClient.java:837)

        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)

        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

        at java.lang.reflect.Method.invoke(Method.java:497)

        at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:89)

        at com.sun.proxy.$Proxy7.getAllDatabases(Unknown Source)

        at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1098)

        at org.apache.hadoop.hive.ql.exec.FunctionRegistry.getFunctionNames(FunctionRegistry.java:671)

        at org.apache.hadoop.hive.ql.exec.FunctionRegistry.getFunctionNames(FunctionRegistry.java:662)

        at org.apache.hadoop.hive.cli.CliDriver.getCommandCompletor(CliDriver.java:540)

        at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:168)

        at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)

        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)

        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

        at java.lang.reflect.Method.invoke(Method.java:497)

        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:569)

        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:166)

        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:189)

        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:110)

        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)

16/06/09 13:57:39 ERROR log: Converting exception to MetaException

16/06/09 13:57:39 ERROR FunctionRegistry: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Got exception: java.lang.NumberFormatException For input string: "2000ms")

spark-sql> 16/06/09 13:57:39 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkExecutor@sda2:47136/user/Executor#-853352827] with ID 2

16/06/09 13:57:39 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkExecutor@so2:35213/user/Executor#-894168911] with ID 0

16/06/09 13:57:39 INFO BlockManagerMasterActor: Registering block manager sda2:58779 with 265.1 MB RAM, BlockManagerId(2, sda2, 58779)

16/06/09 13:57:39 INFO BlockManagerMasterActor: Registering block manager so2:33469 with 265.1 MB RAM, BlockManagerId(0, so2, 33469)

16/06/09 13:57:41 INFO SparkDeploySchedulerBackend: Registered executor: Actor[akka.tcp://sparkExecutor@so-db2:35159/user/Executor#-473750792] with ID 1

16/06/09 13:57:41 INFO BlockManagerMasterActor: Registering block manager so-db2:47990 with 265.1 MB RAM, BlockManagerId(1, so-db2, 47990)

번호 제목 날짜 조회 수
142 kafkaWordCount.scala의 producer와 consumer 클래스를 이용하여 kafka를 이용한 word count 테스트 하기 2016.08.02 587
141 bin/start-hbase.sh실행시 org.apache.hadoop.hbase.util.FileSystemVersionException: HBase file layout needs to be upgraded오류가 발생하면 조치사항 2016.08.01 559
140 start-all.sh로 spark데몬 기동시 "JAVA_HOME is not set"오류 발생시 조치사항 2016.08.01 1255
139 hadoop클러스터를 구성하던 서버중 HA를 담당하는 서버의 hostname등이 변경되어 문제가 발생했을때 조치사항 2016.07.29 468
138 Journal Storage Directory /data/hadoop/journal/data/mycluster not formatted 오류시 조치사항 2016.07.29 1750
137 Apache Spark와 Drools를 이용한 CEP구현 테스트 2016.07.15 1157
136 org.apache.hadoop.hbase.ClockOutOfSyncException: org.apache.hadoop.hbase.ClockOutOfSyncException 오류시 조치사항 2016.07.14 255
135 kafka로 부터 메세지를 stream으로 받아 처리하는 spark샘플소스(spark의 producer와 consumer를 sbt로 컴파일 하고 서버에서 spark-submit하는 방법) 2016.07.13 813
134 avro 사용하기(avsc 스키마 파일 컴파일 방법, consumer, producer샘플소스) 2016.07.08 1751
» spark-sql실행시 ERROR log: Got exception: java.lang.NumberFormatException For input string: "2000ms" 오류발생시 조치사항 2016.06.09 409
132 spark-sql실행시 Caused by: java.lang.NumberFormatException: For input string: "0s" 오류발생시 조치사항 2016.06.09 4900
131 spark-sql실행시 The specified datastore driver ("com.mysql.jdbc.Driver") was not found in the CLASSPATH오류 발생시 조치사항 2016.06.09 1649
130 ./spark-sql 실행시 "java.lang.NumberFormatException: For input string: "1s"오류발생시 조치사항 2016.06.09 655
129 beeline실행시 User: root is not allowed to impersonate오류 발생시 조치사항 2016.06.03 930
128 Caused by: java.net.URISyntaxException: Relative path in absolute URI: ${system:java.io.tmpdir%7D/$%7Bsystem:user.name%7D오류발생시 조치사항 2016.06.03 1323
127 impala 설치/설정 2016.06.03 1313
126 hive 2.0.1 설치및 mariadb로 metastore 설정 2016.06.03 5396
125 Scala버젼 변경 혹은 상황에 맞게 Spark소스 컴파일하기 2016.05.31 913
124 Job이 끝난 log을 볼수 있도록 설정하기 2016.05.30 1129
123 spark client프로그램 기동시 "Error initializing SparkContext"오류 발생할때 조치사항 2016.05.27 698
위로