메뉴 건너뛰기

Cloudera, BigData, Semantic IoT, Hadoop, NoSQL

Cloudera CDH/CDP 및 Hadoop EcoSystem, Semantic IoT등의 개발/운영 기술을 정리합니다. gooper@gooper.com로 문의 주세요.


실행 : python3 DataSetCreator.py -i s2rdf/data/sparql.in -s 0.25

=>http://stackoverflow.com/questions/27792839/spark-fail-when-running-pi-py-example-with-yarn-client-mode 참조

-----------------------------로그내용------------------------------
Input RDF file ->"
16/05/27 18:22:57 INFO SparkContext: Running Spark version 1.6.1
16/05/27 18:22:57 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
16/05/27 18:22:57 WARN SparkConf: Detected deprecated memory fraction settings: [spark.storage.memoryFraction]. As of Spark 1.6, execution and storage memory management are unified. All memory fractions used in the old model are now deprecated and no longer read. If you wish to use the old memory management, you may explicitly enable `spark.memory.useLegacyMode` (not recommended).
16/05/27 18:22:57 INFO SecurityManager: Changing view acls to: hadoop
16/05/27 18:22:57 INFO SecurityManager: Changing modify acls to: hadoop
16/05/27 18:22:57 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop); users with modify permissions: Set(hadoop)
16/05/27 18:22:57 INFO Utils: Successfully started service 'sparkDriver' on port 56181.
16/05/27 18:22:58 INFO Slf4jLogger: Slf4jLogger started
16/05/27 18:22:58 INFO Remoting: Starting remoting
16/05/27 18:22:58 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriverActorSystem@XXX.XXX.XXX.43:34384]
16/05/27 18:22:58 INFO Utils: Successfully started service 'sparkDriverActorSystem' on port 34384.
16/05/27 18:22:58 INFO SparkEnv: Registering MapOutputTracker
16/05/27 18:22:58 INFO SparkEnv: Registering BlockManagerMaster
16/05/27 18:22:58 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-cdc351b1-92b1-405c-9127-fca2f798daf3
16/05/27 18:22:58 INFO MemoryStore: MemoryStore started with capacity 1247.3 MB
16/05/27 18:22:58 INFO SparkEnv: Registering OutputCommitCoordinator
16/05/27 18:22:58 INFO Utils: Successfully started service 'SparkUI' on port 4040.
16/05/27 18:22:58 INFO SparkUI: Started SparkUI at http://XXX.XXX.XXX.43:4040
16/05/27 18:22:58 INFO HttpFileServer: HTTP File server directory is /tmp/spark-de18dde4-d74e-4197-beab-2bc3de517b74/httpd-8faa7605-d0e3-44b9-ba73-d18ce63fe8f1
16/05/27 18:22:58 INFO HttpServer: Starting HTTP Server
16/05/27 18:22:58 INFO Utils: Successfully started service 'HTTP file server' on port 49921.
16/05/27 18:22:58 INFO SparkContext: Added JAR file:/home/hadoop/DataSetCreator/./datasetcreator_2.10-1.1.jar at http://XXX.XXX.XXX.43:49921/jars/datasetcreator_2.10-1.1.jar with timestamp 1464340978585
16/05/27 18:22:58 WARN YarnClientSchedulerBackend: NOTE: SPARK_WORKER_CORES is deprecated. Use SPARK_EXECUTOR_CORES or --executor-cores through spark-submit instead.
16/05/27 18:22:58 INFO ConfiguredRMFailoverProxyProvider: Failing over to rm2
16/05/27 18:22:58 INFO Client: Requesting a new application from cluster with 4 NodeManagers
16/05/27 18:22:58 INFO Client: Verifying our application has not requested more than the maximum memory capability of the cluster (19288 MB per container)
16/05/27 18:22:58 INFO Client: Will allocate AM container, with 896 MB memory including 384 MB overhead
16/05/27 18:22:58 INFO Client: Setting up container launch context for our AM
16/05/27 18:22:58 INFO Client: Setting up the launch environment for our AM container
16/05/27 18:22:58 INFO Client: Preparing resources for our AM container
16/05/27 18:22:59 INFO Client: Uploading resource file:/home/gooper/svc/apps/sda/bin/hadoop/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar -> hdfs://mycluster/user/hadoop/.sparkStaging/application_1464337540213_0018/spark-assembly-1.6.1-hadoop2.6.0.jar
16/05/27 18:23:01 INFO Client: Uploading resource file:/tmp/spark-de18dde4-d74e-4197-beab-2bc3de517b74/__spark_conf__2857474168024892319.zip -> hdfs://mycluster/user/hadoop/.sparkStaging/application_1464337540213_0018/__spark_conf__2857474168024892319.zip
16/05/27 18:23:01 INFO SecurityManager: Changing view acls to: hadoop
16/05/27 18:23:01 INFO SecurityManager: Changing modify acls to: hadoop
16/05/27 18:23:01 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop); users with modify permissions: Set(hadoop)
16/05/27 18:23:01 INFO Client: Submitting application 18 to ResourceManager
16/05/27 18:23:01 INFO YarnClientImpl: Submitted application application_1464337540213_0018
16/05/27 18:23:02 INFO Client: Application report for application_1464337540213_0018 (state: ACCEPTED)
16/05/27 18:23:02 INFO Client: 
         client token: N/A
         diagnostics: N/A
         ApplicationMaster host: N/A
         ApplicationMaster RPC port: -1
         queue: root.hadoop
         start time: 1464340977670
         final status: UNDEFINED
         tracking URL: http://sda2:8088/proxy/application_1464337540213_0018/
         user: hadoop
16/05/27 18:23:03 INFO Client: Application report for application_1464337540213_0018 (state: ACCEPTED)
16/05/27 18:23:04 INFO Client: Application report for application_1464337540213_0018 (state: ACCEPTED)
16/05/27 18:23:04 INFO YarnSchedulerBackend$YarnSchedulerEndpoint: ApplicationMaster registered as NettyRpcEndpointRef(null)
16/05/27 18:23:04 INFO YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> sda1, PROXY_URI_BASES -> http://sda1:8088/proxy/application_1464337540213_0018), /proxy/application_1464337540213_0018
16/05/27 18:23:04 INFO JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
16/05/27 18:23:05 INFO Client: Application report for application_1464337540213_0018 (state: RUNNING)
16/05/27 18:23:05 INFO Client: 
         client token: N/A
         diagnostics: N/A
         ApplicationMaster host: XXX.XXX.XXX.44
         ApplicationMaster RPC port: 0
         queue: root.hadoop
         start time: 1464340977670
         final status: UNDEFINED
         tracking URL: http://sda2:8088/proxy/application_1464337540213_0018/
         user: hadoop
16/05/27 18:23:05 INFO YarnClientSchedulerBackend: Application application_1464337540213_0018 has started running.
16/05/27 18:23:05 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 44676.
16/05/27 18:23:05 INFO NettyBlockTransferService: Server created on 44676
16/05/27 18:23:05 INFO BlockManagerMaster: Trying to register BlockManager
16/05/27 18:23:05 INFO BlockManagerMasterEndpoint: Registering block manager XXX.XXX.XXX.43:44676 with 1247.3 MB RAM, BlockManagerId(driver, XXX.XXX.XXX.43, 44676)
16/05/27 18:23:05 INFO BlockManagerMaster: Registered BlockManager
16/05/27 18:23:05 INFO EventLoggingListener: Logging events to hdfs://mycluster/user/hadoop/spark/application_1464337540213_0018
16/05/27 18:23:08 INFO YarnSchedulerBackend$YarnSchedulerEndpoint: ApplicationMaster registered as NettyRpcEndpointRef(null)
16/05/27 18:23:08 INFO YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> sda1, PROXY_URI_BASES -> http://sda1:8088/proxy/application_1464337540213_0018), /proxy/application_1464337540213_0018
16/05/27 18:23:08 INFO JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
16/05/27 18:23:09 ERROR YarnClientSchedulerBackend: Yarn application has already exited with state FINISHED!
16/05/27 18:23:09 INFO SparkUI: Stopped Spark web UI at http://XXX.XXX.XXX.43:4040
16/05/27 18:23:09 INFO YarnClientSchedulerBackend: Shutting down all executors
16/05/27 18:23:09 INFO YarnClientSchedulerBackend: Asking each executor to shut down
16/05/27 18:23:09 INFO YarnClientSchedulerBackend: Stopped
16/05/27 18:23:09 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
16/05/27 18:23:09 INFO MemoryStore: MemoryStore cleared
16/05/27 18:23:09 INFO BlockManager: BlockManager stopped
16/05/27 18:23:09 INFO BlockManagerMaster: BlockManagerMaster stopped
16/05/27 18:23:09 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
16/05/27 18:23:09 INFO SparkContext: Successfully stopped SparkContext
16/05/27 18:23:09 INFO RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon.
16/05/27 18:23:09 INFO RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports.
16/05/27 18:23:09 INFO RemoteActorRefProvider$RemotingTerminator: Remoting shut down.
16/05/27 18:23:28 INFO YarnClientSchedulerBackend: SchedulerBackend is ready for scheduling beginning after waiting maxRegisteredResourcesWaitingTime: 30000(ms)
16/05/27 18:23:28 ERROR SparkContext: Error initializing SparkContext.
java.lang.NullPointerException
        at org.apache.spark.SparkContext.<init>(SparkContext.scala:584)
        at dataCreator.Settings$.loadSparkContext(Settings.scala:69)
        at dataCreator.Settings$.<init>(Settings.scala:17)
        at dataCreator.Settings$.<clinit>(Settings.scala)
        at runDriver$.main(runDriver.scala:12)
        at runDriver.main(runDriver.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:497)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
16/05/27 18:23:28 INFO SparkContext: SparkContext already stopped.
Exception in thread "main" java.lang.ExceptionInInitializerError
        at runDriver$.main(runDriver.scala:12)
        at runDriver.main(runDriver.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:497)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.NullPointerException
        at org.apache.spark.SparkContext.<init>(SparkContext.scala:584)
        at dataCreator.Settings$.loadSparkContext(Settings.scala:69)
        at dataCreator.Settings$.<init>(Settings.scala:17)
        at dataCreator.Settings$.<clinit>(Settings.scala)
        ... 11 more
16/05/27 18:23:28 INFO ShutdownHookManager: Shutdown hook called
16/05/27 18:23:28 INFO ShutdownHookManager: Deleting directory /tmp/spark-de18dde4-d74e-4197-beab-2bc3de517b74/httpd-8faa7605-d0e3-44b9-ba73-d18ce63fe8f1
16/05/27 18:23:28 INFO ShutdownHookManager: Deleting directory /tmp/spark-de18dde4-d74e-4197-beab-2bc3de517b74



^CTraceback (most recent call last):
  File "DataSetCreator.py", line 128, in <module>
    main(sys.argv[1:])
  File "DataSetCreator.py", line 125, in main
    generateDatsets()
  File "DataSetCreator.py", line 83, in generateDatsets
    delay()
  File "DataSetCreator.py", line 45, in delay
    time.sleep(delTime)
KeyboardInterrupt
번호 제목 날짜 조회 수
221 Flume을 이용한 데이타 수집시 HBase write 성능 튜닝 file 2016.10.31 724
220 kudu table와 impala(hive) table정보가 틀어져서 테이블을 읽지 못하는 경우(Error Loading Metadata) 조치방법 2023.11.10 724
219 solr에서 한글사용시 주의점 2014.09.26 731
218 springframework를 이용한 war를 생성하는 build.gradle파일(참고용) 2016.08.19 732
217 Spark 1.6.1 설치후 HA구성 2016.05.24 734
216 Halyard - RDF4J와 Apache HBase를 이용하여 구현된 TripleStore이며 SPARQL 1.1쿼리를 지원한다. 2016.12.29 734
215 [SBT] assembly시 "[error] deduplicate: different file contents found in the following:"오류 발생시 조치사항 2016.08.04 737
214 Failed to write to server: (no server available): 2022.01.17 738
213 znode /hbase recursive하게 지우기 2015.05.06 739
212 fuseki용 config-examples.ttl 예시 내용 2017.05.17 746
211 서버 5대에 solr 5.5.0 설치하고 index data를 HDFS에 저장/search하도록 설치/설정하는 방법(SolrCloud) 2016.04.08 749
210 CentOS 7.x에 Jupyter설치 2018.04.18 754
209 sendmail전송시 421 4.3.0 collect: Cannot write ./dfv5BA2EBS010579 (bfcommit, uid=0, gid=114): No such file or directory 발생시 조치사항 2017.06.11 756
208 java.lang.OutOfMemoryError: unable to create new native thread오류 발생지 조치사항 2016.10.17 757
207 solr 인스턴스 기동후 shard에 서버가 정상적으로 할당되지 않는 경우 해결책 2016.04.29 770
206 lateral view 예제 2014.09.18 775
205 [Sentry]HDFS의 ACL을 Sentry와 연동후 테스트 2020.06.02 777
204 hue.desktop_document2의 type의 종류 2020.02.10 779
203 hortonworks에서 제공하는 메모리 설정값 계산기 사용법 file 2015.06.14 792
202 SASL configuration failed: javax.security.auth.login.LoginException: java.lang.NullPointerException 오류 해결방법 2015.04.02 803
위로