메뉴 건너뛰기

Cloudera, BigData, Semantic IoT, Hadoop, NoSQL

Cloudera CDH/CDP 및 Hadoop EcoSystem, Semantic IoT등의 개발/운영 기술을 정리합니다. gooper@gooper.com로 문의 주세요.


*출처1 : https://www.cloudera.com/more/training/certification/cca-spark.html




*출처2 : http://www.hadoopexam.com/Cloudera_Certification/CCA175/CCA175_Cloudera_Hadoop_and_Spark_Developer_Tips_and_Tricks.pdf

1. Preparation: I have gone through all the CCA175 Questions and practice the code provided by
http://www.HadoopExam.com Thanks for your questions and code content. The content was
excellent and it helped me a lot. (Especially I have gone through all the Spark Professional
training module as well)
2. No. Of Questions: Generally you will get 10 questions in real exam: Topic will be coverings are
Sqoop, Hive, Pyspark and Scala and avro-tools to extract schema (All questions are covered in
CCA175 Certification Simulator).
3. Code Snippets: will be provided for Pyspark and Scala. You have to edit the snippets accordingly
as per the problem statement.
4. Real Exam Environment: Gateway node will be accessible for execution of the problems during
the exam. Keep in mind there will not be any on-screen timer available during the exam. You
have to keep asking for the time left. There are three sections for each problem i.e.
· Instructions
· Data Set
· Output Requirements.
Please go through all the three sections carefully before start developing the code.
Note: If you started developing code right after looking at the Instruction part of the question,
then later you will realized the exact details of the table like name of the table and HDFS
directory are also mentioned. This can waste your time if have to redo the code or might as well
cost you a question.
5. Editor: nano, gedit are not available. So if you have to edit any code snippets, you have to use vi
alone. Please make yourself familiar with vi editor if you are not.
6. Fill in blanks: You dont have to write entire code for Python and Scala for Apache Spark,
generally they will ask you to do fill in the blanks.
7. Flume: Very few questions on flume.
8. Difficulty Level: If you have enough knowledge, you will feel exam is quite easy. The questions
were logically easy and can be answered in the first attempt if you read the question carefully
(all three sections).
9. Common mistake in Sqoop: People use connector as localhost which is wrong, you have to use
full name instead of localhost (Avoid wasting your time). Use given hostname
10. Hive: Have initial knowledge of hive as well.
11. Spark: Using basic transform functions to get desired output. For instance filter according
particular scenario, sorting and ranking etc.
12. Avro-tool : avro-tool to get schema of avro file. (Very  nicely covered in CCA175
HadoopExam.com Simulator)
13. Big Mistake: Avoid accidently deleting your data: good practice is necessary to avoid such
mistakes. (Once you delete or drop hive table, you have to create it entirely once again.) Same is
instructed by www.HadoopExam.com during their videos  session provided at
http://cca175cloudera.training4exam.com/ (Please go through sample sessions)
14. Spark-sql: They will not ask questions based on Spark Sql learn importantly aggregate, reduce,
sort.
15. Time management: It is very important, (That’s the reason you need too much practice, use
CCA175 simulator to practice all the questions at least a week or two before your real exam).
16. Data sets in real exam is quite larger, hence it will take 2 to 5 mins for execution.

17. Attempts: try to attempt all questions at least 9/10, hence you must be able to score 70%.
18. File format: In most of questions there was tab delimited file to process.
19. Python or Scala: You will get a preloaded python or scala file to work with, so you don't have a
choice whether you want to attempt a question via scala or pyspark. (I have gone through all the
Video sessions provided by www.HadoopExam.com here
20. Connection Issue: If you got disconnected during exam, you may need to contact the proctor
immediately. If he/she is not available log back into examslocal.com and use their online help.
21. Shell scripts: Have good experience to use shell scripts.
22. Question types as mentioned in syllabus : Questions were from Sqoop(import and export),
Hive(table creation and dynamic partitioning), Pyspark and Scala(Joining, sorting and filtering
data), avro-tools. Snippets of code will be provided for Pyspark and Scala. You have to edit the
snippets accordingly as per the problem statement and can the script file(which is another file
apart from snippet) to get the results.
23. Overall exam is easy, but require lot of practice to complete on time and for accurate
solutions of the problem. Hence go through the all below material for CCA175 (It will not take
more than a month, if you are new and already know the Spark and Hadoop then 2-3 weeks
are good enough.
· CCA175 : Hadoop and Spark Developer Certification practice questions
· Hadoop professional training
· Spark professional training.

Wish you all the best

번호 제목 날짜 조회 수
390 우분투 16.04LTS에 Jupyter설치 2018.04.17 3957
389 spark 시동중 applicationHistory 로그 디렉토리가 없다고 하면서 기동되지 않는 경우 2018.06.01 3958
388 CDH에서 Sentry 개념및 설정 file 2018.06.21 3962
387 Cloudera Manager web UI의 언어를 한글에서 영문으로 변경하기 2018.04.03 3965
386 cloudera-scm-agent 설정파일 위치및 재시작 명령문 2018.03.29 3966
385 Cloudera Manager에서 "Mismatched CDH versions: host has NONE but role expects 5 Suppress..."와 같이 오류 발생시 확인사항 2018.06.06 3976
384 sendmail전송시 421 4.3.0 collect: Cannot write ./dfv5BA2EBS010579 (bfcommit, uid=0, gid=114): No such file or directory 발생시 조치사항 2017.06.11 3977
383 embedded-cassandra의 data 저장위치 2019.06.09 3983
382 "You are running Cloudera Manager in non-production mode.." warning메세지가 나타나지 않게 조치하는 방법 2018.05.23 3984
381 [shellscript] 함수에 배열을 인자로 주어서 처리하는 방법 2019.07.16 3985
380 "bad handshake: Error([('SSL routines', 'ssl3_get_server_certificate', 'certificate verify failed')])" 오류는 CA인증을 하지 못해서 발생함 2022.05.13 3990
379 SPIN(SPARQL Inference Notation)이란.. file 2016.02.25 3992
378 [개발] 온라인 IDE - 개발 환경 구축 없어 어디서나 웹브라우저로 개발하기 2022.05.02 3998
377 Hue Load Balancer를 L4로 L/B하는 경우는 L4쪽 도멘인으로 발행된 인증서를 TLS/SSL항목에 설정해주어야 한다. 2021.10.08 4000
376 Hadoop - 클러스터 세팅및 기동 2015.04.28 4001
375 Scala버젼 변경 혹은 상황에 맞게 Spark소스 컴파일하기 2016.05.31 4005
374 kafka의 re-balance를 이용하여 consumer를 multi thread로 돌려서 topic의 partitions을 활용 2015.03.31 4015
373 MongoDB에 있는 특정컬럼의 값을 casting(string->integer)하여 update하기 java 소스 2016.12.19 4016
372 Oracle RAC 구성된 DB서버에 대한 컴포넌트별 설정 방법 2022.02.12 4021
371 solr 인스턴스 기동후 shard에 서버가 정상적으로 할당되지 않는 경우 해결책 2016.04.29 4035
위로