메뉴 건너뛰기

Cloudera, BigData, Semantic IoT, Hadoop, NoSQL

Cloudera CDH/CDP 및 Hadoop EcoSystem, Semantic IoT등의 개발/운영 기술을 정리합니다. gooper@gooper.com로 문의 주세요.


curl -u 'xxxx':'xxxx' 'http://CM_server.domain.com:7180/api/v11/clusters/cluster/services/impala/impalaQueries?from=2017-10... statement RLIKE ".*fawzea.*"' >>f.json


http://CM-server.domain.com:7180/api/v19/clusters/cluster/services/impala/impalaQueries?statment%20RLIKE%20%22.*test.*%22




*참고 : https://cloudera.github.io/cm_api/apidocs/v19/index.html

Introduction

This document describes the Cloudera Manager REST API. All requests and responses are presented in Javascript Object Notation (JSON).

The API resources listed below follow standard Create-Read-Update-Delete (CRUD) semantics where the HTTP request path defines the entity to be acted on and the HTTP method expresses the type of action to perform.

HTTP MethodOperation
POSTCreate entries
GETRead entries
PUTUpdate or edit entries
DELETEDelete entries

All collections in the API use plural names, 'users', instead of the singular, 'user'. To address a specific user in the system, expand the URL path to include the user identifier. For example, '/users/foo' identifies user 'foo' and '/users/bar' identifies user 'bar'.

CollectionPOST (create)GET (read)PUT (update)DELETE (delete)
/usersCreate a new userList all users in the systemBulk update all usersDelete all users
/users/fooerrorRead information about user 'foo'If user 'foo' exists, update their information; otherwise, error.Delete user 'foo'

Keep in mind that not all collections support all operations. For example, events in the system are read-only; you cannot create new events with the API.

You can list the entries in a collection using one of two views: 'summary' or 'full'. The default 'summary' view provides the core information about each entry. The 'full' view is more heavyweight and provides a fully expanded view of each entry. The view is controlled by a query parameter called 'view' e.g. 'GET /users?view=full'.

HTTP requests should set the "Content-Type" header to "application/json" for most calls, unless specified otherwise.

Authentication

The Cloudera Manager API uses HTTP basic access authentication. It accepts the same user credentials as the web interface. Different users may have different levels of access, as defined by their roles. (See the user management API calls for more.) With every authenticated request, the server returns a session cookie, which can be subsequently used for authentication.

Error Code

The API returns HTTP code in the 400s and 500s upon errors:

  • 400 — Invalid request parameters; Malformed requests
  • 401 — Authentication failure
  • 403 — Authorization failure
  • 404 — Object not found
  • 500 — Internal API error
  • 503 — Response temporarily unavailable; caller should retry later

See /tools/echoError for a discussion on the body of an error response.

API Usage Tutorial

Tutorial link.

API Version

The API can be queried for the latest version it supports. By making a GET request to the '/version' URL, a string will be returned with the highest supported version (for example, "v2").

Note that servers that only support API v1 do not support this call. In that case, a 404 error will be returned, and can be used to detect that the server only supports v1.

The version should be prepended to the path of the REST resources listed below. The documentation of the different resources lists the version which added the particular functionality.

REST Resources

This API supports a Representational State Transfer (REST) model for accessing a set of resources through a fixed set of operations. The following resources are accessible through the RESTful model:

There is also a WADL document describing the REST API.

Data Model

All endpoints act on a common set of data. The Cloudera Manager API uses JSON.

Data Elements

번호 제목 날짜 조회 수
550 spark-shell실행시 "A read-only user or a user in a read-only database is not permitted to disable read-only mode on a connection."오류가 발생하는 경우 해결방법 2016.05.20 2953
549 Spark 1.6.1 설치후 HA구성 2016.05.24 4440
548 spark-env.sh에서 사용할 수있는 항목. 2016.05.24 4295
547 "Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources"오류 발생시 조치사항 2016.05.25 4563
546 spark 온라인 책자링크 (제목 : mastering-apache-spark) 2016.05.25 4248
545 spark-submit으로 spark application실행하는 다양한 방법 2016.05.25 4408
544 RDF storage조합에대한 test결과(4store, Jena+HBase, Hive+HBase, CumulusRDF, Couchbase) 페이지 링크 2016.05.26 3018
543 CentOS6에 python3.5.1 소스코드로 빌드하여 설치하기 2016.05.27 4112
542 S2RDF 테스트(벤치마크 테스트를 기준으로 python, scala소스가 만들어져서 기능은 파악되지 못함) [1] file 2016.05.27 2864
541 python실행시 ValueError: zero length field name in format오류 해결방법 2016.05.27 3929
540 python 2.6.6에서 print 'A=' 형태의 사용이 python 3.5.1에서 오류(SyntaxError: Missing parentheses in call to 'print') 발생함.. 2016.05.27 2930
539 DataSetCreator.py 실행시 파일을 찾을 수 없는 오류 2016.05.27 2129
538 --master yarn 옵션으로 spark client프로그램 실행할때 메모리 부족 오류발생시 조치방법 2016.05.27 3522
537 spark client프로그램 기동시 "Error initializing SparkContext"오류 발생할때 조치사항 2016.05.27 4646
536 Job이 끝난 log을 볼수 있도록 설정하기 2016.05.30 4418
535 centos에 sbt 0.13.5 설치 2016.05.30 3201
534 Scala버젼 변경 혹은 상황에 맞게 Spark소스 컴파일하기 2016.05.31 4005
533 "암은 평범한 병, 심호흡만 잘해도 암세포 분열 저지” 2016.06.02 2814
532 Windows에서 sbt개발환경 구축 방법(링크) 2016.06.02 2087
531 hive 2.0.1 설치및 mariadb로 metastore 설정 2016.06.03 9069
위로