메뉴 건너뛰기

Cloudera, BigData, Semantic IoT, Hadoop, NoSQL

Cloudera CDH/CDP 및 Hadoop EcoSystem, Semantic IoT등의 개발/운영 기술을 정리합니다. gooper@gooper.com로 문의 주세요.


curl -u 'xxxx':'xxxx' 'http://CM_server.domain.com:7180/api/v11/clusters/cluster/services/impala/impalaQueries?from=2017-10... statement RLIKE ".*fawzea.*"' >>f.json


http://CM-server.domain.com:7180/api/v19/clusters/cluster/services/impala/impalaQueries?statment%20RLIKE%20%22.*test.*%22




*참고 : https://cloudera.github.io/cm_api/apidocs/v19/index.html

Introduction

This document describes the Cloudera Manager REST API. All requests and responses are presented in Javascript Object Notation (JSON).

The API resources listed below follow standard Create-Read-Update-Delete (CRUD) semantics where the HTTP request path defines the entity to be acted on and the HTTP method expresses the type of action to perform.

HTTP MethodOperation
POSTCreate entries
GETRead entries
PUTUpdate or edit entries
DELETEDelete entries

All collections in the API use plural names, 'users', instead of the singular, 'user'. To address a specific user in the system, expand the URL path to include the user identifier. For example, '/users/foo' identifies user 'foo' and '/users/bar' identifies user 'bar'.

CollectionPOST (create)GET (read)PUT (update)DELETE (delete)
/usersCreate a new userList all users in the systemBulk update all usersDelete all users
/users/fooerrorRead information about user 'foo'If user 'foo' exists, update their information; otherwise, error.Delete user 'foo'

Keep in mind that not all collections support all operations. For example, events in the system are read-only; you cannot create new events with the API.

You can list the entries in a collection using one of two views: 'summary' or 'full'. The default 'summary' view provides the core information about each entry. The 'full' view is more heavyweight and provides a fully expanded view of each entry. The view is controlled by a query parameter called 'view' e.g. 'GET /users?view=full'.

HTTP requests should set the "Content-Type" header to "application/json" for most calls, unless specified otherwise.

Authentication

The Cloudera Manager API uses HTTP basic access authentication. It accepts the same user credentials as the web interface. Different users may have different levels of access, as defined by their roles. (See the user management API calls for more.) With every authenticated request, the server returns a session cookie, which can be subsequently used for authentication.

Error Code

The API returns HTTP code in the 400s and 500s upon errors:

  • 400 — Invalid request parameters; Malformed requests
  • 401 — Authentication failure
  • 403 — Authorization failure
  • 404 — Object not found
  • 500 — Internal API error
  • 503 — Response temporarily unavailable; caller should retry later

See /tools/echoError for a discussion on the body of an error response.

API Usage Tutorial

Tutorial link.

API Version

The API can be queried for the latest version it supports. By making a GET request to the '/version' URL, a string will be returned with the highest supported version (for example, "v2").

Note that servers that only support API v1 do not support this call. In that case, a 404 error will be returned, and can be used to detect that the server only supports v1.

The version should be prepended to the path of the REST resources listed below. The documentation of the different resources lists the version which added the particular functionality.

REST Resources

This API supports a Representational State Transfer (REST) model for accessing a set of resources through a fixed set of operations. The following resources are accessible through the RESTful model:

There is also a WADL document describing the REST API.

Data Model

All endpoints act on a common set of data. The Cloudera Manager API uses JSON.

Data Elements

번호 제목 날짜 조회 수
541 spark-shell실행시 "A read-only user or a user in a read-only database is not permitted to disable read-only mode on a connection."오류가 발생하는 경우 해결방법 2016.05.20 625
540 Spark 1.6.1 설치후 HA구성 2016.05.24 725
539 spark-env.sh에서 사용할 수있는 항목. 2016.05.24 853
538 "Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources"오류 발생시 조치사항 2016.05.25 1132
537 spark 온라인 책자링크 (제목 : mastering-apache-spark) 2016.05.25 159
536 spark-submit으로 spark application실행하는 다양한 방법 2016.05.25 384
535 RDF storage조합에대한 test결과(4store, Jena+HBase, Hive+HBase, CumulusRDF, Couchbase) 페이지 링크 2016.05.26 276
534 CentOS6에 python3.5.1 소스코드로 빌드하여 설치하기 2016.05.27 406
533 S2RDF 테스트(벤치마크 테스트를 기준으로 python, scala소스가 만들어져서 기능은 파악되지 못함) [2] file 2016.05.27 133
532 python실행시 ValueError: zero length field name in format오류 해결방법 2016.05.27 465
531 python 2.6.6에서 print 'A=' 형태의 사용이 python 3.5.1에서 오류(SyntaxError: Missing parentheses in call to 'print') 발생함.. 2016.05.27 313
530 DataSetCreator.py 실행시 파일을 찾을 수 없는 오류 2016.05.27 140
529 --master yarn 옵션으로 spark client프로그램 실행할때 메모리 부족 오류발생시 조치방법 2016.05.27 216
528 spark client프로그램 기동시 "Error initializing SparkContext"오류 발생할때 조치사항 2016.05.27 613
527 Job이 끝난 log을 볼수 있도록 설정하기 2016.05.30 651
526 centos에 sbt 0.13.5 설치 2016.05.30 352
525 Scala버젼 변경 혹은 상황에 맞게 Spark소스 컴파일하기 2016.05.31 469
524 "암은 평범한 병, 심호흡만 잘해도 암세포 분열 저지” 2016.06.02 142
523 Windows에서 sbt개발환경 구축 방법(링크) 2016.06.02 172
522 hive 2.0.1 설치및 mariadb로 metastore 설정 2016.06.03 5292
위로