All Projects β†’ opendistro-for-elasticsearch β†’ Performance Analyzer

opendistro-for-elasticsearch / Performance Analyzer

Licence: other
πŸ“ˆ OpenDistro for Elasticsearch Performance Analyzer

Programming Languages

java
68154 projects - #9 most used programming language

Projects that are alternatives of or similar to Performance Analyzer

Elastik Nearest Neighbors
Go to: https://github.com/alexklibisz/elastiknn
Stars: ✭ 249 (+94.53%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Gem
πŸ’Ž GUI for Data Modeling with Elasticsearch
Stars: ✭ 654 (+410.94%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Alerting
πŸ“Ÿ Open Distro for Elasticsearch Alerting Plugin
Stars: ✭ 259 (+102.34%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Graph Aided Search
Elasticsearch plugin offering Neo4j integration for Personalized Search
Stars: ✭ 153 (+19.53%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Syliuselasticsearchplugin
Elasticsearch integration for Sylius apps.
Stars: ✭ 88 (-31.25%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Mirage
🎨 GUI for simplifying Elasticsearch Query DSL
Stars: ✭ 2,143 (+1574.22%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Elasticsearch Hq
Monitoring and Management Web Application for ElasticSearch instances and clusters.
Stars: ✭ 4,832 (+3675%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Kibana
Your window into the Elastic Stack
Stars: ✭ 16,820 (+13040.63%)
Mutual labels:  elasticsearch, observability
Elasticsearch Analysis Hanlp
HanLP Analysis for Elasticsearch
Stars: ✭ 77 (-39.84%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Elasticsearch Learning To Rank
Plugin to integrate Learning to Rank (aka machine learning for better relevance) with Elasticsearch
Stars: ✭ 1,147 (+796.09%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Esparser
PHP write SQL to convert DSL to query Elasticsearch
Stars: ✭ 142 (+10.94%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Elasticsearch Reindexing
Elasticsearch plugin for reindexing
Stars: ✭ 106 (-17.19%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Elastiknn
Elasticsearch plugin for nearest neighbor search. Store vectors and run similarity search using exact and approximate algorithms.
Stars: ✭ 139 (+8.59%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Emoji Search
πŸ˜„ Emoji synonyms to build your own emoji-capable search engine (elasticsearch, solr)
Stars: ✭ 184 (+43.75%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Elasticsearch Dataformat
Excel/CSV/BulkJSON downloads on Elasticsearch.
Stars: ✭ 135 (+5.47%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Elasticsearch Readonlyrest Plugin
Free Elasticsearch security plugin and Kibana security plugin: super-easy Kibana multi-tenancy, Encryption, Authentication, Authorization, Auditing
Stars: ✭ 917 (+616.41%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Zentity
Entity resolution for Elasticsearch.
Stars: ✭ 97 (-24.22%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Elasticsearch Analysis Kuromoji Ipadic Neologd
Elasticsearch's Analyzer for Kuromoji with Neologd
Stars: ✭ 109 (-14.84%)
Mutual labels:  elasticsearch, elasticsearch-plugin
Od Database
Distributed crawler, database and web frontend for public directories indexing
Stars: ✭ 121 (-5.47%)
Mutual labels:  elasticsearch
Transformalize
Configurable Extract, Transform, and Load
Stars: ✭ 125 (-2.34%)
Mutual labels:  elasticsearch

Java CI CD codecov Documentation Chat PRs welcome!

Performance Analyzer

Performance Analyzer exposes a REST API that allows you to query numerous performance metrics for your cluster, including aggregations of those metrics, independent of the Java Virtual Machine (JVM). PerfTop is the default command line interface (CLI) for displaying those metrics.

Setup

Performance Analyzer API

Performance Analyzer uses a single HTTP method and URI for all requests:

GET <endpoint>/_opendistro/_performanceanalyzer/metrics

Then you provide parameters for metrics, aggregations, dimensions, and nodes (optional):

?metrics=<metrics>&agg=<aggregations>&dim=<dimensions>&nodes=all"
  • metrics - comma separated list of metrics you are interested in. For a full list of metrics, see Metrics Reference.
  • agg - comma separated list of agg to be used on each metric. Possible values are sum, avg, min and max. Length of the list should be equal to the number of metrics specified.
  • dim - comma separated list of dimensions. For the list of dimensions supported by each metric, see Metrics Reference.
  • nodes - If the string all is passed, metrics from all nodes in the cluster are returned. For any other value, metrics from only the local node is returned.

SAMPLE REQUEST

GET _opendistro/_performanceanalyzer/metrics?metrics=Latency,CPU_Utilization&agg=avg,max&dim=ShardID&nodes=all

Batch Metrics API

While the metrics api associated with performance analyzer provides the last 5 seconds worth of metrics, the batch metrics api provides more detailed metrics and from longer periods of time. See the design doc for more information.

In order to access the batch metrics api, first enable it using one of the following HTTP request:

POST localhost:9200/_opendistro/performanceanalyzer/batch/config -H β€˜Content-Type: application/json’ -d β€˜{"enabled": true}’
POST localhost:9200/_opendistro/performanceanalyzer/batch/cluster/config -H β€˜Content-Type: application/json’ -d β€˜{"enabled": true}’

The former enables batch metrics on a single node, while the latter enables it on nodes across the entire cluster. Batch metrics can be disabled using analogous queries with {"enabled": false}.

You can then query either the config or cluster config apis to see how many minutes worth of batch metrics data will be retained by nodes in the cluster (batchMetricsRetentionPeriodMinutes):

GET localhost:9200/_opendistro/_performanceanalyzer/config

{"performanceAnalyzerEnabled":true,"rcaEnabled":false,"loggingEnabled":false,"shardsPerCollection":0,"batchMetricsEnabled":true,"batchMetricsRetentionPeriodMinutes":7}

GET localhost:9200/_opendistro/_performanceanalyzer/cluster/config

{"currentPerformanceAnalyzerClusterState":9,"shardsPerCollection":0,"batchMetricsRetentionPeriodMinutes":7}

The default retention period is 7 minutes. However, the cluster owner can adjust this by setting batch-metrics-retention-period-minutes in performance-analyzer.properties (note, setting this value will require a restart so that the cluster can read the new value upon startup). The value must be between 1 and 60 minutes (inclusive) β€” the range is capped like so in order to prevent excessive data retention on the cluster, which would eat up a lot of storage.

You can then access the batch metrics available at each node via queries of the following format:

GET localhost:9600/_opendistro/_performanceanalyzer/batch?metrics=<metrics>&starttime=<starttime>&endtime=<endtime>&samplingperiod=<samplingperiod>
  • metrics - Comma separated list of metrics you are interested in. For a full list of metrics, see Metrics Reference.
  • starttime - Unix timestamp (difference between the current time and midnight, January 1, 1970 UTC) in milliseconds determining the oldest data point to return. starttime is inclusive β€” data points from at or after the starttime will be returned. Note, the starttime and endtime supplied by the user will both be rounded down to the nearest samplingperiod. starttime must be no less than now - retention_period and it must be less than the endtime (after the rounding).
  • endtime - Unix timestamp in milliseconds determining the freshest data point to return. endtime is exclusive β€” only datapoints from before the endtime will be returned. endtime must be no greater than the system time at the node, and it must be greater than the startime (after being rounded down to the nearest samplingperiod).
  • samplingperiod - Optional parameter indicating the sampling period in seconds (default is 5s). The requested time range will be partitioned according to the sampling period, and data from the first available 5s interval in each partition will be returned to the user. Must be at least 5s, must be less than the retention period, and must be a multiple of 5.

Note, the maximum number of datapoints that a single query can request for via API is capped at 100,800 datapoints (in order to prevent excessive memory consumption by the datapoints). If a query exceeds this limit, an error is returned. The query parameters can be adjusted on such queries to request for fewer datapoints at a time.

Note, unlike with the metrics api, there is no nodes=all parameter for the batch metrics api. You must query a specific node in order to obtain metrics from that node.

Note, the default retention period is 7 minutes because a typical use-case would be to query for 5 minutes worth of data from the node. In order to do this, a client would actually select a starttime of now-6min and an endtime of now-1min (this one minute offset will give sufficient time for the metrics in the time range to be available at the node). Atop this 6 minutes of retention, we need an extra 1 minute of retention to account for the time that would have passed by the time the query arrives at the node, and for the fact that starttime and endtime will be rounded down to the nearest samplingperiod.

SAMPLE REQUEST

GET _opendistro/_performanceanalyzer/batch?metrics=CPU_Utilization,IO_TotThroughput&starttime=1594412250000&endtime=1594412260000&samplingperiod=5

See the design doc for the expected response.

Documentation

Please refer to the technical documentation for detailed information on installing and configuring Performance Analyzer.

Code of Conduct

This project has adopted an Open Source Code of Conduct.

Security issue notifications

If you discover a potential security issue in this project we ask that you notify AWS/Amazon Security via our vulnerability reporting page. Please do not create a public GitHub issue.

Licensing

See the LICENSE file for our project's licensing. We will ask you to confirm the licensing of your contribution.

Copyright

Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].