Skip to content

Commit

Permalink
Merge branch 'master' of github.com:elastic/kibana into sessions/extr…
Browse files Browse the repository at this point in the history
…act-search-abort-controller
  • Loading branch information
Liza K committed Mar 30, 2021
2 parents 512bd7e + e46a74f commit 765c942
Show file tree
Hide file tree
Showing 203 changed files with 3,718 additions and 2,324 deletions.
1 change: 1 addition & 0 deletions .ci/Jenkinsfile_baseline_capture
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ kibanaPipeline(timeoutMinutes: 210) {
) {
withGcpServiceAccount.fromVaultSecret('secret/kibana-issues/dev/ci-artifacts-key', 'value') {
withEnv([
'BUILD_TS_REFS_DISABLE=false', // disabled in root config so we need to override that here
'BUILD_TS_REFS_CACHE_ENABLE=true',
'BUILD_TS_REFS_CACHE_CAPTURE=true',
'DISABLE_BOOTSTRAP_VALIDATION=true',
Expand Down
3 changes: 3 additions & 0 deletions .ci/packer_cache.sh
Original file line number Diff line number Diff line change
Expand Up @@ -2,5 +2,8 @@

set -e

# cache image used by kibana-load-testing project
docker pull "maven:3.6.3-openjdk-8-slim"

./.ci/packer_cache_for_branch.sh master
./.ci/packer_cache_for_branch.sh 7.x
18 changes: 18 additions & 0 deletions docs/apm/getting-started.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,24 @@
<titleabbrev>Get started</titleabbrev>
++++

// Conditionally display a screenshot or video depending on what the
// current documentation version is.

ifeval::["{is-current-version}"=="true"]
++++
<script type="text/javascript" async src="https://play.vidyard.com/embed/v4.js"></script>
<img
style="width: 100%; margin: auto; display: block;"
class="vidyard-player-embed"
src="https://play.vidyard.com/Y4nE2XLYEk75odbRQmUA3g.jpg"
data-uuid="Y4nE2XLYEk75odbRQmUA3g"
data-v="4"
data-type="inline"
/>
</br>
++++
endif::[]

For a quick, high-level overview of the health and performance of your application,
start with:

Expand Down
26 changes: 23 additions & 3 deletions docs/apm/service-maps.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,14 +8,34 @@ requests per minute, and errors per minute.
If enabled, service maps also integrate with machine learning--for real time health indicators based on anomaly detection scores.
All of these features can help you quickly and visually assess your services' status and health.

// Conditionally display a screenshot or video depending on what the
// current documentation version is.

ifeval::["{is-current-version}"=="true"]
++++
<script type="text/javascript" async src="https://play.vidyard.com/embed/v4.js"></script>
<img
style="width: 100%; margin: auto; display: block;"
class="vidyard-player-embed"
src="https://play.vidyard.com/VH8gKnPE3Z2csACZTCeQrw.jpg"
data-uuid="VH8gKnPE3Z2csACZTCeQrw"
data-v="4"
data-type="inline"
/>
</br>
++++
endif::[]

ifeval::["{is-current-version}"=="false"]
[role="screenshot"]
image::apm/images/service-maps.png[Example view of service maps in the APM app in Kibana]
endif::[]

We currently surface two types of service maps:

* Global: All services instrumented with APM agents and the connections between them are shown.
* Service-specific: Highlight connections for a selected service.

[role="screenshot"]
image::apm/images/service-maps.png[Example view of service maps in the APM app in Kibana]

[float]
[[service-maps-how]]
=== How do service maps work?
Expand Down
5 changes: 0 additions & 5 deletions docs/settings/monitoring-settings.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -37,11 +37,6 @@ For more information, see
monitoring back-end does not run and {kib} stats are not sent to the monitoring
cluster.

a|`monitoring.cluster_alerts.`
`email_notifications.email_address` {ess-icon}
| Specifies the email address where you want to receive cluster alerts.
See <<cluster-alert-email-notifications, email notifications>> for details.

| `monitoring.ui.elasticsearch.hosts`
| Specifies the location of the {es} cluster where your monitoring data is stored.
By default, this is the same as <<elasticsearch-hosts, `elasticsearch.hosts`>>. This setting enables
Expand Down
Binary file modified docs/user/dashboard/images/lens_advanced_1_1_2.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified docs/user/dashboard/images/lens_advanced_2_2_1.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file modified docs/user/dashboard/images/lens_advanced_3_1_1.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
64 changes: 0 additions & 64 deletions docs/user/monitoring/cluster-alerts.asciidoc

This file was deleted.

1 change: 0 additions & 1 deletion docs/user/monitoring/index.asciidoc
Original file line number Diff line number Diff line change
@@ -1,6 +1,5 @@
include::xpack-monitoring.asciidoc[]
include::beats-details.asciidoc[leveloffset=+1]
include::cluster-alerts.asciidoc[leveloffset=+1]
include::elasticsearch-details.asciidoc[leveloffset=+1]
include::kibana-alerts.asciidoc[leveloffset=+1]
include::kibana-details.asciidoc[leveloffset=+1]
Expand Down
73 changes: 69 additions & 4 deletions docs/user/monitoring/kibana-alerts.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@ To review and modify all the available alerts, use
This alert is triggered when a node runs a consistently high CPU load. By
default, the trigger condition is set at 85% or more averaged over the last 5
minutes. The alert is grouped across all the nodes of the cluster by running
checks on a schedule time of 1 minute with a re-notify internal of 1 day.
checks on a schedule time of 1 minute with a re-notify interval of 1 day.

[discrete]
[[kibana-alerts-disk-usage-threshold]]
Expand All @@ -38,7 +38,7 @@ checks on a schedule time of 1 minute with a re-notify internal of 1 day.
This alert is triggered when a node is nearly at disk capacity. By
default, the trigger condition is set at 80% or more averaged over the last 5
minutes. The alert is grouped across all the nodes of the cluster by running
checks on a schedule time of 1 minute with a re-notify internal of 1 day.
checks on a schedule time of 1 minute with a re-notify interval of 1 day.

[discrete]
[[kibana-alerts-jvm-memory-threshold]]
Expand All @@ -47,7 +47,7 @@ checks on a schedule time of 1 minute with a re-notify internal of 1 day.
This alert is triggered when a node runs a consistently high JVM memory usage. By
default, the trigger condition is set at 85% or more averaged over the last 5
minutes. The alert is grouped across all the nodes of the cluster by running
checks on a schedule time of 1 minute with a re-notify internal of 1 day.
checks on a schedule time of 1 minute with a re-notify interval of 1 day.

[discrete]
[[kibana-alerts-missing-monitoring-data]]
Expand All @@ -56,7 +56,72 @@ checks on a schedule time of 1 minute with a re-notify internal of 1 day.
This alert is triggered when any stack products nodes or instances stop sending
monitoring data. By default, the trigger condition is set to missing for 15 minutes
looking back 1 day. The alert is grouped across all the nodes of the cluster by running
checks on a schedule time of 1 minute with a re-notify internal of 6 hours.
checks on a schedule time of 1 minute with a re-notify interval of 6 hours.

[discrete]
[[kibana-alerts-thread-pool-rejections]]
== Thread pool rejections (search/write)

This alert is triggered when a node experiences thread pool rejections. By
default, the trigger condition is set at 300 or more over the last 5
minutes. The alert is grouped across all the nodes of the cluster by running
checks on a schedule time of 1 minute with a re-notify interval of 1 day.
Thresholds can be set independently for `search` and `write` type rejections.

[discrete]
[[kibana-alerts-ccr-read-exceptions]]
== CCR read exceptions

This alert is triggered if a read exception has been detected on any of the
replicated clusters. The trigger condition is met if 1 or more read exceptions
are detected in the last hour. The alert is grouped across all replicated clusters
by running checks on a schedule time of 1 minute with a re-notify interval of 6 hours.

[discrete]
[[kibana-alerts-large-shard-size]]
== Large shard size

This alert is triggered if a large (primary) shard size is found on any of the
specified index patterns. The trigger condition is met if an index's shard size is
55gb or higher in the last 5 minutes. The alert is grouped across all indices that match
the default patter of `*` by running checks on a schedule time of 1 minute with a re-notify
interval of 12 hours.

[discrete]
[[kibana-alerts-cluster-alerts]]
== Cluster alerts

These alerts summarize the current status of your {stack}. You can drill down into the metrics
to view more information about your cluster and specific nodes, instances, and indices.

An alert will be triggered if any of the following conditions are met within the last minute:

* {es} cluster health status is yellow (missing at least one replica)
or red (missing at least one primary).
* {es} version mismatch. You have {es} nodes with
different versions in the same cluster.
* {kib} version mismatch. You have {kib} instances with different
versions running against the same {es} cluster.
* Logstash version mismatch. You have Logstash nodes with different
versions reporting stats to the same monitoring cluster.
* {es} nodes changed. You have {es} nodes that were recently added or removed.
* {es} license expiration. The cluster's license is about to expire.
+
--
If you do not preserve the data directory when upgrading a {kib} or
Logstash node, the instance is assigned a new persistent UUID and shows up
as a new instance
--
* Subscription license expiration. When the expiration date
approaches, you will get notifications with a severity level relative to how
soon the expiration date is:
** 60 days: Informational alert
** 30 days: Low-level alert
** 15 days: Medium-level alert
** 7 days: Severe-level alert
+
The 60-day and 30-day thresholds are skipped for Trial licenses, which are only
valid for 30 days.

NOTE: Some action types are subscription features, while others are free.
For a comparison of the Elastic subscription levels, see the alerting section of
Expand Down
3 changes: 2 additions & 1 deletion package.json
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,7 @@
"@elastic/datemath": "link:packages/elastic-datemath",
"@elastic/elasticsearch": "npm:@elastic/elasticsearch-canary@^8.0.0-canary.4",
"@elastic/ems-client": "7.12.0",
"@elastic/eui": "31.7.0",
"@elastic/eui": "31.10.0",
"@elastic/filesaver": "1.1.2",
"@elastic/good": "^9.0.1-kibana3",
"@elastic/maki": "6.3.0",
Expand Down Expand Up @@ -203,6 +203,7 @@
"content-disposition": "0.5.3",
"copy-to-clipboard": "^3.0.8",
"core-js": "^3.6.5",
"css-minimizer-webpack-plugin": "^1.3.0",
"custom-event-polyfill": "^0.3.0",
"cytoscape": "^3.10.0",
"cytoscape-dagre": "^2.2.2",
Expand Down
Loading

0 comments on commit 765c942

Please sign in to comment.