Skip to content

Commit

Permalink
Merge branch 'master' into nls/indices-stats
Browse files Browse the repository at this point in the history
  • Loading branch information
elasticmachine authored Jul 26, 2020
2 parents a2264ce + ff9f06b commit e983bf8
Show file tree
Hide file tree
Showing 1,351 changed files with 43,397 additions and 16,679 deletions.
6 changes: 6 additions & 0 deletions .ci/Jenkinsfile_baseline_trigger
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,12 @@ kibanaLibrary.load()

withGithubCredentials {
branches.each { branch ->
if (branch == '6.8') {
// skip 6.8, it is tracked but we don't need snapshots for it and haven't backported
// the baseline capture scripts to it.
return;
}

stage(branch) {
def commits = getCommits(branch, MAXIMUM_COMMITS_TO_CHECK, MAXIMUM_COMMITS_TO_BUILD)

Expand Down
2 changes: 1 addition & 1 deletion .github/CODEOWNERS
Validating CODEOWNERS rules …
Original file line number Diff line number Diff line change
Expand Up @@ -115,7 +115,6 @@
/src/dev/ @elastic/kibana-operations
/src/setup_node_env/ @elastic/kibana-operations
/src/optimize/ @elastic/kibana-operations
/src/es_archiver/ @elastic/kibana-operations
/packages/*eslint*/ @elastic/kibana-operations
/packages/*babel*/ @elastic/kibana-operations
/packages/kbn-dev-utils*/ @elastic/kibana-operations
Expand All @@ -124,6 +123,7 @@
/packages/kbn-pm/ @elastic/kibana-operations
/packages/kbn-test/ @elastic/kibana-operations
/packages/kbn-ui-shared-deps/ @elastic/kibana-operations
/packages/kbn-es-archiver/ @elastic/kibana-operations
/src/legacy/server/keystore/ @elastic/kibana-operations
/src/legacy/server/pid/ @elastic/kibana-operations
/src/legacy/server/sass/ @elastic/kibana-operations
Expand Down
5 changes: 0 additions & 5 deletions docs/developer/architecture/code-exploration.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -186,11 +186,6 @@ WARNING: Missing README.
Replaces the legacy ui/share module for registering share context menus.
- {kib-repo}blob/{branch}/src/plugins/status_page[statusPage]
WARNING: Missing README.
- {kib-repo}blob/{branch}/src/plugins/telemetry/README.md[telemetry]
Telemetry allows Kibana features to have usage tracked in the wild. The general term "telemetry" refers to multiple things:
Expand Down
11 changes: 6 additions & 5 deletions docs/management/alerting/alerts-and-actions-intro.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -4,22 +4,23 @@

beta[]

The *Alerts and Actions* UI lets you <<alert-management, see and control all the alerts>> in a space, and provides tools to <<connector-management, create and manage connectors>> so that alerts can trigger actions like notification, indexing, and ticketing.
The *Alerts and Actions* UI lets you <<alert-management, see and control all the alerts>> in a space, and provides tools to <<connector-management, create and manage connectors>> so that alerts can trigger actions like notification, indexing, and ticketing.

To manage alerting and connectors, open the menu, then go to *Stack Management > {kib} > Alerts and Actions*.
To manage alerting and connectors, open the menu,
then go to *Stack Management > Alerts and Insights > Alerts and Actions*.

[role="screenshot"]
image:management/alerting/images/alerts-and-actions-ui.png[Example alert listing in the Alerts and Actions UI]

[NOTE]
============================================================================
Similar to dashboards, alerts and connectors reside in a <<xpack-spaces, space>>.
The *Alerts and Actions* UI only shows alerts and connectors for the current space.
The *Alerts and Actions* UI only shows alerts and connectors for the current space.
============================================================================

[NOTE]
============================================================================
{es} also offers alerting capabilities through Watcher, which
can be managed through the <<watcher-ui, Watcher UI>>. See
can be managed through the <<watcher-ui, Watcher UI>>. See
<<alerting-concepts-differences>> for more information.
============================================================================
============================================================================
Original file line number Diff line number Diff line change
Expand Up @@ -2,17 +2,16 @@
[[adding-policy-to-index]]
=== Adding a policy to an index

To add a lifecycle policy to an index and view the status for indices
managed by a policy, open the menu, then go to *Stack Management > {es} > Index Management*. This page lists your
{es} indices, which you can filter by lifecycle status and lifecycle phase.
To add a lifecycle policy to an index and view the status for indices
managed by a policy, open the menu, then go to *Stack Management > Data > Index Management*.
This page lists your
{es} indices, which you can filter by lifecycle status and lifecycle phase.

To add a policy, select the index name and then select *Manage Index > Add lifecycle policy*.
You’ll see the policy name, the phase the index is in, the current
action, and if any errors occurred performing that action.
You’ll see the policy name, the phase the index is in, the current
action, and if any errors occurred performing that action.

To remove a policy from an index, select *Manage Index > Remove lifecycle policy*.

[role="screenshot"]
image::images/index_management_add_policy.png[][UI for adding a policy to an index]


Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
[[example-using-index-lifecycle-policy]]
=== Tutorial: Use {ilm-init} to manage {filebeat} time-based indices

With {ilm} ({ilm-init}), you can create policies that perform actions automatically
With {ilm} ({ilm-init}), you can create policies that perform actions automatically
on indices as they age and grow. {ilm-init} policies help you to manage
performance, resilience, and retention of your data during its lifecycle. This tutorial shows
you how to use {kib}’s *Index Lifecycle Policies* to modify and create {ilm-init}
Expand Down Expand Up @@ -59,7 +59,7 @@ output as described in {filebeat-ref}/filebeat-getting-started.html[Getting Star
{filebeat} includes a default {ilm-init} policy that enables rollover. {ilm-init}
is enabled automatically if you’re using the default `filebeat.yml` and index template.

To view the default policy in {kib}, open the menu, go to * Stack Management > {es} > Index Lifecycle Policies*,
To view the default policy in {kib}, open the menu, go to *Stack Management > Data > Index Lifecycle Policies*,
search for _filebeat_, and choose the _filebeat-version_ policy.

This policy initiates the rollover action when the index size reaches 50GB or
Expand Down Expand Up @@ -114,7 +114,7 @@ If meeting a specific retention time period is most important, you can create a
custom policy. For this option, you will use {filebeat} daily indices without
rollover.

. To create a custom policy, open the menu, go to *Stack Management > {es} > Index Lifecycle Policies*, then click
. To create a custom policy, open the menu, go to *Stack Management > Data > Index Lifecycle Policies*, then click
*Create policy*.

. Activate the warm phase and configure it as follows:
Expand Down Expand Up @@ -156,7 +156,7 @@ image::images/tutorial-ilm-custom-policy.png["Modify the custom policy to add a
[role="screenshot"]
image::images/tutorial-ilm-delete-phase-creation.png["Delete phase"]

. To configure the index to use the new policy, open the menu, then go to *Stack Management > {es} > Index Lifecycle
. To configure the index to use the new policy, open the menu, then go to *Stack Management > Data > Index Lifecycle
Policies*.

.. Find your {ilm-init} policy.
Expand Down
4 changes: 2 additions & 2 deletions docs/management/ingest-pipelines/ingest-pipelines.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ pipelines that perform common transformations and
enrichments on your data. For example, you might remove a field,
rename an existing field, or set a new field.

You’ll find *Ingest Node Pipelines* in *Management > Elasticsearch*. With this feature, you can:
You’ll find *Ingest Node Pipelines* in *Stack Management > Ingest*. With this feature, you can:

* View a list of your pipelines and drill down into details.
* Create a pipeline that defines a series of tasks, known as processors.
Expand All @@ -23,7 +23,7 @@ image:management/ingest-pipelines/images/ingest-pipeline-list.png["Ingest node p
The minimum required permissions to access *Ingest Node Pipelines* are
the `manage_pipeline` and `cluster:monitor/nodes/info` cluster privileges.

You can add these privileges in *Management > Security > Roles*.
You can add these privileges in *Stack Management > Security > Roles*.

[role="screenshot"]
image:management/ingest-pipelines/images/ingest-pipeline-privileges.png["Privileges required for Ingest Node Pipelines"]
Expand Down
7 changes: 4 additions & 3 deletions docs/management/managing-beats.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,8 @@

include::{asciidoc-dir}/../../shared/discontinued.asciidoc[tag=cm-discontinued]

To use the Central Management UI, open the menu, go to *Stack Management > {beats} > Central Management*, then define and
To use {beats} Central Management UI, open the menu, go to *Stack Management > Ingest >
{beats} Central Management*, then define and
manage configurations in a central location in {kib} and quickly deploy
configuration changes to all {beats} running across your enterprise. For more
about central management, see the related {beats} documentation:
Expand All @@ -17,8 +18,8 @@ about central management, see the related {beats} documentation:
This feature requires an Elastic license that includes {beats} central
management.
Don't have a license? You can start a 30-day trial. Open the menu, go to
*Stack Management > Elasticsearch > License Management*. At the end of the trial
Don't have a license? You can start a 30-day trial. Open the menu,
go to *Stack Management > Stack > License Management*. At the end of the trial
period, you can purchase a subscription to keep using central management. For
more information, see https://www.elastic.co/subscriptions and
<<managing-licenses>>.
Expand Down
2 changes: 1 addition & 1 deletion docs/management/managing-ccr.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ remote clusters on a local cluster. {ref}/xpack-ccr.html[Cross-cluster replicati
is commonly used to provide remote backups for disaster recovery and for
geo-proximite copies of data.

To get started, open the menu, then go to *Stack Management > Elasticsearch > Cross-Cluster Replication*.
To get started, open the menu, then go to *Stack Management > Data > Cross-Cluster Replication*.

[role="screenshot"]
image::images/cross-cluster-replication-list-view.png[][Cross-cluster replication list view]
Expand Down
20 changes: 10 additions & 10 deletions docs/management/managing-indices.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ the amount of bookkeeping when working with indices. Instead of manually
setting up your indices, you can create them automatically from a template,
ensuring that your settings, mappings, and aliases are consistently defined.

To manage your indices, open the menu, then go to *Stack Management > {es} > Index Management*.
To manage your indices, open the menu, then go to *Stack Management > Data > Index Management*.

[role="screenshot"]
image::images/management_index_labels.png[Index Management UI]
Expand Down Expand Up @@ -130,17 +130,17 @@ Alternatively, you can click the *Load JSON* link and define the mapping as JSON

[source,js]
----------------------------------
{
{
"properties": {
"geo": {
"properties": {
"coordinates": {
"type": "geo_point"
}
}
}
}
}
"properties": {
"coordinates": {
"type": "geo_point"
}
}
}
}
}
----------------------------------

You can create additional mapping configurations in the *Dynamic templates* and
Expand Down
2 changes: 1 addition & 1 deletion docs/management/managing-licenses.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ with no expiration date. For the full list of features, refer to

If you want to try out the full set of features, you can activate a free 30-day
trial. To view the status of your license, start a trial, or install a new
license, open the menu, then go to *Stack Management > {es} > License Management*.
license, open the menu, then go to *Stack Management > Stack > License Management*.

NOTE: You can start a trial only if your cluster has not already activated a
trial license for the current major product version. For example, if you have
Expand Down
2 changes: 1 addition & 1 deletion docs/management/managing-remote-clusters.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ connection from your cluster to other clusters. This functionality is
required for {ref}/xpack-ccr.html[cross-cluster replication] and
{ref}/modules-cross-cluster-search.html[cross-cluster search].

To get started, open the menu, then go to *Stack Management > {es} > Remote Clusters*.
To get started, open the menu, then go to *Stack Management > Data > Remote Clusters*.

[role="screenshot"]
image::images/remote-clusters-list-view.png[Remote Clusters list view, including Add a remote cluster button]
Expand Down
11 changes: 7 additions & 4 deletions docs/management/rollups/create_and_manage_rollups.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ by an index pattern, and then rolls it into a new index. Rollup indices are a go
compactly store months or years of historical
data for use in visualizations and reports.

To get started, open the menu, then go to *Stack Management > {es} > Rollup Jobs*. With this UI,
To get started, open the menu, then go to *Stack Management > Data > Rollup Jobs*. With this UI,
you can:

* <<create-and-manage-rollup-job, Create a rollup job>>
Expand Down Expand Up @@ -130,8 +130,9 @@ Your next step is to visualize your rolled up data in a vertical bar chart.
Most visualizations support rolled up data, with the exception of Timelion and Vega visualizations.


. Create the rollup index pattern in *Management > Index Patterns* so you can
select your rolled up data for visualizations. Click *Create index pattern*, and select *Rollup index pattern* from the dropdown.
. Go to *Stack Management > {kib} > Index Patterns*.

. Click *Create index pattern*, and select *Rollup index pattern* from the dropdown.
+
[role="screenshot"]
image::images/management-rollup-index-pattern.png[][Create rollup index pattern]
Expand All @@ -144,7 +145,9 @@ is `rollup_logstash,kibana_sample_data_logs`. In this index pattern, `rollup_log
matches the rolled up index pattern and `kibana_sample_data_logs` matches the index
pattern for raw data.

. Go to *Visualize* and create a vertical bar chart. Choose `rollup_logstash,kibana_sample_data_logs`
. Go to *Visualize* and create a vertical bar chart.

. Choose `rollup_logstash,kibana_sample_data_logs`
as your source to see both the raw and rolled up data.
+
[role="screenshot"]
Expand Down
4 changes: 2 additions & 2 deletions docs/management/snapshot-restore/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Snapshots are important because they provide a copy of your data in case
something goes wrong. If you need to roll back to an older version of your data,
you can restore a snapshot from the repository.

To get started, open the menu, then go to *Stack Management > {es} > Snapshot and Restore*.
To get started, open the menu, then go to *Stack Management > Data > Snapshot and Restore*.
With this UI, you can:

* Register a repository for storing your snapshots
Expand Down Expand Up @@ -191,7 +191,7 @@ your master and data nodes. You can do this in one of two ways:
Use *Snapshot and Restore* to register the repository where your snapshots
will live.

. Open the menu, then go to *Stack Management > {es} > Snapshot and Restore*.
. Open the menu, then go to *Stack Management > Data > Snapshot and Restore*.
. Click *Register a repository* in either the introductory message or *Repository view*.
. Enter a name for your repository, for example, `my_backup`.
. Select *Shared file system*.
Expand Down
42 changes: 21 additions & 21 deletions docs/management/upgrade-assistant/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -2,50 +2,50 @@
[[upgrade-assistant]]
== Upgrade Assistant

The Upgrade Assistant helps you prepare for your upgrade to the next major {es} version.
For example, if you are using 6.8, the Upgrade Assistant helps you to upgrade to 7.0.
To access the assistant, open the menu, then go to *Stack Management > {es} > Upgrade Assistant*.
The Upgrade Assistant helps you prepare for your upgrade to the next major {es} version.
For example, if you are using 6.8, the Upgrade Assistant helps you to upgrade to 7.0.
To access the assistant, open the menu, then go to *Stack Management > Stack > Upgrade Assistant*.

The assistant identifies the deprecated settings in your cluster and indices
and guides you through the process of resolving issues, including reindexing.
The assistant identifies the deprecated settings in your cluster and indices
and guides you through the process of resolving issues, including reindexing.

Before you upgrade, make sure that you are using the latest released minor
version of {es} to see the most up-to-date deprecation issues.
Before you upgrade, make sure that you are using the latest released minor
version of {es} to see the most up-to-date deprecation issues.
For example, if you want to upgrade to to 7.0, make sure that you are using 6.8.

[float]
=== Reindexing

The *Indices* page lists the indices that are incompatible with the next
The *Indices* page lists the indices that are incompatible with the next
major version of {es}. You can initiate a reindex to resolve the issues.

[role="screenshot"]
image::images/management-upgrade-assistant-9.0.png[]

For a preview of how the data will change during the reindex, select the
index name. A warning appears if the index requires destructive changes.
Back up your index, then proceed with the reindex by accepting each breaking change.
For a preview of how the data will change during the reindex, select the
index name. A warning appears if the index requires destructive changes.
Back up your index, then proceed with the reindex by accepting each breaking change.

You can follow the progress as the Upgrade Assistant makes the index read-only,
creates a new index, reindexes the documents, and creates an alias that points
from the old index to the new one.
You can follow the progress as the Upgrade Assistant makes the index read-only,
creates a new index, reindexes the documents, and creates an alias that points
from the old index to the new one.

If the reindexing fails or is cancelled, the changes are rolled back, the
new index is deleted, and the original index becomes writable. An error
If the reindexing fails or is cancelled, the changes are rolled back, the
new index is deleted, and the original index becomes writable. An error
message explains the reason for the failure.

You can reindex multiple indices at a time, but keep an eye on the
{es} metrics, including CPU usage, memory pressure, and disk usage. If a
metric is so high it affects query performance, cancel the reindex and
You can reindex multiple indices at a time, but keep an eye on the
{es} metrics, including CPU usage, memory pressure, and disk usage. If a
metric is so high it affects query performance, cancel the reindex and
continue by reindexing fewer indices at a time.

Additional considerations:

* If you use {alert-features}, when you reindex the internal indices
(`.watches`), the {watcher} process pauses and no alerts are triggered.
(`.watches`), the {watcher} process pauses and no alerts are triggered.

* If you use {ml-features}, when you reindex the internal indices (`.ml-state`),
the {ml} jobs pause and models are not trained or updated.
the {ml} jobs pause and models are not trained or updated.

* If you use {security-features}, before you reindex the internal indices
(`.security*`), it is a good idea to create a temporary superuser account in the
Expand Down
6 changes: 2 additions & 4 deletions docs/management/watcher-ui/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,8 @@ Watches are helpful for analyzing mission-critical and business-critical
streaming data. For example, you might watch application logs for performance
outages or audit access logs for security threats.

To get started with the Watcher UI, open then menu, then go to *Stack Management > {es} > Watcher*.
To get started with the Watcher UI, open then menu,
then go to *Stack Management > Alerts and Insights > Watcher*.
With this UI, you can:

* <<watcher-create-threshold-alert, Create a simple threshold watch>>
Expand Down Expand Up @@ -238,6 +239,3 @@ Refer to these examples for creating an advanced watch:

* {ref}/watch-cluster-status.html[Watch the status of an {es} cluster]
* {ref}/watching-meetup-data.html[Watch event data]



14 changes: 14 additions & 0 deletions docs/maps/trouble-shooting.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,20 @@ image::maps/images/inspector.png[]
[float]
=== Solutions to common problems

[float]
==== Index not listed when adding layer

* Verify your geospatial data is correctly mapped as {ref}/geo-point.html[geo_point] or {ref}/geo-shape.html[geo_shape].
** Run `GET myIndexPatternTitle/_field_caps?fields=myGeoFieldName` in <<console-kibana, Console>>, replacing `myIndexPatternTitle` and `myGeoFieldName` with your index pattern title and geospatial field name.
** Ensure response specifies `type` as `geo_point` or `geo_shape`.
* Verify your geospatial data is correctly mapped in your <<managing-fields, Kibana index pattern>>.
** Open your index pattern in <<management, Stack Management>>.
** Ensure your geospatial field type is `geo_point` or `geo_shape`.
** Ensure your geospatial field is searchable and aggregatable.
** If your geospatial field type does not match your Elasticsearch mapping, click the *Refresh* button to refresh the field list from Elasticsearch.
* Index patterns with thousands of fields can exceed the default maximum payload size.
Increase <<settings, `server.maxPayloadBytes`>> for large index patterns.

[float]
==== Features are not displayed

Expand Down
Loading

0 comments on commit e983bf8

Please sign in to comment.