Skip to content

Commit

Permalink
Merge branch 'master' into display-date-in-status-col
Browse files Browse the repository at this point in the history
  • Loading branch information
elasticmachine authored May 28, 2020
2 parents b3284c9 + 79a61d0 commit 6dab983
Show file tree
Hide file tree
Showing 79 changed files with 1,325 additions and 404 deletions.
6 changes: 5 additions & 1 deletion .ci/Jenkinsfile_flaky
Original file line number Diff line number Diff line change
Expand Up @@ -70,6 +70,8 @@ def getWorkerFromParams(isXpack, job, ciGroup) {
"run `node scripts/mocha`"
)
})
} else if (job == 'accessibility') {
return kibanaPipeline.functionalTestProcess('kibana-accessibility', './test/scripts/jenkins_accessibility.sh')
} else if (job == 'firefoxSmoke') {
return kibanaPipeline.functionalTestProcess('firefoxSmoke', './test/scripts/jenkins_firefox_smoke.sh')
} else if(job == 'visualRegression') {
Expand All @@ -79,7 +81,9 @@ def getWorkerFromParams(isXpack, job, ciGroup) {
}
}

if (job == 'firefoxSmoke') {
if (job == 'accessibility') {
return kibanaPipeline.functionalTestProcess('xpack-accessibility', './test/scripts/jenkins_xpack_accessibility.sh')
} else if (job == 'firefoxSmoke') {
return kibanaPipeline.functionalTestProcess('xpack-firefoxSmoke', './test/scripts/jenkins_xpack_firefox_smoke.sh')
} else if(job == 'visualRegression') {
return kibanaPipeline.functionalTestProcess('xpack-visualRegression', './test/scripts/jenkins_xpack_visual_regression.sh')
Expand Down
12 changes: 6 additions & 6 deletions .ci/packer_cache.sh
Original file line number Diff line number Diff line change
Expand Up @@ -35,20 +35,20 @@ mkdir -p ".geckodriver"
cp "node_modules/geckodriver/geckodriver.tar.gz" .geckodriver/geckodriver.tar.gz
echo "$geckodriverPkgVersion" > .geckodriver/pkgVersion

echo "Creating bootstrap_cache archive"

# archive cacheable directories
mkdir -p "$HOME/.kibana/bootstrap_cache"
tar -cf "$HOME/.kibana/bootstrap_cache/$branch.tar" \
node_modules \
packages/*/node_modules \
x-pack/node_modules \
x-pack/legacy/plugins/*/node_modules \
x-pack/legacy/plugins/reporting/.chromium \
test/plugin_functional/plugins/*/node_modules \
examples/*/node_modules \
.es \
.chromedriver \
.geckodriver;

echo "Adding node_modules"
# Find all of the node_modules directories that aren't test fixtures, and aren't inside other node_modules directories, and append them to the tar
find . -type d -name node_modules -not -path '*__fixtures__*' -prune -print0 | xargs -0I % tar -rf "$HOME/.kibana/bootstrap_cache/$branch.tar" "%"

echo "created $HOME/.kibana/bootstrap_cache/$branch.tar"

if [ "$branch" == "master" ]; then
Expand Down
1 change: 1 addition & 0 deletions .eslintignore
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@
/.es
/build
/built_assets
/config/apm.dev.js
/data
/html_docs
/optimize
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,6 +31,7 @@ export declare class Field implements IFieldType
| [indexPattern](./kibana-plugin-plugins-data-public.indexpatternfield.indexpattern.md) | | <code>IndexPattern</code> | |
| [lang](./kibana-plugin-plugins-data-public.indexpatternfield.lang.md) | | <code>string</code> | |
| [name](./kibana-plugin-plugins-data-public.indexpatternfield.name.md) | | <code>string</code> | |
| [readFromDocValues](./kibana-plugin-plugins-data-public.indexpatternfield.readfromdocvalues.md) | | <code>boolean</code> | |
| [script](./kibana-plugin-plugins-data-public.indexpatternfield.script.md) | | <code>string</code> | |
| [scripted](./kibana-plugin-plugins-data-public.indexpatternfield.scripted.md) | | <code>boolean</code> | |
| [searchable](./kibana-plugin-plugins-data-public.indexpatternfield.searchable.md) | | <code>boolean</code> | |
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
<!-- Do not edit this file. It is automatically generated by API Documenter. -->

[Home](./index.md) &gt; [kibana-plugin-plugins-data-public](./kibana-plugin-plugins-data-public.md) &gt; [IndexPatternField](./kibana-plugin-plugins-data-public.indexpatternfield.md) &gt; [readFromDocValues](./kibana-plugin-plugins-data-public.indexpatternfield.readfromdocvalues.md)

## IndexPatternField.readFromDocValues property

<b>Signature:</b>

```typescript
readFromDocValues?: boolean;
```
Binary file added docs/maps/images/embed_in_dashboard.jpeg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
36 changes: 30 additions & 6 deletions docs/maps/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -11,17 +11,41 @@ With *Elastic Maps*, you can:

* Create maps with multiple layers and indices.
* Upload GeoJSON files into Elasticsearch.
* Embed your map in Dashboards.
* Plot individual documents or use aggregations to plot any data set, no matter how large.
* Create choropleth maps.
* Use data driven styling to symbolize features from property values.
* Focus the data you want to display with searches.
* Embed your map in dashboards.
* Symbolize features using data values.
* Focus in on just the data you want.

Start your tour of *Elastic Maps* with the <<maps-getting-started, getting started tutorial>>.
*Ready to get started?* Start your tour of *Elastic Maps* with the <<maps-getting-started, getting started tutorial>>.

[float]
=== Create maps with multiple layers and indices
You can use multiple layers and indices to show all your data in a single map. This enables your map to show how data sits relative to physical features like weather patterns, human-made features like international borders, and business-specific features like sales regions. You can plot individual documents or use aggregations to plot any data set, no matter how large.

[role="screenshot"]
image::maps/images/sample_data_ecommerce.png[]

[float]
=== Upload GeoJSON files into Elasticsearch
Elastic Maps makes it easy to import geospatial data into the Elastic Stack. Using the GeoJSON Upload feature, you can drag and drop your point and shape data files directly into Elasticsearch, and then use them as layers in the map.

[float]
=== Embed your map in dashboards
Viewing data from different angles provides better insights. Dimensions that are obscured in one visualization might be illuminated in another. Add your map to a dashboard and view your geospatial data alongside bar charts, pie charts, tag clouds, and more.

This choropleth map shows the density of non-emergency service requests in San Diego by council district. The map is embedded in a dashboard, so users can better understand when services are requested and gain insight into the top requested services.

[role="screenshot"]
image::maps/images/embed_in_dashboard.jpeg[]

[float]
=== Symbolize features using data values
You can customize each layer to highlight meaningful dimensions in your data. For example, you can use dark colors to symbolize areas with more web log traffic, and lighter colors to symbolize areas with less traffic.

[float]
=== Focus in on just the data you want
You can search across your Elasticsearch layers to focus in on just the data you want. Draw a polygon on the map or use the shape from features to create spatial filters to narrow search results to documents that either intersect with, are within, or do not intersect with the specified geometry. Filter individual layers to compares facets.


--

include::maps-getting-started.asciidoc[]
Expand Down
2 changes: 1 addition & 1 deletion docs/user/alerting/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -160,7 +160,7 @@ If you are using an *on-premises* Elastic Stack deployment:

If you are using an *on-premises* Elastic Stack deployment with <<using-kibana-with-security, *security*>>:

* Transport Layer Security (TLS) must be configured for communication <<configuring-tls-kib-es, between {es} and {kib}>>. {kib} alerting uses <<api-keys, API keys>> to secure background alert checks and actions, and API keys require {ref}/configuring-tls.html#tls-http[TLS on the HTTP interface].
* You must enable Transport Layer Security (TLS) for communication <<configuring-tls-kib-es, between {es} and {kib}>>. {kib} alerting uses <<api-keys, API keys>> to secure background alert checks and actions, and API keys require {ref}/configuring-tls.html#tls-http[TLS on the HTTP interface]. A proxy will not suffice.

[float]
[[alerting-security]]
Expand Down
176 changes: 112 additions & 64 deletions docs/user/management.asciidoc
Original file line number Diff line number Diff line change
@@ -1,123 +1,171 @@
[[management]]
= Management
= Stack Management

[partintro]
--
*Management* is home to UIs for managing all things Elastic Stack&mdash;
*Stack Management* is home to UIs for managing all things Elastic Stack&mdash;
indices, clusters, licenses, UI settings, index patterns, spaces, and more.

[float]
[[manage-Elasticsearch]]
== Manage {es}
[[manage-ingest]]
== Ingest

[cols="50, 50"]
|===

a| <<managing-cross-cluster-replication, *Cross-Cluster Replication*>>

Replicate indices on a remote cluster and copy them to a follower index on a local cluster.
This is important for
disaster recovery. It also keeps data local for faster queries.

| <<index-lifecycle-policies, *Index Lifecycle Policies*>>

Create a policy for defining the lifecycle of an index as it ages
through the hot, warm, cold, and delete phases.
Such policies help you control operation costs
because you can put data in different resource tiers.
| <<ingest-node-pipelines, Ingest Node Pipelines>>
| Create and manage {es}
pipelines that enable you to perform common transformations and
enrichments on your data.

a| <<managing-indices, *Index Management*>>
| {logstash-ref}/logstash-centralized-pipeline-management.html[Logstash Pipelines]
| Create, edit, and delete your Logstash pipeline configurations.

View index settings, mappings, and statistics and perform operations, such as refreshing,
flushing, and clearing the cache. Practicing good index management ensures
that your data is stored cost effectively.
| <<managing-beats, Beats Central Management>>
| Manage your Beats configurations in a central location and
quickly deploy configuration changes to all Beats running across your enterprise.

a| <<ingest-node-pipelines, *Ingest Node Pipelines*>>

Create and manage {es}
pipelines that enable you to perform common transformations and
enrichments on your data.
|===

| <<managing-licenses, *License Management*>>
[float]
[[manage-data]]
== Data

View the status of your license, start a trial, or install a new license. For
the full list of features that are included in your license,
see the https://www.elastic.co/subscriptions[subscription page].
[cols="50, 50"]
|===

| <<working-remote-clusters, *Remote Clusters*>>
a| <<managing-indices, Index Management>>
| View index settings, mappings, and statistics and perform operations, such as refreshing,
flushing, and clearing the cache. Practicing good index management ensures
that your data is stored cost effectively.

Manage your remote clusters for use with cross-cluster search and cross-cluster replication.
You can add and remove remote clusters, and check their connectivity.
| <<index-lifecycle-policies, Index Lifecycle Policies>>
|Create a policy for defining the lifecycle of an index as it ages
through the hot, warm, cold, and delete phases.
Such policies help you control operation costs
because you can put data in different resource tiers.

| <<data-rollups, *Rollup Jobs*>>
| <<snapshot-repositories, Snapshot and Restore>>
|Define a policy that creates, schedules, and automatically deletes snapshots to ensure that you
have backups of your cluster in case something goes wrong.

Create a job that periodically aggregates data from one or more indices, and then
| <<data-rollups, Rollup Jobs>>
|Create a job that periodically aggregates data from one or more indices, and then
rolls it into a new, compact index. Rollup indices are a good way to store months or
years of historical data in combination with your raw data.

| <<snapshot-repositories, *Snapshot and Restore*>>
| {ref}/transforms.html[Transforms]
|Use transforms to pivot existing {es} indices into summarized or entity-centric indices.

Define a policy that creates, schedules, and automatically deletes snapshots to ensure that you
have backups of your cluster in case something goes wrong.
| <<managing-cross-cluster-replication, Cross-Cluster Replication>>
|Replicate indices on a remote cluster and copy them to a follower index on a local cluster.
This is important for
disaster recovery. It also keeps data local for faster queries.

| {ref}/transforms.html[*Transforms*]
| <<working-remote-clusters, Remote Clusters>>
|Manage your remote clusters for use with cross-cluster search and cross-cluster replication.
You can add and remove remote clusters, and check their connectivity.
|===

Use transforms to pivot existing {es} indices into summarized or entity-centric indices.
[float]
[[manage-alerts-insights]]
== Alerts and Insights

| <<upgrade-assistant, *Upgrade Assistant*>>
[cols="50, 50"]
|===

Identify the issues that you need to address before upgrading to the
next major version of {es}, and then reindex, if needed.
| <<managing-alerts-and-actions, Alerts&nbsp;and Actions>>
| Centrally manage your alerts across {kib}. Create and manage reusable
connectors for triggering actions.

| <<watcher-ui, *Watcher*>>
| <<reporting-getting-started, Reporting>>
| Monitor the generation of reports&mdash;PDF, PNG, and CSV&mdash;and download reports that you previously generated.
A report can contain a dashboard, visualization, saved search, or Canvas workpad.

| {ml-docs}/ml-jobs.html[Machine Learning Jobs]
| View your {anomaly-jobs} and {dfanalytics-jobs}. Open the Single Metric
Viewer or Anomaly Explorer to see your {ml} results.

Detect changes in your data by creating, managing, and monitoring alerts.
For example, create an alert when the maximum total CPU usage on a machine goes
| <<watcher-ui, Watcher>>
| Detect changes in your data by creating, managing, and monitoring alerts.
For example, you might create an alert when the maximum total CPU usage on a machine goes
above a certain percentage.

|===

[float]
[[manage-kibana]]
== Manage {kib}
[[manage-security]]
== Security

[cols="50, 50"]
|===

a| <<advanced-options, *Advanced Settings*>>

Customize {kib} to suit your needs. Change the format for displaying dates, turn on dark mode,
set the timespan for notification messages, and much more.
a| <<xpack-security, Users>>
|View the users that have been defined on your cluster.
Add or delete users and assign roles that give users
specific privileges.

| <<managing-alerts-and-actions, *Alerts and Actions*>>
| <<xpack-kibana-role-management, Roles>>
|View the roles that exist on your cluster. Customize
the actions that a user with the role can perform, on a cluster, index, and space level.

Centrally manage your alerts across {kib}. Create and manage reusable
connectors for triggering actions.
| <<api-keys, API Keys>>
| Create secondary credentials so that you can send requests on behalf of the user.
Secondary credentials have the same or lower access rights.

| <<managing-fields, *Index Patterns*>>
| <<role-mappings, Role Mappings>>
| Assign roles to your users using a set of rules. Role mappings are required
when authenticating via an external identity provider, such as Active Directory,
Kerberos, PKI, OIDC, and SAML.

Create and manage the index patterns that help you retrieve your data from {es}.
|===

| <<reporting-getting-started, *Reporting*>>
[float]
[[manage-kibana]]
== {kib}

Monitor the generation of reports&mdash;PDF, PNG, and CSV&mdash;and download reports that you previously generated.
A report can contain a dashboard, visualization, saved search, or Canvas workpad.
[cols="50, 50"]
|===

| <<managing-saved-objects, *Saved Objects*>>
a| <<managing-fields, Index Patterns>>
|Create and manage the index patterns that retrieve your data from {es}.

Copy, edit, delete, import, and export your saved objects.
| <<managing-saved-objects, Saved Objects>>
| Copy, edit, delete, import, and export your saved objects.
These include dashboards, visualizations, maps, index patterns, Canvas workpads, and more.

| <<xpack-spaces, *Spaces*>>

Create spaces to organize your dashboards and other saved objects into categories.
| <<xpack-spaces, Spaces>>
| Create spaces to organize your dashboards and other saved objects into categories.
A space is isolated from all other spaces,
so you can tailor it to your needs without impacting others.

| &nbsp;
a| <<advanced-options, Advanced Settings>>
| Customize {kib} to suit your needs. Change the format for displaying dates, turn on dark mode,
set the timespan for notification messages, and much more.

|===

[float]
[[manage-stack]]
== Stack

[cols="50, 50"]
|===

| <<managing-licenses, License Management>>
| View the status of your license, start a trial, or install a new license. For
the full list of features that are included in your license,
see the https://www.elastic.co/subscriptions[subscription page].

| <<upgrade-assistant, Upgrade Assistant>>
| Identify the issues that you need to address before upgrading to the
next major version of {es}, and then reindex, if needed.

|===



--

include::{kib-repo-dir}/management/advanced-options.asciidoc[]
Expand Down
Loading

0 comments on commit 6dab983

Please sign in to comment.