Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[7.17][EMS] Update to ems-client@7.17.2 #187672

Closed
wants to merge 2 commits into from

Merge branch '7.17' into update/ems-client

2f0a91b
Select commit
Loading
Failed to load commit list.
Closed

[7.17][EMS] Update to ems-client@7.17.2 #187672

Merge branch '7.17' into update/ems-client
2f0a91b
Select commit
Loading
Failed to load commit list.
checks-reporter / X-Pack Chrome Functional tests / Group 6 succeeded Jul 9, 2024 in 35m 14s

node scripts/functional_tests --bail --kibana-install-dir /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana-build-xpack --include-tag ciGroup6

Details

[truncated]
taCreateIndexService] [ftr] [.apm-agent-configuration] creating index, cause [api], templates [], shards [1]/[1]
   │ info [o.e.c.r.a.AllocationService] [ftr] updating number_of_replicas to [0] for indices [.apm-agent-configuration]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.apm-custom-link][0], [.apm-agent-configuration][0]]]).
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.kibana_security_session_index_template_1] for index patterns [.kibana_security_session_1]
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_task_manager_7.17.23_001/bhX8oVcNSn2meTyiRoXPew] update_mapping [_doc]
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_7.17.23_001/VbcrXMQsRKywp3Z19G0Y7g] update_mapping [_doc]
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.kibana_security_session_1] creating index, cause [api], templates [.kibana_security_session_index_template_1], shards [1]/[0]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [.alerts-ilm-policy]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.kibana_security_session_1][0]]]).
   │ proc [kibana]   log   [16:21:22.441] [info][plugins][ruleRegistry] Installed common resources shared between all indices
   │ proc [kibana]   log   [16:21:22.443] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.uptime.alerts
   │ proc [kibana]   log   [16:21:22.444] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.logs.alerts
   │ proc [kibana]   log   [16:21:22.444] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.metrics.alerts
   │ proc [kibana]   log   [16:21:22.445] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.apm.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.metrics.alerts-mappings]
   │ proc [kibana]   log   [16:21:22.551] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.metrics.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.uptime.alerts-mappings]
   │ proc [kibana]   log   [16:21:22.600] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.uptime.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.logs.alerts-mappings]
   │ proc [kibana]   log   [16:21:22.662] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.logs.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.apm.alerts-mappings]
   │ proc [kibana]   log   [16:21:22.718] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.apm.alerts
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_7.17.23_001/VbcrXMQsRKywp3Z19G0Y7g] update_mapping [_doc]
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_7.17.23_001/VbcrXMQsRKywp3Z19G0Y7g] update_mapping [_doc]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [kibana-event-log-policy]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.kibana-event-log-7.17.23-snapshot-template] for index patterns [.kibana-event-log-7.17.23-snapshot-*]
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.kibana-event-log-7.17.23-snapshot-000001] creating index, cause [api], templates [.kibana-event-log-7.17.23-snapshot-template], shards [1]/[1]
   │ info [o.e.c.r.a.AllocationService] [ftr] updating number_of_replicas to [0] for indices [.kibana-event-log-7.17.23-snapshot-000001]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.kibana-event-log-7.17.23-snapshot-000001][0]]]).
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.kibana-event-log-7.17.23-snapshot-000001] from [null] to [{"phase":"new","action":"complete","name":"complete"}] in policy [kibana-event-log-policy]
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.kibana-event-log-7.17.23-snapshot-000001] from [{"phase":"new","action":"complete","name":"complete"}] to [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] in policy [kibana-event-log-policy]
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.kibana-event-log-7.17.23-snapshot-000001] from [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] to [{"phase":"hot","action":"rollover","name":"check-rollover-ready"}] in policy [kibana-event-log-policy]
   │ proc [kibana]   log   [16:21:23.819] [info][chromium][plugins][reporting] Browser executable: /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana-build-xpack/x-pack/plugins/reporting/chromium/headless_shell-linux_x64/headless_shell
   │ proc [kibana]   log   [16:21:23.861] [info][plugins][reporting][store] Creating ILM policy for managing reporting indices: kibana-reporting
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [kibana-reporting]
   │ proc [kibana]   log   [16:21:24.042] [info][plugins][securitySolution] Dependent plugin setup complete - Starting ManifestTask
   │ proc [kibana]   log   [16:21:24.747] [info][0][1][endpoint:metadata-check-transforms-task:0][plugins][securitySolution] no endpoint metadata transforms found
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_7.17.23_001/VbcrXMQsRKywp3Z19G0Y7g] update_mapping [_doc]
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.ds-ilm-history-5-2024.07.09-000001] creating index, cause [initialize_data_stream], templates [ilm-history], shards [1]/[0]
   │ info [o.e.c.m.MetadataCreateDataStreamService] [ftr] adding data stream [ilm-history-5] with write index [.ds-ilm-history-5-2024.07.09-000001], backing indices [], and aliases []
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.ds-ilm-history-5-2024.07.09-000001][0]]]).
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.ds-ilm-history-5-2024.07.09-000001] from [null] to [{"phase":"new","action":"complete","name":"complete"}] in policy [ilm-history-ilm-policy]
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.ds-ilm-history-5-2024.07.09-000001] from [{"phase":"new","action":"complete","name":"complete"}] to [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] in policy [ilm-history-ilm-policy]
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.ds-ilm-history-5-2024.07.09-000001] from [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] to [{"phase":"hot","action":"rollover","name":"check-rollover-ready"}] in policy [ilm-history-ilm-policy]
   │ proc [kibana]   log   [16:21:30.498] [info][status] Kibana is now available (was degraded)
   │ info Only running suites which are compatible with ES version 7.17.23
   │ info Only running suites (and their sub-suites) if they include the tag(s): [ 'ciGroup6' ]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] updated role [system_indices_superuser]
   │ info [o.e.x.s.a.u.TransportPutUserAction] [ftr] updated user [system_indices_superuser]
   │ info Only running suites which are compatible with ES version 7.17.23
   │ info Only running suites (and their sub-suites) if they include the tag(s): [ 'ciGroup6' ]
   │ info Starting tests
   │ warn debug logs are being captured, only error logs will be written to the console
   │
     └-: security APIs - Anonymous access
       └-> "before all" hook: beforeTestSuite.trigger in "security APIs - Anonymous access"
       └-: Anonymous authentication
         └-> "before all" hook: beforeTestSuite.trigger for "should reject API requests if client is not authenticated"
         └-> should reject API requests if client is not authenticated
           └-> "before each" hook: global before each for "should reject API requests if client is not authenticated"
           └- ✓ pass  (59ms)
         └-> does not prevent basic login
           └-> "before each" hook: global before each for "does not prevent basic login"
           └- ✓ pass  (561ms)
         └-: login
           └-> "before all" hook: beforeTestSuite.trigger for "should properly set cookie and authenticate user"
           └-> should properly set cookie and authenticate user
             └-> "before each" hook: global before each for "should properly set cookie and authenticate user"
             └- ✓ pass  (1.1s)
           └-> should fail if `Authorization` header is present, but not valid
             └-> "before each" hook: global before each for "should fail if `Authorization` header is present, but not valid"
             └- ✓ pass  (79ms)
           └-> "after all" hook: afterTestSuite.trigger for "should fail if `Authorization` header is present, but not valid"
         └-: API access with active session
           └-> "before all" hook: beforeTestSuite.trigger for "should not extend cookie for system AND non-system API calls"
           └-> should not extend cookie for system AND non-system API calls
             └-> "before each" hook: global before each for "should not extend cookie for system AND non-system API calls"
             └-> "before each" hook for "should not extend cookie for system AND non-system API calls"
             └- ✓ pass  (37ms)
           └-> should fail and preserve session cookie if unsupported authentication schema is used
             └-> "before each" hook: global before each for "should fail and preserve session cookie if unsupported authentication schema is used"
             └-> "before each" hook for "should fail and preserve session cookie if unsupported authentication schema is used"
             └- ✓ pass  (36ms)
           └-> "after all" hook: afterTestSuite.trigger for "should fail and preserve session cookie if unsupported authentication schema is used"
         └-: logging out
           └-> "before all" hook: beforeTestSuite.trigger for "should redirect to `logged_out` page after successful logout"
           └-> should redirect to `logged_out` page after successful logout
             └-> "before each" hook: global before each for "should redirect to `logged_out` page after successful logout"
             └- ✓ pass  (2.0s)
           └-> should redirect to home page if session cookie is not provided
             └-> "before each" hook: global before each for "should redirect to home page if session cookie is not provided"
             └- ✓ pass  (7ms)
           └-> "after all" hook: afterTestSuite.trigger for "should redirect to home page if session cookie is not provided"
         └-> "after all" hook: afterTestSuite.trigger for "does not prevent basic login"
       └-: Anonymous capabilities
         └-> "before all" hook: beforeTestSuite.trigger in "Anonymous capabilities"
         └-> "before all" hook in "Anonymous capabilities"
         └-: without anonymous service account
           └-> "before all" hook: beforeTestSuite.trigger for "all capabilities should be disabled"
           └-> all capabilities should be disabled
             └-> "before each" hook: global before each for "all capabilities should be disabled"
             └- ✓ pass  (781ms)
           └-> "after all" hook: afterTestSuite.trigger for "all capabilities should be disabled"
         └-: with anonymous service account without roles
           └-> "before all" hook: beforeTestSuite.trigger for "all capabilities should be disabled"
           └-> all capabilities should be disabled
             └-> "before each" hook: global before each for "all capabilities should be disabled"
             └- ✓ pass  (272ms)
           └-> "after all" hook: afterTestSuite.trigger for "all capabilities should be disabled"
         └-: with properly configured anonymous service account
           └-> "before all" hook: beforeTestSuite.trigger for "capabilities should be properly defined"
           └-> "before all" hook for "capabilities should be properly defined"
           └-> capabilities should be properly defined
             └-> "before each" hook: global before each for "capabilities should be properly defined"
             └- ✓ pass  (1.5s)
           └-> "after all" hook for "capabilities should be properly defined"
           └-> "after all" hook: afterTestSuite.trigger for "capabilities should be properly defined"
         └-> "after all" hook in "Anonymous capabilities"
         └-> "after all" hook: afterTestSuite.trigger in "Anonymous capabilities"
       └-> "after all" hook: afterTestSuite.trigger in "security APIs - Anonymous access"
   │
   │11 passing (11.3s)
   │
   │ proc [kibana]   log   [16:21:43.225] [info][plugins-system][standard] Stopping all plugins.
   │ proc [kibana]   log   [16:21:43.226] [info][kibana-monitoring][monitoring][monitoring][plugins] Monitoring stats collection is stopped
   │ info [kibana] exited with null after 38.6 seconds
   │ info [es] stopping node ftr
   │ info [o.e.x.m.p.NativeController] [ftr] Native controller process has stopped - no new native processes can be started
   │ info [o.e.n.Node] [ftr] stopping ...
   │ info [o.e.x.w.WatcherService] [ftr] stopping watch service, reason [shutdown initiated]
   │ info [o.e.x.w.WatcherLifeCycleService] [ftr] watcher has stopped and shutdown
   │ info [o.e.n.Node] [ftr] stopped
   │ info [o.e.n.Node] [ftr] closing ...
   │ info [o.e.n.Node] [ftr] closed
   │ info [es] stopped
   │ info [es] no debug files found, assuming es did not write any
   │ info [es] cleanup complete
--- [18/18] Running x-pack/test/observability_functional/with_rac_write.config.ts
 info Installing from snapshot
   │ info version: 7.17.23
   │ info install path: /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr
   │ info license: trial
   │ info Downloading snapshot manifest from https://storage.googleapis.com/kibana-ci-es-snapshots-daily/7.17.23/archives/20240708-131119_42b93a53/manifest.json
   │ info verifying cache of https://storage.googleapis.com/kibana-ci-es-snapshots-daily/7.17.23/archives/20240708-131119_42b93a53/elasticsearch-7.17.23-SNAPSHOT-linux-x86_64.tar.gz
   │ info etags match, reusing cache from 2024-07-09T15:50:27.337Z
   │ info extracting /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/cache/elasticsearch-7.17.23-SNAPSHOT-linux-x86_64.tar.gz
   │ info extracted to /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr
   │ info created /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/ES_TMPDIR
   │ info setting secure setting bootstrap.password to changeme
 info [es] starting node ftr on port 9220
 info Starting
   │ info moved /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/node_modules/@kbn/dev-utils/certs/elasticsearch_nopassword.p12 in config to /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/config/elasticsearch_nopassword.p12
   │ERROR Jul 09, 2024 4:21:54 PM sun.util.locale.provider.LocaleProviderAdapter <clinit>
   │      WARNING: COMPAT locale provider will be removed in a future release
   │      
   │ info [o.e.n.Node] [ftr] version[7.17.23-SNAPSHOT], pid[11963], build[default/tar/42b93a534929add031e668becc4565463f2c4b32/2024-07-08T13:06:16.104506372Z], OS[Linux/5.15.0-1062-gcp/amd64], JVM[Oracle Corporation/OpenJDK 64-Bit Server VM/22.0.1/22.0.1+8-16]
   │ info [o.e.n.Node] [ftr] JVM home [/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/jdk], using bundled JDK [true]
   │ info [o.e.n.Node] [ftr] JVM arguments [-Xshare:auto, -Des.networkaddress.cache.ttl=60, -Des.networkaddress.cache.negative.ttl=10, -XX:+AlwaysPreTouch, -Xss1m, -Djava.awt.headless=true, -Dfile.encoding=UTF-8, -Djna.nosys=true, -XX:-OmitStackTraceInFastThrow, -XX:+ShowCodeDetailsInExceptionMessages, -Dio.netty.noUnsafe=true, -Dio.netty.noKeySetOptimization=true, -Dio.netty.recycler.maxCapacityPerThread=0, -Dio.netty.allocator.numDirectArenas=0, -Dlog4j.shutdownHookEnabled=false, -Dlog4j2.disable.jmx=true, -Dlog4j2.formatMsgNoLookups=true, -Djava.locale.providers=SPI,COMPAT, --add-opens=java.base/java.io=ALL-UNNAMED, -Djava.security.manager=allow, -XX:+UseG1GC, -Djava.io.tmpdir=/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/ES_TMPDIR, -XX:+HeapDumpOnOutOfMemoryError, -XX:+ExitOnOutOfMemoryError, -XX:HeapDumpPath=data, -XX:ErrorFile=logs/hs_err_pid%p.log, -Xlog:gc*,gc+age=trace,safepoint:file=logs/gc.log:utctime,pid,tags:filecount=32,filesize=64m, -XX:+UnlockDiagnosticVMOptions, -XX:G1NumCollectionsKeepPinned=10000000, -Xms1536m, -Xmx1536m, -XX:MaxDirectMemorySize=805306368, -XX:G1HeapRegionSize=4m, -XX:InitiatingHeapOccupancyPercent=30, -XX:G1ReservePercent=15, -Des.path.home=/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr, -Des.path.conf=/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/config, -Des.distribution.flavor=default, -Des.distribution.type=tar, -Des.bundled_jdk=true]
   │ info [o.e.n.Node] [ftr] version [7.17.23-SNAPSHOT] is a pre-release version of Elasticsearch and is not suitable for production
   │ info [o.e.p.PluginsService] [ftr] loaded module [aggs-matrix-stats]
   │ info [o.e.p.PluginsService] [ftr] loaded module [analysis-common]
   │ info [o.e.p.PluginsService] [ftr] loaded module [constant-keyword]
   │ info [o.e.p.PluginsService] [ftr] loaded module [frozen-indices]
   │ info [o.e.p.PluginsService] [ftr] loaded module [ingest-common]
   │ info [o.e.p.PluginsService] [ftr] loaded module [ingest-geoip]
   │ info [o.e.p.PluginsService] [ftr] loaded module [ingest-user-agent]
   │ info [o.e.p.PluginsService] [ftr] loaded module [kibana]
   │ info [o.e.p.PluginsService] [ftr] loaded module [lang-expression]
   │ info [o.e.p.PluginsService] [ftr] loaded module [lang-mustache]
   │ info [o.e.p.PluginsService] [ftr] loaded module [lang-painless]
   │ info [o.e.p.PluginsService] [ftr] loaded module [legacy-geo]
   │ info [o.e.p.PluginsService] [ftr] loaded module [mapper-extras]
   │ info [o.e.p.PluginsService] [ftr] loaded module [mapper-version]
   │ info [o.e.p.PluginsService] [ftr] loaded module [parent-join]
   │ info [o.e.p.PluginsService] [ftr] loaded module [percolator]
   │ info [o.e.p.PluginsService] [ftr] loaded module [rank-eval]
   │ info [o.e.p.PluginsService] [ftr] loaded module [reindex]
   │ info [o.e.p.PluginsService] [ftr] loaded module [repositories-metering-api]
   │ info [o.e.p.PluginsService] [ftr] loaded module [repository-encrypted]
   │ info [o.e.p.PluginsService] [ftr] loaded module [repository-url]
   │ info [o.e.p.PluginsService] [ftr] loaded module [runtime-fields-common]
   │ info [o.e.p.PluginsService] [ftr] loaded module [search-business-rules]
   │ info [o.e.p.PluginsService] [ftr] loaded module [searchable-snapshots]
   │ info [o.e.p.PluginsService] [ftr] loaded module [snapshot-repo-test-kit]
   │ info [o.e.p.PluginsService] [ftr] loaded module [spatial]
   │ info [o.e.p.PluginsService] [ftr] loaded module [test-delayed-aggs]
   │ info [o.e.p.PluginsService] [ftr] loaded module [test-die-with-dignity]
   │ info [o.e.p.PluginsService] [ftr] loaded module [test-error-query]
   │ info [o.e.p.PluginsService] [ftr] loaded module [transform]
   │ info [o.e.p.PluginsService] [ftr] loaded module [transport-netty4]
   │ info [o.e.p.PluginsService] [ftr] loaded module [unsigned-long]
   │ info [o.e.p.PluginsService] [ftr] loaded module [vector-tile]
   │ info [o.e.p.PluginsService] [ftr] loaded module [vectors]
   │ info [o.e.p.PluginsService] [ftr] loaded module [wildcard]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-aggregate-metric]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-analytics]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-async]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-async-search]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-autoscaling]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-ccr]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-core]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-data-streams]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-deprecation]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-enrich]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-eql]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-fleet]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-graph]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-identity-provider]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-ilm]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-logstash]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-ml]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-monitoring]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-ql]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-rollup]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-security]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-shutdown]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-sql]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-stack]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-text-structure]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-voting-only-node]
   │ info [o.e.p.PluginsService] [ftr] loaded module [x-pack-watcher]
   │ info [o.e.p.PluginsService] [ftr] no plugins loaded
   │ info [o.e.e.NodeEnvironment] [ftr] using [1] data paths, mounts [[/opt/local-ssd (/dev/nvme0n1)]], net usable_space [343.5gb], net total_space [368gb], types [ext4]
   │ info [o.e.e.NodeEnvironment] [ftr] heap size [1.5gb], compressed ordinary object pointers [true]
   │ info [o.e.n.Node] [ftr] node name [ftr], node ID [8Ezxt7quRWWSmD9eHHjR9Q], cluster name [job-kibana-default-ciGroup6-cluster-ftr], roles [transform, data_frozen, master, remote_cluster_client, data, ml, data_content, data_hot, data_warm, data_cold, ingest]
   │ info [o.e.x.m.p.l.CppLogMessageHandler] [ftr] [controller/12132] [Main.cc@122] controller (64 bit): Version 7.17.23-SNAPSHOT (Build 3e4489a02bea5d) Copyright (c) 2024 Elasticsearch BV
   │ info [o.e.x.s.a.Realms] [ftr] license mode is [trial], currently licensed security realms are [reserved/reserved,file/default_file,native/default_native]
   │ info [o.e.x.s.a.s.FileRolesStore] [ftr] parsed [0] roles from file [/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/config/roles.yml]
   │ info [o.e.i.g.ConfigDatabases] [ftr] initialized default databases [[GeoLite2-Country.mmdb, GeoLite2-City.mmdb, GeoLite2-ASN.mmdb]], config databases [[]] and watching [/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/config/ingest-geoip] for changes
   │ info [o.e.i.g.DatabaseNodeService] [ftr] initialized database registry, using geoip-databases directory [/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/.es/job-kibana-default-ciGroup6-cluster-ftr/ES_TMPDIR/geoip-databases/8Ezxt7quRWWSmD9eHHjR9Q]
   │ info [o.e.t.NettyAllocator] [ftr] creating NettyAllocator with the following configs: [name=elasticsearch_configured, chunk_size=1mb, suggested_max_allocation_size=1mb, factors={es.unsafe.use_netty_default_chunk_and_page_size=false, g1gc_enabled=true, g1gc_region_size=4mb}]
   │ info [o.e.i.r.RecoverySettings] [ftr] using rate limit [40mb] with [default=40mb, read=0b, write=0b, max=0b]
   │ info [o.e.d.DiscoveryModule] [ftr] using discovery type [single-node] and seed hosts providers [settings]
   │ info [o.e.g.DanglingIndicesState] [ftr] gateway.auto_import_dangling_indices is disabled, dangling indices will not be automatically detected or imported and must be managed manually
   │ info [o.e.n.Node] [ftr] initialized
   │ info [o.e.n.Node] [ftr] starting ...
   │ info [o.e.x.s.c.f.PersistentCache] [ftr] persistent cache index loaded
   │ info [o.e.x.d.l.DeprecationIndexingComponent] [ftr] deprecation component started
   │ info [o.e.t.TransportService] [ftr] publish_address {127.0.0.1:9300}, bound_addresses {[::1]:9300}, {127.0.0.1:9300}
   │ info [o.e.x.m.Monitoring] [ftr] creating template [.monitoring-alerts-7] with version [7]
   │ info [o.e.x.m.Monitoring] [ftr] creating template [.monitoring-es] with version [7]
   │ info [o.e.x.m.Monitoring] [ftr] creating template [.monitoring-kibana] with version [7]
   │ info [o.e.x.m.Monitoring] [ftr] creating template [.monitoring-logstash] with version [7]
   │ info [o.e.x.m.Monitoring] [ftr] creating template [.monitoring-beats] with version [7]
   │ info [o.e.c.c.Coordinator] [ftr] setting initial configuration to VotingConfiguration{8Ezxt7quRWWSmD9eHHjR9Q}
   │ info [o.e.c.s.MasterService] [ftr] elected-as-master ([1] nodes joined)[{ftr}{8Ezxt7quRWWSmD9eHHjR9Q}{uT2nragHRRWGXpDgJc_GCg}{127.0.0.1}{127.0.0.1:9300}{cdfhilmrstw} elect leader, _BECOME_MASTER_TASK_, _FINISH_ELECTION_], term: 1, version: 1, delta: master node changed {previous [], current [{ftr}{8Ezxt7quRWWSmD9eHHjR9Q}{uT2nragHRRWGXpDgJc_GCg}{127.0.0.1}{127.0.0.1:9300}{cdfhilmrstw}]}
   │ info [o.e.c.c.CoordinationState] [ftr] cluster UUID set to [xORQTK1bQs2xE5QpkLP0VQ]
   │ info [o.e.c.s.ClusterApplierService] [ftr] master node changed {previous [], current [{ftr}{8Ezxt7quRWWSmD9eHHjR9Q}{uT2nragHRRWGXpDgJc_GCg}{127.0.0.1}{127.0.0.1:9300}{cdfhilmrstw}]}, term: 1, version: 1, reason: Publication{term=1, version=1}
   │ info [o.e.h.AbstractHttpServerTransport] [ftr] publish_address {127.0.0.1:9220}, bound_addresses {[::1]:9220}, {127.0.0.1:9220}
   │ info [o.e.n.Node] [ftr] started
   │ info [o.e.g.GatewayService] [ftr] recovered [0] indices into cluster_state
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.ml-notifications-000002] for index patterns [.ml-notifications-000002]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.ml-state] for index patterns [.ml-state*]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.ml-anomalies-] for index patterns [.ml-anomalies-*]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.ml-stats] for index patterns [.ml-stats-*]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [data-streams-mappings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [metrics-mappings]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [logs-settings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [logs-mappings]
   │ info [o.e.x.s.s.SecurityIndexManager] [ftr] security index does not exist, creating [.security-7] with alias [.security]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [synthetics-settings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [metrics-settings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [synthetics-mappings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.watch-history-13] for index patterns [.watcher-history-13*]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [ilm-history] for index patterns [ilm-history-5*]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.slm-history] for index patterns [.slm-history-5*]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.deprecation-indexing-settings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.deprecation-indexing-mappings]
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.security-7] creating index, cause [api], templates [], shards [1]/[0]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [logs] for index patterns [logs-*-*]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [metrics] for index patterns [metrics-*-*]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [synthetics] for index patterns [synthetics-*-*]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.deprecation-indexing-template] for index patterns [.logs-deprecation.*]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.security-7][0]]]).
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [ml-size-based-ilm-policy]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [logs]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [synthetics]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [7-days-default]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [metrics]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [90-days-default]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [180-days-default]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [30-days-default]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [365-days-default]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [watch-history-ilm-policy]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [ilm-history-ilm-policy]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [slm-history-ilm-policy]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [.deprecation-indexing-ilm-policy]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [.fleet-actions-results-ilm-policy]
   │ info [o.e.l.LicenseService] [ftr] license [2ea253e4-1204-4242-9c3d-30923dd1a360] mode [trial] - valid
   │ info [o.e.x.s.a.Realms] [ftr] license mode is [trial], currently licensed security realms are [reserved/reserved,file/default_file,native/default_native]
   │ info [o.e.x.s.s.SecurityStatusChangeListener] [ftr] Active license is now [TRIAL]; Security is enabled
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [system_indices_superuser]
   │ info [o.e.x.s.a.u.TransportPutUserAction] [ftr] added user [system_indices_superuser]
   │ info starting [kibana] > /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana-build-xpack/bin/kibana --logging.json=false --server.port=5620 --elasticsearch.username=kibana_system --elasticsearch.password=changeme --data.search.aggs.shardDelay.enabled=true --security.showInsecureClusterWarning=false --telemetry.banner=false --telemetry.optIn=false --telemetry.sendUsageTo=staging --server.maxPayload=1679958 --plugin-path=/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/test/common/fixtures/plugins/newsfeed --newsfeed.service.urlRoot=http://localhost:5620 --newsfeed.service.pathTemplate=/api/_newsfeed-FTS-external-service-simulators/kibana/v{VERSION}.json --logging.appenders.deprecation.type=console --logging.appenders.deprecation.layout.type=json --logging.loggers[0].name=elasticsearch.deprecation --logging.loggers[0].level=all --logging.loggers[0].appenders[0]=deprecation --status.allowAnonymous=true --server.uuid=5b2de169-2785-441b-ae8c-186a1936b17d --xpack.maps.showMapsInspectorAdapter=true --xpack.maps.preserveDrawingBuffer=true --xpack.security.encryptionKey="wuGNaIhoMpk5sO4UBxgr3NyW1sFcLgIf" --xpack.encryptedSavedObjects.encryptionKey="DkdXazszSCYexXqz4YktBGHCRkV6hyNK" --xpack.discoverEnhanced.actions.exploreDataInContextMenu.enabled=true --savedObjects.maxImportPayloadBytes=10485760 --xpack.siem.enabled=true --elasticsearch.hosts=https://localhost:9220 --elasticsearch.ssl.certificateAuthorities=/opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana/node_modules/@kbn/dev-utils/certs/ca.crt
   │ proc [kibana] Kibana is currently running with legacy OpenSSL providers enabled! For details and instructions on how to disable see https://www.elastic.co/guide/en/kibana/7.17/production.html#openssl-legacy-provider
   │ proc [kibana]   log   [16:22:26.115] [info][plugins-service] Plugin "metricsEntities" is disabled.
   │ proc [kibana]   log   [16:22:26.194] [info][server][Preboot][http] http server running at http://localhost:5620
   │ proc [kibana]   log   [16:22:26.241] [warning][config][deprecation] Starting in 8.0, the Kibana logging format will be changing. This may affect you if you are doing any special handling of your Kibana logs, such as ingesting logs into Elasticsearch for further analysis. If you are using the new logging configuration, you are already receiving logs in both old and new formats, and the old format will simply be going away. If you are not yet using the new logging configuration, the log format will change upon upgrade to 8.0. Beginning in 8.0, the format of JSON logs will be ECS-compatible JSON, and the default pattern log format will be configurable with our new logging system. Please refer to the documentation for more information about the new logging format.
   │ proc [kibana]   log   [16:22:26.242] [warning][config][deprecation] The default mechanism for Reporting privileges will work differently in future versions, which will affect the behavior of this cluster. Set "xpack.reporting.roles.enabled" to "false" to adopt the future behavior before upgrading.
   │ proc [kibana]   log   [16:22:26.243] [warning][config][deprecation] Setting "security.showInsecureClusterWarning" has been replaced by "xpack.security.showInsecureClusterWarning"
   │ proc [kibana]   log   [16:22:26.243] [warning][config][deprecation] User sessions will automatically time out after 8 hours of inactivity starting in 8.0. Override this value to change the timeout.
   │ proc [kibana]   log   [16:22:26.244] [warning][config][deprecation] Users are automatically required to log in again after 30 days starting in 8.0. Override this value to change the timeout.
   │ proc [kibana]   log   [16:22:26.245] [warning][config][deprecation] Setting "xpack.siem.enabled" has been replaced by "xpack.securitySolution.enabled"
   │ proc [kibana]   log   [16:22:26.375] [info][plugins-system][standard] Setting up [114] plugins: [newsfeedFixtures,translations,licensing,globalSearch,globalSearchProviders,features,licenseApiGuard,code,usageCollection,xpackLegacy,taskManager,telemetryCollectionManager,telemetryCollectionXpack,kibanaUsageCollection,share,embeddable,uiActionsEnhanced,screenshotMode,banners,telemetry,newsfeed,mapsEms,mapsLegacy,kibanaLegacy,fieldFormats,expressions,dataViews,charts,esUiShared,bfetch,data,savedObjects,presentationUtil,expressionShape,expressionRevealImage,expressionRepeatImage,expressionMetric,expressionImage,customIntegrations,home,searchprofiler,painlessLab,grokdebugger,management,watcher,licenseManagement,advancedSettings,spaces,security,savedObjectsTagging,reporting,canvas,lists,ingestPipelines,fileUpload,encryptedSavedObjects,dataEnhanced,cloud,snapshotRestore,eventLog,actions,alerting,triggersActionsUi,transform,stackAlerts,ruleRegistry,visualizations,visTypeXy,visTypeVislib,visTypeVega,visTypeTimelion,visTypeTagcloud,visTypeTable,visTypePie,visTypeMetric,visTypeMarkdown,tileMap,regionMap,expressionTagcloud,expressionMetricVis,console,graph,fleet,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,dashboard,maps,dashboardMode,dashboardEnhanced,visualize,visTypeTimeseries,rollup,indexPatternFieldEditor,lens,cases,timelines,discover,osquery,observability,discoverEnhanced,dataVisualizer,ml,uptime,securitySolution,infra,upgradeAssistant,monitoring,logstash,enterpriseSearch,apm,savedObjectsManagement,indexPatternManagement]
   │ proc [kibana]   log   [16:22:26.393] [info][plugins][taskManager] TaskManager is identified by the Kibana UUID: 5b2de169-2785-441b-ae8c-186a1936b17d
   │ proc [kibana]   log   [16:22:26.514] [warning][config][plugins][security] Session cookies will be transmitted over insecure connections. This is not recommended.
   │ proc [kibana]   log   [16:22:26.533] [warning][config][plugins][security] Session cookies will be transmitted over insecure connections. This is not recommended.
   │ proc [kibana]   log   [16:22:26.549] [warning][config][plugins][reporting] Generating a random key for xpack.reporting.encryptionKey. To prevent sessions from being invalidated on restart, please set xpack.reporting.encryptionKey in the kibana.yml or use the bin/kibana-encryption-keys command.
   │ proc [kibana]   log   [16:22:26.571] [info][encryptedSavedObjects][plugins] Hashed 'xpack.encryptedSavedObjects.encryptionKey' for this instance: nnkvE7kjGgidcjXzmLYBbIh4THhRWI1/7fUjAEaJWug=
   │ proc [kibana]   log   [16:22:26.609] [info][plugins][ruleRegistry] Installing common resources shared between all indices
   │ proc [kibana]   log   [16:22:27.066] [info][config][plugins][reporting] Chromium sandbox provides an additional layer of protection, and is supported for Linux Ubuntu 20.04 OS. Automatically enabling Chromium sandbox.
   │ proc [kibana]   log   [16:22:27.387] [info][savedobjects-service] Waiting until all Elasticsearch nodes are compatible with Kibana before starting saved objects migrations...
   │ proc [kibana]   log   [16:22:27.388] [info][savedobjects-service] Starting saved objects migrations
   │ proc [kibana]   log   [16:22:27.451] [info][savedobjects-service] [.kibana] INIT -> CREATE_NEW_TARGET. took: 16ms.
   │ proc [kibana]   log   [16:22:27.457] [info][savedobjects-service] [.kibana_task_manager] INIT -> CREATE_NEW_TARGET. took: 18ms.
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.kibana_task_manager_7.17.23_001] creating index, cause [api], templates [], shards [1]/[1]
   │ info [o.e.c.r.a.AllocationService] [ftr] updating number_of_replicas to [0] for indices [.kibana_task_manager_7.17.23_001]
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.kibana_7.17.23_001] creating index, cause [api], templates [], shards [1]/[1]
   │ info [o.e.c.r.a.AllocationService] [ftr] updating number_of_replicas to [0] for indices [.kibana_7.17.23_001]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.kibana_task_manager_7.17.23_001][0], [.kibana_7.17.23_001][0]]]).
   │ proc [kibana]   log   [16:22:27.711] [info][savedobjects-service] [.kibana] CREATE_NEW_TARGET -> MARK_VERSION_INDEX_READY. took: 260ms.
   │ proc [kibana]   log   [16:22:27.713] [info][savedobjects-service] [.kibana_task_manager] CREATE_NEW_TARGET -> MARK_VERSION_INDEX_READY. took: 256ms.
   │ proc [kibana]   log   [16:22:27.810] [info][savedobjects-service] [.kibana] MARK_VERSION_INDEX_READY -> DONE. took: 99ms.
   │ proc [kibana]   log   [16:22:27.810] [info][savedobjects-service] [.kibana] Migration completed after 375ms
   │ proc [kibana]   log   [16:22:27.839] [info][savedobjects-service] [.kibana_task_manager] MARK_VERSION_INDEX_READY -> DONE. took: 126ms.
   │ proc [kibana]   log   [16:22:27.839] [info][savedobjects-service] [.kibana_task_manager] Migration completed after 400ms
   │ proc [kibana]   log   [16:22:27.851] [info][plugins-system][standard] Starting [114] plugins: [newsfeedFixtures,translations,licensing,globalSearch,globalSearchProviders,features,licenseApiGuard,code,usageCollection,xpackLegacy,taskManager,telemetryCollectionManager,telemetryCollectionXpack,kibanaUsageCollection,share,embeddable,uiActionsEnhanced,screenshotMode,banners,telemetry,newsfeed,mapsEms,mapsLegacy,kibanaLegacy,fieldFormats,expressions,dataViews,charts,esUiShared,bfetch,data,savedObjects,presentationUtil,expressionShape,expressionRevealImage,expressionRepeatImage,expressionMetric,expressionImage,customIntegrations,home,searchprofiler,painlessLab,grokdebugger,management,watcher,licenseManagement,advancedSettings,spaces,security,savedObjectsTagging,reporting,canvas,lists,ingestPipelines,fileUpload,encryptedSavedObjects,dataEnhanced,cloud,snapshotRestore,eventLog,actions,alerting,triggersActionsUi,transform,stackAlerts,ruleRegistry,visualizations,visTypeXy,visTypeVislib,visTypeVega,visTypeTimelion,visTypeTagcloud,visTypeTable,visTypePie,visTypeMetric,visTypeMarkdown,tileMap,regionMap,expressionTagcloud,expressionMetricVis,console,graph,fleet,indexManagement,remoteClusters,crossClusterReplication,indexLifecycleManagement,dashboard,maps,dashboardMode,dashboardEnhanced,visualize,visTypeTimeseries,rollup,indexPatternFieldEditor,lens,cases,timelines,discover,osquery,observability,discoverEnhanced,dataVisualizer,ml,uptime,securitySolution,infra,upgradeAssistant,monitoring,logstash,enterpriseSearch,apm,savedObjectsManagement,indexPatternManagement]
   │ proc [kibana]   log   [16:22:29.017] [info][monitoring][monitoring][plugins] config sourced from: production cluster
   │ proc [kibana]   log   [16:22:30.316] [info][server][Kibana][http] http server running at http://localhost:5620
   │ proc [kibana]   log   [16:22:30.383] [info][status] Kibana is now degraded
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_task_manager_7.17.23_001/G_IyX1aVT6eiYF0TckAGJQ] update_mapping [_doc]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-technical-mappings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-ecs-mappings]
   │ proc [kibana]   log   [16:22:30.658] [info][kibana-monitoring][monitoring][monitoring][plugins] Starting monitoring stats collection
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.apm-custom-link] creating index, cause [api], templates [], shards [1]/[1]
   │ info [o.e.c.r.a.AllocationService] [ftr] updating number_of_replicas to [0] for indices [.apm-custom-link]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.apm-custom-link][0]]]).
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.kibana_security_session_index_template_1] for index patterns [.kibana_security_session_1]
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.apm-agent-configuration] creating index, cause [api], templates [], shards [1]/[1]
   │ info [o.e.c.r.a.AllocationService] [ftr] updating number_of_replicas to [0] for indices [.apm-agent-configuration]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.apm-agent-configuration][0]]]).
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.kibana_security_session_1] creating index, cause [api], templates [.kibana_security_session_index_template_1], shards [1]/[0]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.kibana_security_session_1][0]]]).
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_7.17.23_001/fDDZik0JSKyteoq1_smeZw] update_mapping [_doc]
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_7.17.23_001/fDDZik0JSKyteoq1_smeZw] update_mapping [_doc]
   │ info [o.e.c.m.MetadataMappingService] [ftr] [.kibana_7.17.23_001/fDDZik0JSKyteoq1_smeZw] update_mapping [_doc]
   │ proc [kibana]   log   [16:22:31.712] [info][plugins][securitySolution] Dependent plugin setup complete - Starting ManifestTask
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [.alerts-ilm-policy]
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [kibana-event-log-policy]
   │ proc [kibana]   log   [16:22:31.814] [info][plugins][ruleRegistry] Installed common resources shared between all indices
   │ proc [kibana]   log   [16:22:31.814] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.uptime.alerts
   │ proc [kibana]   log   [16:22:31.816] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.logs.alerts
   │ proc [kibana]   log   [16:22:31.817] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.metrics.alerts
   │ proc [kibana]   log   [16:22:31.829] [info][plugins][ruleRegistry] Installing resources for index .alerts-observability.apm.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.metrics.alerts-mappings]
   │ proc [kibana]   log   [16:22:31.889] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.metrics.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.uptime.alerts-mappings]
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.logs.alerts-mappings]
   │ proc [kibana]   log   [16:22:31.961] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.uptime.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding component template [.alerts-observability.apm.alerts-mappings]
   │ proc [kibana]   log   [16:22:32.063] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.logs.alerts
   │ proc [kibana]   log   [16:22:32.122] [info][plugins][ruleRegistry] Installed resources for index .alerts-observability.apm.alerts
   │ info [o.e.c.m.MetadataIndexTemplateService] [ftr] adding index template [.kibana-event-log-7.17.23-snapshot-template] for index patterns [.kibana-event-log-7.17.23-snapshot-*]
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.kibana-event-log-7.17.23-snapshot-000001] creating index, cause [api], templates [.kibana-event-log-7.17.23-snapshot-template], shards [1]/[1]
   │ info [o.e.c.r.a.AllocationService] [ftr] updating number_of_replicas to [0] for indices [.kibana-event-log-7.17.23-snapshot-000001]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.kibana-event-log-7.17.23-snapshot-000001][0]]]).
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.kibana-event-log-7.17.23-snapshot-000001] from [null] to [{"phase":"new","action":"complete","name":"complete"}] in policy [kibana-event-log-policy]
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.kibana-event-log-7.17.23-snapshot-000001] from [{"phase":"new","action":"complete","name":"complete"}] to [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] in policy [kibana-event-log-policy]
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.kibana-event-log-7.17.23-snapshot-000001] from [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] to [{"phase":"hot","action":"rollover","name":"check-rollover-ready"}] in policy [kibana-event-log-policy]
   │ proc [kibana]   log   [16:22:32.956] [info][chromium][plugins][reporting] Browser executable: /opt/local-ssd/buildkite/builds/bk-agent-prod-gcp-1720539735500162592/elastic/kibana-pull-request/kibana-build-xpack/x-pack/plugins/reporting/chromium/headless_shell-linux_x64/headless_shell
   │ proc [kibana]   log   [16:22:32.999] [info][plugins][reporting][store] Creating ILM policy for managing reporting indices: kibana-reporting
   │ info [o.e.x.i.a.TransportPutLifecycleAction] [ftr] adding index lifecycle policy [kibana-reporting]
   │ proc [kibana]   log   [16:22:33.648] [info][0][1][endpoint:metadata-check-transforms-task:0][plugins][securitySolution] no endpoint metadata transforms found
   │ info [o.e.c.m.MetadataCreateIndexService] [ftr] [.ds-ilm-history-5-2024.07.09-000001] creating index, cause [initialize_data_stream], templates [ilm-history], shards [1]/[0]
   │ info [o.e.c.m.MetadataCreateDataStreamService] [ftr] adding data stream [ilm-history-5] with write index [.ds-ilm-history-5-2024.07.09-000001], backing indices [], and aliases []
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.ds-ilm-history-5-2024.07.09-000001] from [null] to [{"phase":"new","action":"complete","name":"complete"}] in policy [ilm-history-ilm-policy]
   │ info [o.e.c.r.a.AllocationService] [ftr] Cluster health status changed from [YELLOW] to [GREEN] (reason: [shards started [[.ds-ilm-history-5-2024.07.09-000001][0]]]).
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.ds-ilm-history-5-2024.07.09-000001] from [{"phase":"new","action":"complete","name":"complete"}] to [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] in policy [ilm-history-ilm-policy]
   │ info [o.e.x.i.IndexLifecycleTransition] [ftr] moving index [.ds-ilm-history-5-2024.07.09-000001] from [{"phase":"hot","action":"unfollow","name":"branch-check-unfollow-prerequisites"}] to [{"phase":"hot","action":"rollover","name":"check-rollover-ready"}] in policy [ilm-history-ilm-policy]
   │ proc [kibana]   log   [16:22:37.835] [info][status] Kibana is now available (was degraded)
   │ info Only running suites which are compatible with ES version 7.17.23
   │ info Only running suites (and their sub-suites) if they include the tag(s): [ 'ciGroup6' ]
   │ info Remote initialized: chrome-headless-shell 126.0.6478.126
   │ info chromedriver version: 126.0.6478.126 (d36ace6122e0a59570e258d82441395206d60e1c-refs/branch-heads/6478@{#1591})
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] updated role [system_indices_superuser]
   │ info [o.e.x.s.a.u.TransportPutUserAction] [ftr] updated user [system_indices_superuser]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [test_logstash_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_canvas_all]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_discover_all]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_dashboard_read]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_discover_read]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_visualize_read]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_visualize_all]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_dashboard_all]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_maps_all]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_maps_read]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [geoshape_data_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [antimeridian_points_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [antimeridian_shapes_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [meta_for_geoshape_data_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [geoconnections_data_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [test_logs_data_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [geoall_data_writer]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_index_pattern_management_all]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_devtools_read]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_ccr_role]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_upgrade_assistant_role]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [manage_rollups_role]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [test_rollup_reader]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [test_api_keys]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [manage_security]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [ccr_user]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [manage_ilm]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [index_management_user]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [snapshot_restore_user]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [ingest_pipelines_user]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [license_management_user]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [logstash_read_user]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [remote_clusters_user]
   │ info [o.e.x.s.a.r.TransportPutRoleAction] [ftr] added role [global_alerts_logs_all_else_read]
   │ info [o.e.x.s.a.u.TransportPutUserAction] [ftr] added user [test_user]
   │ info Only running suites which are compatible with ES version 7.17.23
   │ info Only running suites (and their sub-suites) if they include the tag(s): [ 'ciGroup6' ]
   │ info Starting tests
   │ warn debug logs are being captured, only error logs will be written to the console
   │
     └-: ObservabilityApp
       └-> "before all" hook: beforeTestSuite.trigger in "ObservabilityApp"
       └-: feature controls
         └-> "before all" hook: beforeTestSuite.trigger in "feature controls"
         └-: observability security feature controls
           └-> "before all" hook: beforeTestSuite.trigger for "Shows the no data page on load"
           └-> "before all" hook for "Shows the no data page on load"
           └-> Shows the no data page on load
             └-> "before each" hook: global before each for "Shows the no data page on load"
             └- ✓ pass  (7.9s)
           └-: observability cases all privileges
             └-> "before all" hook: beforeTestSuite.trigger for "shows observability/cases navlink"
             └-> "before all" hook for "shows observability/cases navlink"
             └-> shows observability/cases navlink
               └-> "before each" hook: global before each for "shows observability/cases navlink"
               └- ✓ pass  (6.5s)
             └-> landing page shows "Create new case" button
               └-> "before each" hook: global before each for "landing page shows "Create new case" button"
               └- ✓ pass  (1.4s)
             └-> doesn't show read-only badge
               └-> "before each" hook: global before each for "doesn't show read-only badge"
               └- ✓ pass  (1.7s)
             └-> allows a case to be created
               └-> "before each" hook: global before each for "allows a case to be created"
               └- ✓ pass  (1.5s)
             └-> allows a case to be edited
               └-> "before each" hook: global before each for "allows a case to be edited"
               └- ✓ pass  (1.5s)
             └-> "after all" hook for "allows a case to be edited"
             └-> "after all" hook: afterTestSuite.trigger for "allows a case to be edited"
           └-: observability cases read-only privileges
             └-> "before all" hook: beforeTestSuite.trigger for "shows observability/cases navlink"
             └-> "before all" hook for "shows observability/cases navlink"
             └-> shows observability/cases navlink
               └-> "before each" hook: global before each for "shows observability/cases navlink"
               └- ✓ pass  (6.3s)
             └-> landing page shows disabled "Create new case" button
               └-> "before each" hook: global before each for "landing page shows disabled "Create new case" button"
               └- ✓ pass  (1.7s)
             └-> shows read-only glasses badge
               └-> "before each" hook: global before each for "shows read-only glasses badge"
               └- ✓ pass  (1.0s)
             └-> does not allow a case to be created
               └-> "before each" hook: global before each for "does not allow a case to be created"
               └- ✓ pass  (1.7s)
             └-> does not allow a case to be edited
               └-> "before each" hook: global before each for "does not allow a case to be edited"
               └- ✓ pass  (1.6s)
             └-> "after all" hook for "does not allow a case to be edited"
             └-> "after all" hook: afterTestSuite.trigger for "does not allow a case to be edited"
           └-: no observability privileges
             └-> "before all" hook: beforeTestSuite.trigger for "returns a 403"
             └-> "before all" hook for "returns a 403"
             └-> returns a 403
               └-> "before each" hook: global before each for "returns a 403"
               └- ✓ pass  (1.3s)
             └-> create new case returns a 403
             └-> "after all" hook for "create new case returns a 403"
             └-> "after all" hook: afterTestSuite.trigger for "create new case returns a 403"
           └-> "after all" hook for "Shows the no data page on load"
           └-> "after all" hook: afterTestSuite.trigger for "Shows the no data page on load"
         └-> "after all" hook: afterTestSuite.trigger in "feature controls"
       └-: ExploratoryView
         └-> should able to open exploratory view from ux app
         └-> renders lens visualization
         └-> can do a breakdown per series
       └-: Observability alerts
         └-: With no data
           └-> Shows the no data screen
         └-: Alerts table
           └-> Renders the table
           └-> Renders the correct number of cells
           └-: Filtering
             └-> Autocompletion works
             └-> Applies filters correctly
             └-> Displays a no data state when filters produce zero results
           └-: Date selection
             └-> Correctly applies date picker selections
           └-: Flyout
             └-> Can be opened
             └-> Can be closed
             └-: When open
               └-> Displays the correct title
               └-> Displays the correct content
               └-> Displays a View in App button
           └-: Cell actions
             └-> Copy button works
             └-> Filter for value works
           └-: Bulk Actions
             └-> Only logs alert should be enable for bulk actions
             └-> validate formatting of the bulk actions button
             └-> validate functionality of the bulk actions button
             └-> validate functionality of the single alert actions button
       └-: alert workflow status
         └-> is filtered to only show "open" alerts by default
         └-> can be set to "acknowledged" using the row menu
         └-> can be filtered to only show "acknowledged" alerts using the filter button
         └-> can be set to "closed" using the row menu
         └-> can be filtered to only show "closed" alerts using the filter button
         └-> can be set to "open" using the row menu
         └-> can be filtered to only show "open" alerts using the filter button
       └-: Observability alerts pagination
         └-: When less than 10 alerts are found
           └-> Does not render page size selector
           └-> Does not render pagination controls
         └-: When 10 alerts are found
           └-: Page size selector
             └-> Renders page size selector
             └-> Default rows per page is 50
             └-> Shows up to 10 rows per page
             └-> Shows up to 25 rows per page
           └-: Pagination controls
             └-> Renders previous page button
             └-> Renders next page button
             └-> Previous page button is disabled
             └-> Goes to next page
             └-> Goes to previous page
       └-: Observability alerts / Add to case
         └-> "before all" hook: beforeTestSuite.trigger in "Observability alerts / Add to case"
         └-> "before all" hook in "Observability alerts / Add to case"
         └-: When user has all priviledges for cases
           └-> renders case options in the overflow menu
           └-> opens a flyout when Add to new case is clicked
           └-> opens a modal when Add to existing case is clicked
         └-: When user has read permissions for cases
           └-> "before all" hook: beforeTestSuite.trigger for "does not render case options in the overflow menu"
           └-> "before all" hook for "does not render case options in the overflow menu"
           └-> does not render case options in the overflow menu
             └-> "before each" hook: global before each for "does not render case options in the overflow menu"
             └- ✓ pass  (4.1s)
           └-> "after all" hook for "does not render case options in the overflow menu"
           └-> "after all" hook: afterTestSuite.trigger for "does not render case options in the overflow menu"
         └-> "after all" hook in "Observability alerts / Add to case"
         └-> "after all" hook: afterTestSuite.trigger in "Observability alerts / Add to case"
       └-> "after all" hook: afterTestSuite.trigger in "ObservabilityApp"
   │
   │13 passing (1.0m)
   │43 pending
   │
   │ warn browser[SEVERE] ERROR FETCHING BROWSR LOGS: This driver instance does not have a valid session ID (did you call WebDriver.quit()?) and may no longer be used.
   │ proc [kibana]   log   [16:23:58.305] [info][plugins-system][standard] Stopping all plugins.
   │ proc [kibana]   log   [16:23:58.306] [info][kibana-monitoring][monitoring][monitoring][plugins] Monitoring stats collection is stopped
   │ info [kibana] exited with null after 104.7 seconds
   │ info [es] stopping node ftr
   │ info [o.e.x.m.p.NativeController] [ftr] Native controller process has stopped - no new native processes can be started
   │ info [o.e.n.Node] [ftr] stopping ...
   │ info [o.e.x.w.WatcherService] [ftr] stopping watch service, reason [shutdown initiated]
   │ info [o.e.x.w.WatcherLifeCycleService] [ftr] watcher has stopped and shutdown
   │ info [o.e.n.Node] [ftr] stopped
   │ info [o.e.n.Node] [ftr] closing ...
   │ info [o.e.n.Node] [ftr] closed
   │ info [es] stopped
   │ info [es] no debug files found, assuming es did not write any
   │ info [es] cleanup complete