# @TEST-DOC: Query the Prometheus endpoint and smoke check that zeek_version_info{...} is contained in the response for all cluster nodes. # Not compilable to C++ due to globals being initialized to a record that # has an opaque type as a field. # @TEST-REQUIRES: test "${ZEEK_USE_CPP}" != "1" # @TEST-REQUIRES: which jq # # @TEST-PORT: BROKER_MANAGER_PORT # @TEST-PORT: BROKER_LOGGER1_PORT # @TEST-PORT: BROKER_PROXY1_PORT # @TEST-PORT: BROKER_WORKER1_PORT # # @TEST-PORT: METRICS_PORT1 # @TEST-PORT: METRICS_PORT2 # @TEST-PORT: METRICS_PORT3 # @TEST-PORT: METRICS_PORT4 # # @TEST-REQUIRES: which curl # @TEST-EXEC: zeek --parse-only %INPUT # @TEST-EXEC: btest-bg-run manager ZEEKPATH=$ZEEKPATH:.. CLUSTER_NODE=manager zeek -b %INPUT # @TEST-EXEC: btest-bg-run logger-1 ZEEKPATH=$ZEEKPATH:.. CLUSTER_NODE=logger-1 zeek -b %INPUT # @TEST-EXEC: btest-bg-run proxy-1 ZEEKPATH=$ZEEKPATH:.. CLUSTER_NODE=proxy-1 zeek -b %INPUT # @TEST-EXEC: btest-bg-run worker-1 ZEEKPATH=$ZEEKPATH:.. CLUSTER_NODE=worker-1 zeek -b %INPUT # @TEST-EXEC: btest-bg-wait 30 # @TEST-EXEC: btest-diff manager/services.out # @TEST-START-FILE cluster-layout.zeek redef Cluster::nodes = { ["manager"] = [$node_type=Cluster::MANAGER, $ip=127.0.0.1, $p=to_port(getenv("BROKER_MANAGER_PORT")), $metrics_port=to_port(getenv("METRICS_PORT1"))], ["logger-1"] = [$node_type=Cluster::LOGGER, $ip=127.0.0.1, $p=to_port(getenv("BROKER_LOGGER1_PORT")), $manager="manager", $metrics_port=to_port(getenv("METRICS_PORT2"))], ["proxy-1"] = [$node_type=Cluster::PROXY, $ip=127.0.0.1, $p=to_port(getenv("BROKER_PROXY1_PORT")), $manager="manager", $metrics_port=to_port(getenv("METRICS_PORT3"))], ["worker-1"] = [$node_type=Cluster::WORKER, $ip=127.0.0.1, $p=to_port(getenv("BROKER_WORKER1_PORT")), $manager="manager", $metrics_port=to_port(getenv("METRICS_PORT4"))], }; # @TEST-END-FILE # @TEST-START-FILE request-services.sh #! /usr/bin/env bash # This script makes repeat curl requests to find all of the metrics data from the # hosts listed in the services output from the manager, and outputs it all into a # single file. services_url=$1 output_file=$2 services_data=$(curl -s -m 5 ${services_url}) for host in $(echo ${services_data} | jq -r '.[0].targets[]' | sort); do metrics=$(curl -m 5 --trace trace-${host}.out http://${host}/metrics) if [ $? -eq 0 ] ; then version_info=$(echo ${metrics} | grep -Eo "zeek_version_info\{[^}]+\}" | grep -o 'node="[^"]*"') echo ${version_info} >> ${output_file}; else echo "Failed to request data from ${host}" >> ${output_file} fi done # @TEST-END-FILE @load policy/frameworks/cluster/experimental @load base/frameworks/telemetry # So the cluster nodes don't terminate right away. #redef exit_only_after_terminate=T; @if ( Cluster::node == "manager" ) @load base/utils/exec # Query the Prometheus endpoint using curl for testing, oh my. event run_test() { local services_url = fmt("http://localhost:%s/services.json", port_to_count(Telemetry::metrics_port)); local req_cmd = fmt("sh ../request-services.sh %s %s", services_url, "services.out"); when [req_cmd] ( local result = Exec::run([$cmd=req_cmd]) ) { if ( result$exit_code != 0 ) { # This is bad. print "ERROR: Failed to request service information"; exit(1); } terminate(); } timeout 10sec { # This is bad. print "ERROR: Timed out requesting service information"; exit(1); } } # Use a dynamic metrics port for testing to avoid colliding on 9911/tcp # when running tests in parallel. event zeek_init() { print Cluster::node, "Telemetry::metrics_port from cluster config", Telemetry::metrics_port; } event Cluster::Experimental::cluster_started() { # Run the test once all nodes are up schedule 2 secs { run_test() }; } @endif # If any node goes down, terminate() right away. event Cluster::node_down(name: string, id: string) { print fmt("node_down on %s", Cluster::node); terminate(); }