mirror of
https://github.com/zeek/zeek.git
synced 2025-10-02 06:38:20 +00:00

* origin/topic/robin/conn-ids: Moving uid from conn_id to connection, and making output determistic if a hash seed is given. Extending conn_id with a globally unique identifiers.
1402 lines
38 KiB
Text
1402 lines
38 KiB
Text
# $Id: bro.init 6887 2009-08-20 05:17:33Z vern $
|
|
|
|
@load const.bif.bro
|
|
@load types.bif.bro
|
|
|
|
global bro_signal: event(signal: count);
|
|
|
|
# Called (one day) if there's no handler for an internal event.
|
|
global no_handler: event(name: string, val: any);
|
|
|
|
# Type declarations
|
|
type string_array: table[count] of string;
|
|
type string_set: set[string];
|
|
type index_vec: vector of count;
|
|
type string_vec: vector of string;
|
|
|
|
type transport_proto: enum { unknown_transport, tcp, udp, icmp };
|
|
|
|
type conn_id: record {
|
|
orig_h: addr;
|
|
orig_p: port;
|
|
resp_h: addr;
|
|
resp_p: port;
|
|
} &log;
|
|
|
|
type icmp_conn: record {
|
|
orig_h: addr;
|
|
resp_h: addr;
|
|
itype: count;
|
|
icode: count;
|
|
len: count;
|
|
};
|
|
|
|
type icmp_hdr: record {
|
|
icmp_type: count; # type of message
|
|
};
|
|
|
|
type icmp_context: record {
|
|
id: conn_id;
|
|
len: count;
|
|
proto: count;
|
|
frag_offset: count;
|
|
bad_hdr_len: bool;
|
|
bad_checksum: bool;
|
|
MF: bool;
|
|
DF: bool;
|
|
};
|
|
|
|
type addr_set: set[addr];
|
|
|
|
type dns_mapping: record {
|
|
creation_time: time;
|
|
|
|
req_host: string;
|
|
req_addr: addr;
|
|
|
|
valid: bool;
|
|
hostname: string;
|
|
addrs: addr_set;
|
|
};
|
|
|
|
type ftp_port: record {
|
|
h: addr;
|
|
p: port;
|
|
valid: bool; # true if format was right
|
|
};
|
|
|
|
type endpoint: record {
|
|
size: count;
|
|
state: count;
|
|
};
|
|
|
|
type endpoint_stats: record {
|
|
num_pkts: count;
|
|
num_rxmit: count;
|
|
num_rxmit_bytes: count;
|
|
num_in_order: count;
|
|
num_OO: count;
|
|
num_repl: count;
|
|
endian_type: count;
|
|
};
|
|
|
|
type AnalyzerID: count;
|
|
|
|
type connection: record {
|
|
id: conn_id;
|
|
orig: endpoint;
|
|
resp: endpoint;
|
|
start_time: time;
|
|
duration: interval;
|
|
service: string_set; # if empty, service hasn't been determined
|
|
addl: string;
|
|
hot: count; # how hot; 0 = don't know or not hot
|
|
history: string;
|
|
uid: string;
|
|
};
|
|
|
|
type SYN_packet: record {
|
|
is_orig: bool;
|
|
DF: bool;
|
|
ttl: count;
|
|
size: count;
|
|
win_size: count;
|
|
win_scale: int;
|
|
MSS: count;
|
|
SACK_OK: bool;
|
|
};
|
|
|
|
type bro_resources: record {
|
|
version: string; # Bro version string
|
|
debug: bool; # true if compiled with --enable-debug
|
|
start_time: time; # start time of process
|
|
real_time: interval; # elapsed real time since Bro started running
|
|
user_time: interval; # user CPU seconds
|
|
system_time: interval; # system CPU seconds
|
|
mem: count; # maximum memory consumed, in KB
|
|
minor_faults: count; # page faults not requiring actual I/O
|
|
major_faults: count; # page faults requiring actual I/O
|
|
num_swap: count; # times swapped out
|
|
blocking_input: count; # blocking input operations
|
|
blocking_output: count; # blocking output operations
|
|
num_context: count; # number of involuntary context switches
|
|
|
|
num_TCP_conns: count; # current number of TCP connections
|
|
num_UDP_conns: count;
|
|
num_ICMP_conns: count;
|
|
num_fragments: count; # current number of fragments pending reassembly
|
|
num_packets: count; # total number packets processed to date
|
|
num_timers: count; # current number of pending timers
|
|
num_events_queued: count; # total number of events queued so far
|
|
num_events_dispatched: count; # same for events dispatched
|
|
|
|
max_TCP_conns: count; # maximum number of TCP connections, etc.
|
|
max_UDP_conns: count;
|
|
max_ICMP_conns: count;
|
|
max_fragments: count;
|
|
max_timers: count;
|
|
};
|
|
|
|
|
|
# Summary statistics of all DFA_State_Caches.
|
|
type matcher_stats: record {
|
|
matchers: count; # number of distinct RE matchers
|
|
dfa_states: count; # number of DFA states across all matchers
|
|
computed: count; # number of computed DFA state transitions
|
|
mem: count; # number of bytes used by DFA states
|
|
hits: count; # number of cache hits
|
|
misses: count; # number of cache misses
|
|
avg_nfa_states: count; # average # NFA states across all matchers
|
|
};
|
|
|
|
# Info provided to gap_report, and also available by get_gap_summary().
|
|
type gap_info: record {
|
|
ack_events: count; # how many ack events *could* have had gaps
|
|
ack_bytes: count; # how many bytes those covered
|
|
gap_events: count; # how many *did* have gaps
|
|
gap_bytes: count; # how many bytes were missing in the gaps:
|
|
};
|
|
|
|
# This record should be read-only.
|
|
type packet: record {
|
|
conn: connection;
|
|
is_orig: bool;
|
|
seq: count; # seq=k => it is the kth *packet* of the connection
|
|
timestamp: time;
|
|
};
|
|
|
|
type var_sizes: table[string] of count; # indexed by var's name, returns size
|
|
|
|
type script_id: record {
|
|
type_name: string;
|
|
exported: bool;
|
|
constant: bool;
|
|
enum_constant: bool;
|
|
redefinable: bool;
|
|
value: any &optional;
|
|
};
|
|
|
|
type id_table: table[string] of script_id;
|
|
|
|
# {precompile,install}_pcap_filter identify the filter by IDs
|
|
type PcapFilterID: enum { None };
|
|
|
|
type IPAddrAnonymization: enum {
|
|
KEEP_ORIG_ADDR,
|
|
SEQUENTIALLY_NUMBERED,
|
|
RANDOM_MD5,
|
|
PREFIX_PRESERVING_A50,
|
|
PREFIX_PRESERVING_MD5,
|
|
};
|
|
|
|
type IPAddrAnonymizationClass: enum {
|
|
ORIG_ADDR, # client address
|
|
RESP_ADDR, # server address
|
|
OTHER_ADDR,
|
|
};
|
|
|
|
|
|
# Events are generated by event_peer's (which may be either ourselves, or
|
|
# some remote process).
|
|
type peer_id: count;
|
|
|
|
type event_peer: record {
|
|
id: peer_id; # locally unique ID of peer (returned by connect())
|
|
host: addr;
|
|
p: port;
|
|
is_local: bool; # true if this peer describes the current process.
|
|
descr: string; # source's external_source_description
|
|
class: string &optional; # self-assigned class of the peer
|
|
};
|
|
|
|
type rotate_info: record {
|
|
old_name: string; # original filename
|
|
new_name: string; # file name after rotation
|
|
open: time; # time when opened
|
|
close: time; # time when closed
|
|
};
|
|
|
|
|
|
### The following aren't presently used, though they should be.
|
|
# # Structures needed for subsequence computations (str_smith_waterman):
|
|
# #
|
|
# type sw_variant: enum {
|
|
# SW_SINGLE,
|
|
# SW_MULTIPLE,
|
|
# };
|
|
|
|
type sw_params: record {
|
|
# Minimum size of a substring, minimum "granularity".
|
|
min_strlen: count &default = 3;
|
|
|
|
# Smith-Waterman flavor to use.
|
|
sw_variant: count &default = 0;
|
|
};
|
|
|
|
type sw_align: record {
|
|
str: string; # string a substring is part of
|
|
index: count; # at which offset
|
|
};
|
|
|
|
type sw_align_vec: vector of sw_align;
|
|
|
|
type sw_substring: record {
|
|
str: string; # a substring
|
|
aligns: sw_align_vec; # all strings of which it's a substring
|
|
new: bool; # true if start of new alignment
|
|
};
|
|
|
|
type sw_substring_vec: vector of sw_substring;
|
|
|
|
# Policy-level handling of pcap packets.
|
|
type pcap_packet: record {
|
|
ts_sec: count;
|
|
ts_usec: count;
|
|
caplen: count;
|
|
len: count;
|
|
data: string;
|
|
};
|
|
|
|
# GeoIP support.
|
|
type geo_location: record {
|
|
country_code: string;
|
|
region: string;
|
|
city: string;
|
|
latitude: double;
|
|
longitude: double;
|
|
};
|
|
|
|
type entropy_test_result: record {
|
|
entropy: double;
|
|
chi_square: double;
|
|
mean: double;
|
|
monte_carlo_pi: double;
|
|
serial_correlation: double;
|
|
};
|
|
|
|
# Prototypes of Bro built-in functions.
|
|
@load strings.bif.bro
|
|
@load bro.bif.bro
|
|
|
|
@load logging # sic! Not logging.bif.
|
|
@load logging-ascii
|
|
|
|
global bro_alarm_file: file &redef;
|
|
global alarm_hook: function(msg: string): bool &redef;
|
|
global log_file_name: function(tag: string): string &redef;
|
|
global open_log_file: function(tag: string): file &redef;
|
|
|
|
# Where to store the persistent state.
|
|
const state_dir = ".state" &redef;
|
|
|
|
# Length of the delays added when storing state incrementally.
|
|
const state_write_delay = 0.01 secs &redef;
|
|
|
|
global done_with_network = F;
|
|
event net_done(t: time) { done_with_network = T; }
|
|
|
|
const SIGHUP = 1;
|
|
event bro_signal(signal: count)
|
|
{
|
|
if ( signal == SIGHUP )
|
|
{
|
|
flush_all();
|
|
checkpoint_state();
|
|
}
|
|
}
|
|
|
|
function log_file_name(tag: string): string
|
|
{
|
|
local suffix = getenv("BRO_LOG_SUFFIX") == "" ? "log" : getenv("BRO_LOG_SUFFIX");
|
|
return fmt("%s.%s", tag, suffix);
|
|
}
|
|
|
|
function open_log_file(tag: string): file
|
|
{
|
|
return open(log_file_name(tag));
|
|
}
|
|
|
|
function add_interface(iold: string, inew: string): string
|
|
{
|
|
if ( iold == "" )
|
|
return inew;
|
|
else
|
|
return fmt("%s %s", iold, inew);
|
|
}
|
|
|
|
global interfaces = "" &add_func = add_interface;
|
|
|
|
function add_signature_file(sold: string, snew: string): string
|
|
{
|
|
if ( sold == "" )
|
|
return snew;
|
|
else
|
|
return cat(sold, " ", snew);
|
|
}
|
|
|
|
global signature_files = "" &add_func = add_signature_file;
|
|
|
|
const passive_fingerprint_file = "sigs/p0fsyn" &redef;
|
|
|
|
const ftp = 21/tcp;
|
|
const ssh = 22/tcp;
|
|
const telnet = 23/tcp;
|
|
const smtp = 25/tcp;
|
|
const domain = 53/tcp; # note, doesn't include UDP version
|
|
const gopher = 70/tcp;
|
|
const finger = 79/tcp;
|
|
const http = 80/tcp;
|
|
const ident = 113/tcp;
|
|
const bgp = 179/tcp;
|
|
const rlogin = 513/tcp;
|
|
|
|
const TCP_INACTIVE = 0;
|
|
const TCP_SYN_SENT = 1;
|
|
const TCP_SYN_ACK_SENT = 2;
|
|
const TCP_PARTIAL = 3;
|
|
const TCP_ESTABLISHED = 4;
|
|
const TCP_CLOSED = 5;
|
|
const TCP_RESET = 6;
|
|
|
|
# If true, don't verify checksums. Useful for running on altered trace
|
|
# files, and for saving a few cycles, but of course dangerous, too ...
|
|
# Note that the -C command-line option overrides the setting of this
|
|
# variable.
|
|
const ignore_checksums = F &redef;
|
|
|
|
# If true, instantiate connection state when a partial connection
|
|
# (one missing its initial establishment negotiation) is seen.
|
|
const partial_connection_ok = T &redef;
|
|
|
|
# If true, instantiate connection state when a SYN ack is seen
|
|
# but not the initial SYN (even if partial_connection_ok is false).
|
|
const tcp_SYN_ack_ok = T &redef;
|
|
|
|
# If a connection state is removed there may still be some undelivered
|
|
# data waiting in the reassembler. If true, pass this to the signature
|
|
# engine before flushing the state.
|
|
const tcp_match_undelivered = T &redef;
|
|
|
|
# Check up on the result of an initial SYN after this much time.
|
|
const tcp_SYN_timeout = 5 secs &redef;
|
|
|
|
# After a connection has closed, wait this long for further activity
|
|
# before checking whether to time out its state.
|
|
const tcp_session_timer = 6 secs &redef;
|
|
|
|
# When checking a closed connection for further activity, consider it
|
|
# inactive if there hasn't been any for this long. Complain if the
|
|
# connection is reused before this much time has elapsed.
|
|
const tcp_connection_linger = 5 secs &redef;
|
|
|
|
# Wait this long upon seeing an initial SYN before timing out the
|
|
# connection attempt.
|
|
const tcp_attempt_delay = 5 secs &redef;
|
|
|
|
# Upon seeing a normal connection close, flush state after this much time.
|
|
const tcp_close_delay = 5 secs &redef;
|
|
|
|
# Upon seeing a RST, flush state after this much time.
|
|
const tcp_reset_delay = 5 secs &redef;
|
|
|
|
# Generate a connection_partial_close event this much time after one half
|
|
# of a partial connection closes, assuming there has been no subsequent
|
|
# activity.
|
|
const tcp_partial_close_delay = 3 secs &redef;
|
|
|
|
# If a connection belongs to an application that we don't analyze,
|
|
# time it out after this interval. If 0 secs, then don't time it out.
|
|
const non_analyzed_lifetime = 0 secs &redef;
|
|
|
|
# If a connection is inactive, time it out after this interval.
|
|
# If 0 secs, then don't time it out.
|
|
const tcp_inactivity_timeout = 5 min &redef;
|
|
const udp_inactivity_timeout = 1 min &redef;
|
|
const icmp_inactivity_timeout = 1 min &redef;
|
|
|
|
# This many FINs/RSTs in a row constitutes a "storm".
|
|
const tcp_storm_thresh = 1000 &redef;
|
|
|
|
# The FINs/RSTs must come with this much time or less between them.
|
|
const tcp_storm_interarrival_thresh = 1 sec &redef;
|
|
|
|
# Maximum amount of data that might plausibly be sent in an initial
|
|
# flight (prior to receiving any acks). Used to determine whether we
|
|
# must not be seeing our peer's acks. Set to zero to turn off this
|
|
# determination.
|
|
const tcp_max_initial_window = 4096;
|
|
|
|
# If we're not seeing our peer's acks, the maximum volume of data above
|
|
# a sequence hole that we'll tolerate before assuming that there's
|
|
# been a packet drop and we should give up on tracking a connection.
|
|
# If set to zero, then we don't ever give up.
|
|
const tcp_max_above_hole_without_any_acks = 4096;
|
|
|
|
# If we've seen this much data without any of it being acked, we give up
|
|
# on that connection to avoid memory exhaustion due to buffering all that
|
|
# stuff. If set to zero, then we don't ever give up. Ideally, Bro would
|
|
# track the current window on a connection and use it to infer that data
|
|
# has in fact gone too far, but for now we just make this quite beefy.
|
|
const tcp_excessive_data_without_further_acks = 10 * 1024 * 1024;
|
|
|
|
# For services without a handler, these sets define which
|
|
# side of a connection is to be reassembled.
|
|
const tcp_reassembler_ports_orig: set[port] = {} &redef;
|
|
const tcp_reassembler_ports_resp: set[port] = {} &redef;
|
|
|
|
# These sets define destination ports for which the contents
|
|
# of the originator (responder, respectively) stream should
|
|
# be delivered via tcp_contents.
|
|
const tcp_content_delivery_ports_orig: table[port] of bool = {} &redef;
|
|
const tcp_content_delivery_ports_resp: table[port] of bool = {} &redef;
|
|
|
|
# To have all TCP orig->resp/resp->orig traffic reported via tcp_contents,
|
|
# redef these to T.
|
|
const tcp_content_deliver_all_orig = F &redef;
|
|
const tcp_content_deliver_all_resp = F &redef;
|
|
|
|
# These sets define destination ports for which the contents
|
|
# of the originator (responder, respectively) stream should
|
|
# be delivered via udp_contents.
|
|
const udp_content_delivery_ports_orig: table[port] of bool = {} &redef;
|
|
const udp_content_delivery_ports_resp: table[port] of bool = {} &redef;
|
|
|
|
# To have all UDP orig->resp/resp->orig traffic reported via udp_contents,
|
|
# redef these to T.
|
|
const udp_content_deliver_all_orig = F &redef;
|
|
const udp_content_deliver_all_resp = F &redef;
|
|
|
|
# Check for expired table entries after this amount of time
|
|
const table_expire_interval = 10 secs &redef;
|
|
|
|
# When expiring/serializing, don't work on more than this many table
|
|
# entries at a time.
|
|
const table_incremental_step = 5000 &redef;
|
|
|
|
# When expiring, wait this amount of time before checking the next chunk
|
|
# of entries.
|
|
const table_expire_delay = 0.01 secs &redef;
|
|
|
|
# Time to wait before timing out a DNS/NTP/RPC request.
|
|
const dns_session_timeout = 10 sec &redef;
|
|
const ntp_session_timeout = 300 sec &redef;
|
|
const rpc_timeout = 24 sec &redef;
|
|
|
|
# Time window for reordering packets (to deal with timestamp
|
|
# discrepency between multiple packet sources).
|
|
const packet_sort_window = 0 usecs &redef;
|
|
|
|
# How long to hold onto fragments for possible reassembly. A value
|
|
# of 0.0 means "forever", which resists evasion, but can lead to
|
|
# state accrual.
|
|
const frag_timeout = 0.0 sec &redef;
|
|
|
|
# If positive, indicates the encapsulation header size that should
|
|
# be skipped over for each captured packet ....
|
|
const encap_hdr_size = 0 &redef;
|
|
# ... or just for the following UDP port.
|
|
const tunnel_port = 0/udp &redef;
|
|
|
|
const UDP_INACTIVE = 0;
|
|
const UDP_ACTIVE = 1; # means we've seen something from this endpoint
|
|
|
|
const ENDIAN_UNKNOWN = 0;
|
|
const ENDIAN_LITTLE = 1;
|
|
const ENDIAN_BIG = 2;
|
|
const ENDIAN_CONFUSED = 3;
|
|
|
|
function append_addl(c: connection, addl: string)
|
|
{
|
|
if ( c$addl == "" )
|
|
c$addl= addl;
|
|
|
|
else if ( addl !in c$addl )
|
|
c$addl = fmt("%s %s", c$addl, addl);
|
|
}
|
|
|
|
function append_addl_marker(c: connection, addl: string, marker: string)
|
|
{
|
|
if ( c$addl == "" )
|
|
c$addl= addl;
|
|
|
|
else if ( addl !in c$addl )
|
|
c$addl = fmt("%s%s%s", c$addl, marker, addl);
|
|
}
|
|
|
|
|
|
# Values for set_contents_file's "direction" argument.
|
|
const CONTENTS_NONE = 0; # turn off recording of contents
|
|
const CONTENTS_ORIG = 1; # record originator contents
|
|
const CONTENTS_RESP = 2; # record responder contents
|
|
const CONTENTS_BOTH = 3; # record both originator and responder contents
|
|
|
|
const ICMP_UNREACH_NET = 0;
|
|
const ICMP_UNREACH_HOST = 1;
|
|
const ICMP_UNREACH_PROTOCOL = 2;
|
|
const ICMP_UNREACH_PORT = 3;
|
|
const ICMP_UNREACH_NEEDFRAG = 4;
|
|
const ICMP_UNREACH_ADMIN_PROHIB = 13;
|
|
# The above list isn't exhaustive ...
|
|
|
|
|
|
# Definitions for access to packet headers. Currently only used for
|
|
# discarders.
|
|
const IPPROTO_IP = 0; # dummy for IP
|
|
const IPPROTO_ICMP = 1; # control message protocol
|
|
const IPPROTO_IGMP = 2; # group mgmt protocol
|
|
const IPPROTO_IPIP = 4; # IP encapsulation in IP
|
|
const IPPROTO_TCP = 6; # TCP
|
|
const IPPROTO_UDP = 17; # user datagram protocol
|
|
const IPPROTO_RAW = 255; # raw IP packet
|
|
|
|
type ip_hdr: record {
|
|
hl: count; # header length (in bytes)
|
|
tos: count; # type of service
|
|
len: count; # total length
|
|
id: count; # identification
|
|
ttl: count; # time to live
|
|
p: count; # protocol
|
|
src: addr; # source address
|
|
dst: addr; # dest address
|
|
};
|
|
|
|
# TCP flags.
|
|
const TH_FIN = 1;
|
|
const TH_SYN = 2;
|
|
const TH_RST = 4;
|
|
const TH_PUSH = 8;
|
|
const TH_ACK = 16;
|
|
const TH_URG = 32;
|
|
const TH_FLAGS = 63; # (TH_FIN|TH_SYN|TH_RST|TH_ACK|TH_URG)
|
|
|
|
type tcp_hdr: record {
|
|
sport: port; # source port
|
|
dport: port; # destination port
|
|
seq: count; # sequence number
|
|
ack: count; # acknowledgement number
|
|
hl: count; # header length (in bytes)
|
|
dl: count; # data length (xxx: not in original tcphdr!)
|
|
flags: count; # flags
|
|
win: count; # window
|
|
};
|
|
|
|
type udp_hdr: record {
|
|
sport: port; # source port
|
|
dport: port; # destination port
|
|
ulen: count; # udp length
|
|
};
|
|
|
|
|
|
# Holds an ip_hdr and one of tcp_hdr, udp_hdr, or icmp_hdr.
|
|
type pkt_hdr: record {
|
|
ip: ip_hdr;
|
|
tcp: tcp_hdr &optional;
|
|
udp: udp_hdr &optional;
|
|
icmp: icmp_hdr &optional;
|
|
};
|
|
|
|
|
|
# If you add elements here, then for a given BPF filter as index, when
|
|
# a packet matching that filter is captured, the corresponding event handler
|
|
# will be invoked.
|
|
global secondary_filters: table[string] of event(filter: string, pkt: pkt_hdr)
|
|
&redef;
|
|
|
|
global discarder_maxlen = 128 &redef; # maximum amount of data passed to fnc
|
|
|
|
global discarder_check_ip: function(i: ip_hdr): bool;
|
|
global discarder_check_tcp: function(i: ip_hdr, t: tcp_hdr, d: string): bool;
|
|
global discarder_check_udp: function(i: ip_hdr, u: udp_hdr, d: string): bool;
|
|
global discarder_check_icmp: function(i: ip_hdr, ih: icmp_hdr): bool;
|
|
# End of definition of access to packet headers, discarders.
|
|
|
|
|
|
type net_stats: record {
|
|
# All counts are cumulative.
|
|
pkts_recvd: count; # pkts received by Bro
|
|
pkts_dropped: count; # pkts dropped
|
|
pkts_link: count; # pkts seen on link (not always available)
|
|
};
|
|
|
|
|
|
const watchdog_interval = 10 sec &redef;
|
|
const heartbeat_interval = 10 sec &redef;
|
|
|
|
# The maximum number of timers to expire after processing each new
|
|
# packet. The value trades off spreading out the timer expiration load
|
|
# with possibly having to hold state longer. A value of 0 means
|
|
# "process all expired timers with each new packet".
|
|
const max_timer_expires = 300 &redef;
|
|
|
|
# With a similar trade-off, this gives the number of remote events
|
|
# to process in a batch before interleaving other activity.
|
|
const max_remote_events_processed = 10 &redef;
|
|
|
|
# These need to match the definitions in Login.h.
|
|
const LOGIN_STATE_AUTHENTICATE = 0; # trying to authenticate
|
|
const LOGIN_STATE_LOGGED_IN = 1; # successful authentication
|
|
const LOGIN_STATE_SKIP = 2; # skip any further processing
|
|
const LOGIN_STATE_CONFUSED = 3; # we're confused
|
|
|
|
# It would be nice to replace these function definitions with some
|
|
# form of parameterized types.
|
|
function min_double(a: double, b: double): double { return a < b ? a : b; }
|
|
function max_double(a: double, b: double): double { return a > b ? a : b; }
|
|
function min_interval(a: interval, b: interval): interval { return a < b ? a : b; }
|
|
function max_interval(a: interval, b: interval): interval { return a > b ? a : b; }
|
|
function min_count(a: count, b: count): count { return a < b ? a : b; }
|
|
function max_count(a: count, b: count): count { return a > b ? a : b; }
|
|
|
|
global skip_authentication: set[string] &redef;
|
|
global direct_login_prompts: set[string] &redef;
|
|
global login_prompts: set[string] &redef;
|
|
global login_non_failure_msgs: set[string] &redef;
|
|
global login_failure_msgs: set[string] &redef;
|
|
global login_success_msgs: set[string] &redef;
|
|
global login_timeouts: set[string] &redef;
|
|
|
|
type mime_header_rec: record {
|
|
name: string;
|
|
value: string;
|
|
};
|
|
type mime_header_list: table[count] of mime_header_rec;
|
|
global mime_segment_length = 1024 &redef;
|
|
global mime_segment_overlap_length = 0 &redef;
|
|
|
|
type pm_mapping: record {
|
|
program: count;
|
|
version: count;
|
|
p: port;
|
|
};
|
|
|
|
type pm_mappings: table[count] of pm_mapping;
|
|
|
|
type pm_port_request: record {
|
|
program: count;
|
|
version: count;
|
|
is_tcp: bool;
|
|
};
|
|
|
|
type pm_callit_request: record {
|
|
program: count;
|
|
version: count;
|
|
proc: count;
|
|
arg_size: count;
|
|
};
|
|
|
|
# See const.bif
|
|
# const RPC_SUCCESS = 0;
|
|
# const RPC_PROG_UNAVAIL = 1;
|
|
# const RPC_PROG_MISMATCH = 2;
|
|
# const RPC_PROC_UNAVAIL = 3;
|
|
# const RPC_GARBAGE_ARGS = 4;
|
|
# const RPC_SYSTEM_ERR = 5;
|
|
# const RPC_TIMEOUT = 6;
|
|
# const RPC_AUTH_ERROR = 7;
|
|
# const RPC_UNKNOWN_ERROR = 8;
|
|
|
|
const RPC_status = {
|
|
[RPC_SUCCESS] = "ok",
|
|
[RPC_PROG_UNAVAIL] = "prog unavail",
|
|
[RPC_PROG_MISMATCH] = "mismatch",
|
|
[RPC_PROC_UNAVAIL] = "proc unavail",
|
|
[RPC_GARBAGE_ARGS] = "garbage args",
|
|
[RPC_SYSTEM_ERR] = "system err",
|
|
[RPC_TIMEOUT] = "timeout",
|
|
[RPC_AUTH_ERROR] = "auth error",
|
|
[RPC_UNKNOWN_ERROR] = "unknown"
|
|
};
|
|
|
|
|
|
type nfs3_file_type: enum {
|
|
NFS3_REG, NFS3_DIR, NFS3_BLK, NFS3_CHR, NFS3_LNK, NFS3_SOCK, NFS3_FIFO,
|
|
};
|
|
|
|
type nfs3_attrs: record {
|
|
ftype: nfs3_file_type;
|
|
mode: count;
|
|
nlink: count;
|
|
uid: count;
|
|
gid: count;
|
|
size: double;
|
|
used: double;
|
|
rdev1: count;
|
|
rdev2: count;
|
|
fsid: double;
|
|
fileid: double;
|
|
atime: time;
|
|
mtime: time;
|
|
ctime: time;
|
|
};
|
|
|
|
type nfs3_opt_attrs: record {
|
|
attrs: nfs3_attrs &optional;
|
|
};
|
|
|
|
type nfs3_lookup_args: record {
|
|
fh: string; # file handle of directory in which to search
|
|
name: string; # name of file to look for in directory
|
|
};
|
|
|
|
type nfs3_lookup_reply: record {
|
|
fh: string; # file handle of object looked up
|
|
file_attr: nfs3_opt_attrs; # optional attributes associated w/ file
|
|
dir_attr: nfs3_opt_attrs; # optional attributes associated w/ dir.
|
|
};
|
|
|
|
type nfs3_fsstat: record {
|
|
attrs: nfs3_opt_attrs;
|
|
tbytes: double;
|
|
fbytes: double;
|
|
abytes: double;
|
|
tfiles: double;
|
|
ffiles: double;
|
|
afiles: double;
|
|
invarsec: interval;
|
|
};
|
|
|
|
|
|
type ntp_msg: record {
|
|
id: count;
|
|
code: count;
|
|
stratum: count;
|
|
poll: count;
|
|
precision: int;
|
|
distance: interval;
|
|
dispersion: interval;
|
|
ref_t: time;
|
|
originate_t: time;
|
|
receive_t: time;
|
|
xmit_t: time;
|
|
};
|
|
|
|
|
|
# Maps Samba command numbers to descriptive names.
|
|
global samba_cmds: table[count] of string &redef
|
|
&default = function(c: count): string
|
|
{ return fmt("samba-unknown-%d", c); };
|
|
|
|
type smb_hdr : record {
|
|
command: count;
|
|
status: count;
|
|
flags: count;
|
|
flags2: count;
|
|
tid: count;
|
|
pid: count;
|
|
uid: count;
|
|
mid: count;
|
|
};
|
|
|
|
type smb_trans : record {
|
|
word_count: count;
|
|
total_param_count: count;
|
|
total_data_count: count;
|
|
max_param_count: count;
|
|
max_data_count: count;
|
|
max_setup_count: count;
|
|
# flags: count;
|
|
# timeout: count;
|
|
param_count: count;
|
|
param_offset: count;
|
|
data_count: count;
|
|
data_offset: count;
|
|
setup_count: count;
|
|
setup0: count;
|
|
setup1: count;
|
|
setup2: count;
|
|
setup3: count;
|
|
byte_count: count;
|
|
parameters: string;
|
|
};
|
|
|
|
type smb_trans_data : record {
|
|
data : string;
|
|
};
|
|
|
|
type smb_tree_connect : record {
|
|
flags: count;
|
|
password: string;
|
|
path: string;
|
|
service: string;
|
|
};
|
|
|
|
type smb_negotiate : table[count] of string;
|
|
|
|
# A list of router addresses offered by the server.
|
|
type dhcp_router_list: table[count] of addr;
|
|
|
|
type dhcp_msg: record {
|
|
op: count; # message OP code. 1 = BOOTREQUEST, 2 = BOOTREPLY
|
|
m_type: count; # the type of DHCP message
|
|
xid: count; # transaction ID of a DHCP session
|
|
h_addr: string; # hardware address of the client
|
|
ciaddr: addr; # original IP address of the client
|
|
yiaddr: addr; # IP address assigned to the client
|
|
};
|
|
|
|
type dns_msg: record {
|
|
id: count;
|
|
|
|
opcode: count;
|
|
rcode: count;
|
|
|
|
QR: bool;
|
|
AA: bool;
|
|
TC: bool;
|
|
RD: bool;
|
|
RA: bool;
|
|
Z: count;
|
|
|
|
num_queries: count;
|
|
num_answers: count;
|
|
num_auth: count;
|
|
num_addl: count;
|
|
};
|
|
|
|
type dns_soa: record {
|
|
mname: string; # primary source of data for zone
|
|
rname: string; # mailbox for responsible person
|
|
serial: count; # version number of zone
|
|
refresh: interval; # seconds before refreshing
|
|
retry: interval; # how long before retrying failed refresh
|
|
expire: interval; # when zone no longer authoritative
|
|
minimum: interval; # minimum TTL to use when exporting
|
|
};
|
|
|
|
type dns_edns_additional: record {
|
|
query: string;
|
|
qtype: count;
|
|
t: count;
|
|
payload_size: count;
|
|
extended_rcode: count;
|
|
version: count;
|
|
z_field: count;
|
|
TTL: interval;
|
|
is_query: count;
|
|
};
|
|
|
|
type dns_tsig_additional: record {
|
|
query: string;
|
|
qtype: count;
|
|
alg_name: string;
|
|
sig: string;
|
|
time_signed: time;
|
|
fudge: time;
|
|
orig_id: count;
|
|
rr_error: count;
|
|
is_query: count;
|
|
};
|
|
|
|
# Different values for "answer_type" in the following. DNS_QUERY
|
|
# shouldn't occur, it's just for completeness.
|
|
const DNS_QUERY = 0;
|
|
const DNS_ANS = 1;
|
|
const DNS_AUTH = 2;
|
|
const DNS_ADDL = 3;
|
|
|
|
type dns_answer: record {
|
|
answer_type: count;
|
|
query: string;
|
|
qtype: count;
|
|
qclass: count;
|
|
TTL: interval;
|
|
};
|
|
|
|
# For servers in these sets, omit processing the AUTH or ADDL records
|
|
# they include in their replies.
|
|
global dns_skip_auth: set[addr] &redef;
|
|
global dns_skip_addl: set[addr] &redef;
|
|
|
|
# If the following are true, then all AUTH or ADDL records are skipped.
|
|
global dns_skip_all_auth = T &redef;
|
|
global dns_skip_all_addl = T &redef;
|
|
|
|
# If a DNS request includes more than this many queries, assume it's
|
|
# non-DNS traffic and do not process it. Set to 0 to turn off this
|
|
# functionality.
|
|
global dns_max_queries = 5;
|
|
|
|
# The maxiumum size in bytes for an SSL cipherspec. If we see a packet that
|
|
# has bigger cipherspecs, we won't do a comparisons of cipherspecs.
|
|
const ssl_max_cipherspec_size = 68 &redef;
|
|
|
|
# SSL and X.509 types.
|
|
type cipher_suites_list: set[count];
|
|
type SSL_sessionID: table[count] of count;
|
|
type X509_extension: table[count] of string;
|
|
type X509: record {
|
|
issuer: string;
|
|
subject: string;
|
|
orig_addr: addr;
|
|
};
|
|
|
|
type http_stats_rec: record {
|
|
num_requests: count;
|
|
num_replies: count;
|
|
request_version: double;
|
|
reply_version: double;
|
|
};
|
|
|
|
type http_message_stat: record {
|
|
start: time; # when the request/reply line was complete
|
|
interrupted: bool; # whether the message is interrupted
|
|
finish_msg: string; # reason phrase if interrupted
|
|
body_length: count; # length of body processed
|
|
# (before finished/interrupted)
|
|
content_gap_length: count; # total len of gaps within body_length
|
|
header_length: count; # length of headers
|
|
# (including the req/reply line,
|
|
# but not CR/LF's)
|
|
};
|
|
|
|
global http_entity_data_delivery_size = 1500 &redef;
|
|
|
|
# Truncate URIs longer than this to prevent over-long URIs (usually sent
|
|
# by worms) from slowing down event processing. A value of -1 means "do
|
|
# not truncate".
|
|
const truncate_http_URI = -1 &redef;
|
|
|
|
# IRC-related globals to which the event engine is sensitive.
|
|
type irc_join_info: record {
|
|
nick: string;
|
|
channel: string;
|
|
password: string;
|
|
usermode: string;
|
|
};
|
|
type irc_join_list: set[irc_join_info];
|
|
global irc_servers : set[addr] &redef;
|
|
|
|
# Stepping-stone globals.
|
|
const stp_delta: interval &redef;
|
|
const stp_idle_min: interval &redef;
|
|
|
|
# Don't do analysis on these sources. Used to avoid overload from scanners.
|
|
global stp_skip_src: set[addr] &redef;
|
|
|
|
const interconn_min_interarrival: interval &redef;
|
|
const interconn_max_interarrival: interval &redef;
|
|
const interconn_max_keystroke_pkt_size: count &redef;
|
|
const interconn_default_pkt_size: count &redef;
|
|
const interconn_stat_period: interval &redef;
|
|
const interconn_stat_backoff: double &redef;
|
|
|
|
type interconn_endp_stats: record {
|
|
num_pkts: count;
|
|
num_keystrokes_two_in_row: count;
|
|
num_normal_interarrivals: count;
|
|
num_8k0_pkts: count;
|
|
num_8k4_pkts: count;
|
|
is_partial: bool;
|
|
num_bytes: count;
|
|
num_7bit_ascii: count;
|
|
num_lines: count;
|
|
num_normal_lines: count;
|
|
};
|
|
|
|
const backdoor_stat_period: interval &redef;
|
|
const backdoor_stat_backoff: double &redef;
|
|
|
|
type backdoor_endp_stats: record {
|
|
is_partial: bool;
|
|
num_pkts: count;
|
|
num_8k0_pkts: count;
|
|
num_8k4_pkts: count;
|
|
num_lines: count;
|
|
num_normal_lines: count;
|
|
num_bytes: count;
|
|
num_7bit_ascii: count;
|
|
};
|
|
|
|
type signature_state: record {
|
|
id: string; # ID of the signature
|
|
conn: connection; # Current connection
|
|
is_orig: bool; # True if current endpoint is originator
|
|
payload_size: count; # Payload size of the first pkt of curr. endpoint
|
|
|
|
};
|
|
|
|
type software_version: record {
|
|
major: int; # Major version number
|
|
minor: int; # Minor version number
|
|
minor2: int; # Minor subversion number
|
|
addl: string; # Additional version string (e.g. "beta42")
|
|
};
|
|
|
|
type software: record {
|
|
name: string; # Unique name of a software, e.g., "OS"
|
|
version: software_version;
|
|
};
|
|
|
|
# The following describe the quality of signature matches used
|
|
# for passive fingerprinting.
|
|
type OS_version_inference: enum {
|
|
direct_inference, generic_inference, fuzzy_inference,
|
|
};
|
|
|
|
type OS_version: record {
|
|
genre: string; # Linux, Windows, AIX, ...
|
|
detail: string; # kernel version or such
|
|
dist: count; # how far is the host away from the sensor (TTL)?
|
|
match_type: OS_version_inference;
|
|
};
|
|
|
|
# Defines for which subnets we should do passive fingerprinting.
|
|
global generate_OS_version_event: set[subnet] &redef;
|
|
|
|
# Type used to report load samples via load_sample(). For now,
|
|
# it's a set of names (event names, source file names, and perhaps
|
|
# <source file, line number>'s, which were seen during the sample.
|
|
type load_sample_info: set[string];
|
|
|
|
# NetFlow-related data structures.
|
|
|
|
# The following provides a mean to sort together flow headers and flow
|
|
# records at the script level. rcvr_id equals the name of the file
|
|
# (e.g., netflow.dat) or the socket address (e.g., 127.0.0.1:5555),
|
|
# or an explicit name if specified to -y or -Y; pdu_id is just a serial
|
|
# number, ignoring any overflows.
|
|
type nfheader_id: record {
|
|
rcvr_id: string;
|
|
pdu_id: count;
|
|
};
|
|
|
|
type nf_v5_header: record {
|
|
h_id: nfheader_id; # ID for sorting, per the above
|
|
cnt: count;
|
|
sysuptime: interval; # router's uptime
|
|
exporttime: time; # when the data was exported
|
|
flow_seq: count;
|
|
eng_type: count;
|
|
eng_id: count;
|
|
sample_int: count;
|
|
exporter: addr;
|
|
};
|
|
|
|
type nf_v5_record: record {
|
|
h_id: nfheader_id;
|
|
id: conn_id;
|
|
nexthop: addr;
|
|
input: count;
|
|
output: count;
|
|
pkts: count;
|
|
octets: count;
|
|
first: time;
|
|
last: time;
|
|
tcpflag_fin: bool; # Taken from tcpflags in NF V5; or directly.
|
|
tcpflag_syn: bool;
|
|
tcpflag_rst: bool;
|
|
tcpflag_psh: bool;
|
|
tcpflag_ack: bool;
|
|
tcpflag_urg: bool;
|
|
proto: count;
|
|
tos: count;
|
|
src_as: count;
|
|
dst_as: count;
|
|
src_mask: count;
|
|
dst_mask: count;
|
|
};
|
|
|
|
|
|
# The peer record and the corresponding set type used by the
|
|
# BitTorrent analyzer.
|
|
type bittorrent_peer: record {
|
|
h: addr;
|
|
p: port;
|
|
};
|
|
type bittorrent_peer_set: set[bittorrent_peer];
|
|
|
|
# The benc value record and the corresponding table type used by the
|
|
# BitTorrenttracker analyzer. Note that "benc" = Bencode ("Bee-Encode"),
|
|
# per http://en.wikipedia.org/wiki/Bencode.
|
|
type bittorrent_benc_value: record {
|
|
i: int &optional;
|
|
s: string &optional;
|
|
d: string &optional;
|
|
l: string &optional;
|
|
};
|
|
type bittorrent_benc_dir: table[string] of bittorrent_benc_value;
|
|
|
|
# The header table type used by the bittorrenttracker analyzer.
|
|
type bt_tracker_headers: table[string] of string;
|
|
|
|
@load event.bif.bro
|
|
|
|
function subst(s: string, from: pattern, to: string): string
|
|
{
|
|
local p = split_all(s, from);
|
|
|
|
for ( p_i in p )
|
|
if ( p_i % 2 == 0 )
|
|
p[p_i] = to;
|
|
|
|
return cat_string_array(p);
|
|
}
|
|
|
|
|
|
type pattern_match_result: record {
|
|
matched: bool;
|
|
str: string;
|
|
off: count;
|
|
};
|
|
|
|
function match_pattern(s: string, p:pattern): pattern_match_result
|
|
{
|
|
local a = split_n(s, p, T, 1);
|
|
|
|
if ( length(a) == 1 )
|
|
# no match
|
|
return [$matched = F, $str = "", $off = 0];
|
|
else
|
|
return [$matched = T, $str = a[2], $off = byte_len(a[1]) + 1];
|
|
}
|
|
|
|
function cut_tail(s: string, tail_len: count): string
|
|
{
|
|
local len = byte_len(s);
|
|
|
|
if ( tail_len > len )
|
|
tail_len = len;
|
|
|
|
return sub_bytes(s, 1, int_to_count(len - tail_len));
|
|
}
|
|
|
|
# Given a string, returns an escaped version. This means that
|
|
# (1) any occurrences of any character in "chars" are escaped using '\', and
|
|
# (2) any '\'s are likewise escaped.
|
|
function string_escape(s: string, chars: string): string
|
|
{
|
|
s = subst_string(s, "\\", "\\\\");
|
|
for ( c in chars )
|
|
s = subst_string(s, c, cat("\\", c));
|
|
return s;
|
|
}
|
|
|
|
@load pcap.bro
|
|
|
|
# Rotate logs every x seconds.
|
|
const log_rotate_interval = 0 sec &redef;
|
|
|
|
# If set, rotate logs at given time + i * log_rotate_interval.
|
|
# (string is time in 24h format, e.g., "18:00").
|
|
const log_rotate_base_time = "" &redef;
|
|
|
|
# Rotate logs when they reach this size (in bytes). Note, the
|
|
# parameter is a double rather than a count to enable easy expression
|
|
# of large values such as 1e7 or exceeding 2^32.
|
|
const log_max_size = 0.0 &redef;
|
|
|
|
# Default public key for encrypting log files.
|
|
const log_encryption_key = "<undefined>" &redef;
|
|
|
|
# Write profiling info into this file.
|
|
global profiling_file: file &redef;
|
|
|
|
# Update interval for profiling (0 disables).
|
|
const profiling_interval = 0 secs &redef;
|
|
|
|
# Multiples of profiling_interval at which (expensive) memory
|
|
# profiling is done (0 disables).
|
|
const expensive_profiling_multiple = 0 &redef;
|
|
|
|
# If true, then write segment profiling information (very high volume!)
|
|
# in addition to statistics.
|
|
const segment_profiling = F &redef;
|
|
|
|
# Output packet profiling information every <freq> secs (mode 1),
|
|
# every <freq> packets (mode 2), or every <freq> bytes (mode 3).
|
|
# Mode 0 disables.
|
|
type pkt_profile_modes: enum {
|
|
PKT_PROFILE_MODE_NONE,
|
|
PKT_PROFILE_MODE_SECS,
|
|
PKT_PROFILE_MODE_PKTS,
|
|
PKT_PROFILE_MODE_BYTES,
|
|
};
|
|
const pkt_profile_mode = PKT_PROFILE_MODE_NONE &redef;
|
|
|
|
# Frequency associated with packet profiling.
|
|
const pkt_profile_freq = 0.0 &redef;
|
|
|
|
# File where packet profiles are logged.
|
|
global pkt_profile_file: file &redef;
|
|
|
|
# Rate at which to generate load_sample events, *if* you've also
|
|
# defined a load_sample handler. Units are inverse number of packets;
|
|
# e.g., a value of 20 means "roughly one in every 20 packets".
|
|
global load_sample_freq = 20 &redef;
|
|
|
|
# Rate at which to generate gap_report events assessing to what
|
|
# degree the measurement process appears to exhibit loss.
|
|
const gap_report_freq = 1.0 sec &redef;
|
|
|
|
# Globals associated with entire-run statistics on gaps (useful
|
|
# for final summaries).
|
|
|
|
# The CA certificate file to authorize remote Bros.
|
|
const ssl_ca_certificate = "<undefined>" &redef;
|
|
|
|
# File containing our private key and our certificate.
|
|
const ssl_private_key = "<undefined>" &redef;
|
|
|
|
# The passphrase for our private key. Keeping this undefined
|
|
# causes Bro to prompt for the passphrase.
|
|
const ssl_passphrase = "<undefined>" &redef;
|
|
|
|
# Whether the Bro-level packet filter drops packets per default or not.
|
|
const packet_filter_default = F &redef;
|
|
|
|
# Maximum size of regular expression groups for signature matching.
|
|
const sig_max_group_size = 50 &redef;
|
|
|
|
# If true, send logger messages to syslog.
|
|
const enable_syslog = T &redef;
|
|
|
|
# This is transmitted to peers receiving our events.
|
|
const peer_description = "" &redef;
|
|
|
|
# If true, broadcast events/state received from one peer to other peers.
|
|
# NOTE: These options are only temporary. They will disappear when we get a
|
|
# more sophisticated script-level communication framework.
|
|
const forward_remote_events = F &redef;
|
|
const forward_remote_state_changes = F &redef;
|
|
|
|
const PEER_ID_NONE = 0;
|
|
|
|
# Whether to use the connection tracker.
|
|
const use_connection_compressor = T &redef;
|
|
|
|
# Whether compressor should handle refused connections itself.
|
|
const cc_handle_resets = F &redef;
|
|
|
|
# Whether compressor should only take care of initial SYNs.
|
|
# (By default on, this is basically "connection compressor lite".)
|
|
const cc_handle_only_syns = T &redef;
|
|
|
|
# Whether compressor instantiates full state when originator sends a
|
|
# non-control packet.
|
|
const cc_instantiate_on_data = F &redef;
|
|
|
|
# Signature payload pattern types
|
|
const SIG_PATTERN_PAYLOAD = 0;
|
|
const SIG_PATTERN_HTTP = 1;
|
|
const SIG_PATTERN_FTP = 2;
|
|
const SIG_PATTERN_FINGER = 3;
|
|
|
|
# Log-levels for remote_log.
|
|
# Eventually we should create a general logging framework and merge these in.
|
|
const REMOTE_LOG_INFO = 1;
|
|
const REMOTE_LOG_ERROR = 2;
|
|
|
|
# Sources for remote_log.
|
|
const REMOTE_SRC_CHILD = 1;
|
|
const REMOTE_SRC_PARENT = 2;
|
|
const REMOTE_SRC_SCRIPT = 3;
|
|
|
|
# Synchronize trace processing at a regular basis in pseudo-realtime mode.
|
|
const remote_trace_sync_interval = 0 secs &redef;
|
|
|
|
# Number of peers across which to synchronize trace processing.
|
|
const remote_trace_sync_peers = 0 &redef;
|
|
|
|
# Whether for &synchronized state to send the old value as a consistency check.
|
|
const remote_check_sync_consistency = F &redef;
|
|
|
|
# Prepend the peer description, if set.
|
|
function prefixed_id(id: count): string
|
|
{
|
|
if ( peer_description == "" )
|
|
return fmt("%s", id);
|
|
else
|
|
return cat(peer_description, "-", id);
|
|
}
|
|
|
|
# Analyzer tags. The core automatically defines constants
|
|
# ANALYZER_<analyzer-name>*, e.g., ANALYZER_HTTP.
|
|
type AnalyzerTag: count;
|
|
|
|
# DPM configuration.
|
|
|
|
type dpd_protocol_config: record {
|
|
ports: set[port] &optional;
|
|
};
|
|
|
|
const dpd_config: table[AnalyzerTag] of dpd_protocol_config = {} &redef;
|
|
|
|
# Reassemble the beginning of all TCP connections before doing
|
|
# signature-matching for protocol detection.
|
|
const dpd_reassemble_first_packets = T &redef;
|
|
|
|
# Size of per-connection buffer in bytes. If the buffer is full, data is
|
|
# deleted and lost to analyzers that are activated afterwards.
|
|
const dpd_buffer_size = 1024 &redef;
|
|
|
|
# If true, stops signature matching if dpd_buffer_size has been reached.
|
|
const dpd_match_only_beginning = T &redef;
|
|
|
|
# If true, don't consider any ports for deciding which analyzer to use.
|
|
const dpd_ignore_ports = F &redef;
|
|
|
|
# Ports which the core considers being likely used by servers.
|
|
const likely_server_ports: set[port] &redef;
|
|
|
|
# Set of all ports for which we know an analyzer.
|
|
global dpd_analyzer_ports: table[port] of set[AnalyzerTag];
|
|
|
|
event bro_init()
|
|
{
|
|
for ( a in dpd_config )
|
|
for ( p in dpd_config[a]$ports )
|
|
{
|
|
if ( p !in dpd_analyzer_ports )
|
|
{
|
|
local empty_set: set[AnalyzerTag];
|
|
dpd_analyzer_ports[p] = empty_set;
|
|
}
|
|
|
|
add dpd_analyzer_ports[p][a];
|
|
}
|
|
}
|
|
|
|
@load server-ports
|
|
|
|
# Per-incident timer managers are drained after this amount of inactivity.
|
|
const timer_mgr_inactivity_timeout = 1 min &redef;
|
|
|
|
# Time-out for expected connections.
|
|
const expected_connection_timeout = 5 min &redef;
|
|
|
|
# If true, output profiling for time-machine queries.
|
|
const time_machine_profiling = F &redef;
|
|
|
|
# If true, warns about unused event handlers at startup.
|
|
const check_for_unused_event_handlers = F &redef;
|
|
|
|
# If true, dumps all invoked event handlers at startup.
|
|
const dump_used_event_handlers = F &redef;
|
|
|
|
# If true, we suppress prints to local files if we have a receiver for
|
|
# print_hook events. Ignored for files with a &disable_print_hook attribute.
|
|
const suppress_local_output = F &redef;
|
|
|
|
# Holds the filename of the trace file given with -w (empty if none).
|
|
const trace_output_file = "";
|
|
|
|
# If a trace file is given, dump *all* packets seen by Bro into it.
|
|
# By default, Bro applies (very few) heuristics to reduce the volume.
|
|
# A side effect of setting this to true is that we can write the
|
|
# packets out before we actually process them, which can be helpful
|
|
# for debugging in case the analysis triggers a crash.
|
|
const record_all_packets = F &redef;
|
|
|
|
# Some connections (e.g., SSH) retransmit the acknowledged last
|
|
# byte to keep the connection alive. If ignore_keep_alive_rexmit
|
|
# is set to T, such retransmissions will be excluded in the rexmit
|
|
# counter in conn_stats.
|
|
const ignore_keep_alive_rexmit = F &redef;
|
|
|
|
# Skip HTTP data portions for performance considerations (the skipped
|
|
# portion will not go through TCP reassembly).
|
|
const skip_http_data = F &redef;
|
|
|
|
# Whether the analysis engine parses IP packets encapsulated in
|
|
# UDP tunnels. See also: udp_tunnel_port, policy/udp-tunnel.bro.
|
|
const parse_udp_tunnels = F &redef;
|