mirror of
https://github.com/zeek/zeek.git
synced 2025-10-04 07:38:19 +00:00
Merge remote-tracking branch 'origin/master' into topic/bif_cleanup
Conflicts: src/bro.bif
This commit is contained in:
commit
6c255d13ff
247 changed files with 6038 additions and 3503 deletions
161
CHANGES
161
CHANGES
|
@ -1,4 +1,165 @@
|
||||||
|
|
||||||
|
2.0 | 2012-01-11 13:52:22 -0800
|
||||||
|
|
||||||
|
* Adding script reference documentation. (The Team).
|
||||||
|
|
||||||
|
2.0-beta-194 | 2012-01-10 10:44:32 -0800
|
||||||
|
|
||||||
|
* Added an option for filtering out URLs before they are turned into
|
||||||
|
HTTP::Incorrect_File_Type notices. (Seth Hall)
|
||||||
|
|
||||||
|
* Fix ref counting bug in BIFs that call internal_type. Addresses
|
||||||
|
#740. (Jon Siwek)
|
||||||
|
|
||||||
|
* Adding back the stats.bro file. (Seth Hall)
|
||||||
|
|
||||||
|
|
||||||
|
2.0-beta-188 | 2012-01-10 09:49:29 -0800
|
||||||
|
|
||||||
|
* Change SFTP/SCP log rotators to use 4-digit year in filenames
|
||||||
|
Fixes #745. (Jon Siwek)
|
||||||
|
|
||||||
|
* Adding back the stats.bro file. Addresses #656. (Seth Hall)
|
||||||
|
|
||||||
|
2.0-beta-185 | 2012-01-09 18:00:50 -0800
|
||||||
|
|
||||||
|
* Tweaks for OpenBSD support. (Jon Siwek)
|
||||||
|
|
||||||
|
2.0-beta-181 | 2012-01-08 20:49:04 -0800
|
||||||
|
|
||||||
|
* Add SFTP log postprocessor that transfers logs to remote hosts.
|
||||||
|
Addresses #737. (Jon Siwek)
|
||||||
|
|
||||||
|
* Add FAQ entry about disabling NIC offloading features. (Jon Siwek)
|
||||||
|
|
||||||
|
* Add a file NEWS with release notes. (Robin Sommer)
|
||||||
|
|
||||||
|
2.0-beta-177 | 2012-01-05 15:01:07 -0800
|
||||||
|
|
||||||
|
* Replace the --snaplen/-l command line option with a
|
||||||
|
scripting-layer option called "snaplen" (which can also be
|
||||||
|
redefined on the command line, e.g. `bro -i eth0 snaplen=65535`).
|
||||||
|
|
||||||
|
* Reduce snaplen default from 65535 to old default of 8192. Fixes
|
||||||
|
#720. (Jon Siwek)
|
||||||
|
|
||||||
|
2.0-beta-174 | 2012-01-04 12:47:10 -0800
|
||||||
|
|
||||||
|
* SSL improvements. (Seth Hall)
|
||||||
|
|
||||||
|
- Added the ssl_session_ticket_handshake event back.
|
||||||
|
|
||||||
|
- Fixed a few bugs.
|
||||||
|
|
||||||
|
- Removed the SSLv2.cc file since it's not used.
|
||||||
|
|
||||||
|
2.0-beta-169 | 2012-01-04 12:44:39 -0800
|
||||||
|
|
||||||
|
* Tuning the pretty-printed alarm mails, which now include the
|
||||||
|
covered time range into the subject. (Robin Sommer)
|
||||||
|
|
||||||
|
* Adding top-level "test" target to Makefile. (Robin Sommer)
|
||||||
|
|
||||||
|
* Adding SWIG as dependency to INSTALL. (Robin Sommer)
|
||||||
|
|
||||||
|
2.0-beta-155 | 2012-01-03 15:42:32 -0800
|
||||||
|
|
||||||
|
* Remove dead code related to record type inheritance. (Jon Siwek)
|
||||||
|
|
||||||
|
2.0-beta-152 | 2012-01-03 14:51:34 -0800
|
||||||
|
|
||||||
|
* Notices now record the transport-layer protocol. (Bernhard Amann)
|
||||||
|
|
||||||
|
2.0-beta-150 | 2012-01-03 14:42:45 -0800
|
||||||
|
|
||||||
|
* CMake 2.6 top-level 'install' target compat. Fixes #729. (Jon Siwek)
|
||||||
|
|
||||||
|
* Minor fixes to test process. Addresses #298.
|
||||||
|
|
||||||
|
* Increase timeout interval of communication-related btests. (Jon Siwek)
|
||||||
|
|
||||||
|
2.0-beta-145 | 2011-12-19 11:37:15 -0800
|
||||||
|
|
||||||
|
* Empty fields are now logged as "(empty)" by default. (Robin
|
||||||
|
Sommer)
|
||||||
|
|
||||||
|
* In log headers, only escape information when necessary. (Robin
|
||||||
|
Sommer)
|
||||||
|
|
||||||
|
2.0-beta-139 | 2011-12-19 07:06:29 -0800
|
||||||
|
|
||||||
|
* The hostname notice email extension works now, plus a general
|
||||||
|
mechanism for adding delayed information to notices. (Seth Hall)
|
||||||
|
|
||||||
|
* Fix &default fields in records not being initialized in coerced
|
||||||
|
assignments. Addresses #722. (Jon Siwek)
|
||||||
|
|
||||||
|
* Make log headers include the type of data stored inside a set or
|
||||||
|
vector ("vector[string]"). (Bernhard Amann)
|
||||||
|
|
||||||
|
2.0-beta-126 | 2011-12-18 15:18:05 -0800
|
||||||
|
|
||||||
|
* DNS updates. (Seth Hall)
|
||||||
|
|
||||||
|
- Fixed some bugs with capturing data in the base DNS script.
|
||||||
|
|
||||||
|
- Answers and TTLs are now vectors.
|
||||||
|
|
||||||
|
- A warning that was being generated (dns_reply_seen_after_done)
|
||||||
|
from transaction ID reuse is fixed.
|
||||||
|
|
||||||
|
* SSL updates. (Seth Hall)
|
||||||
|
|
||||||
|
- Added is_orig fields to the SSL events and adapted script.
|
||||||
|
|
||||||
|
- Added a field named last_alert to the SSL log.
|
||||||
|
|
||||||
|
- The x509_certificate function has an is_orig field now instead
|
||||||
|
of is_server and its position in the argument list has moved.
|
||||||
|
|
||||||
|
- A bit of reorganization and cleanup in the core analyzer. (Seth
|
||||||
|
Hall)
|
||||||
|
|
||||||
|
2.0-beta-121 | 2011-12-18 15:10:15 -0800
|
||||||
|
|
||||||
|
* Enable warnings for malformed Broxygen xref roles. (Jon Siwek)
|
||||||
|
|
||||||
|
* Fix Broxygen confusing scoped IDs at start of line as function
|
||||||
|
parameter. (Jon Siwek)
|
||||||
|
|
||||||
|
* Allow Broxygen markup "##<" for more general use. (Jon Siwek)
|
||||||
|
|
||||||
|
2.0-beta-116 | 2011-12-16 02:38:27 -0800
|
||||||
|
|
||||||
|
* Cleanup some misc Broxygen css/js stuff. (Jon Siwek)
|
||||||
|
|
||||||
|
* Add search box to Broxygen docs. Fixes #726. (Jon Siwek)
|
||||||
|
|
||||||
|
* Fixed major bug with cluster synchronization, which was not
|
||||||
|
working. (Seth Hall)
|
||||||
|
|
||||||
|
* Fix missing action in notice policy for looking up GeoIP data.
|
||||||
|
(Jon Siwek)
|
||||||
|
|
||||||
|
* Better persistent state configuration warning messages (fixes
|
||||||
|
#433). (Jon Siwek)
|
||||||
|
|
||||||
|
* Renaming HTTP::SQL_Injection_Attack_Against to
|
||||||
|
HTTP::SQL_Injection_Victim. (Seth Hall).
|
||||||
|
|
||||||
|
* Fixed DPD signatures for IRC. Fixes #311. (Seth Hall)
|
||||||
|
|
||||||
|
* Removing Off_Port_Protocol_Found notice. (Seth Hall)
|
||||||
|
|
||||||
|
* Teach Broxygen to more generally reference attribute values by name. (Jon Siwek)
|
||||||
|
|
||||||
|
* SSH::Interesting_Hostname_Login cleanup. Fixes #664. (Seth Hall)
|
||||||
|
|
||||||
|
* Fixed bug that was causing the malware hash registry script to
|
||||||
|
break. (Seth Hall)
|
||||||
|
|
||||||
|
* Remove remnant of libmagic optionality. (Jon Siwek)
|
||||||
|
|
||||||
2.0-beta-98 | 2011-12-07 08:12:08 -0800
|
2.0-beta-98 | 2011-12-07 08:12:08 -0800
|
||||||
|
|
||||||
* Adapting test-suite's diff-all so that it expands globs in both
|
* Adapting test-suite's diff-all so that it expands globs in both
|
||||||
|
|
2
COPYING
2
COPYING
|
@ -1,4 +1,4 @@
|
||||||
Copyright (c) 1995-2011, The Regents of the University of California
|
Copyright (c) 1995-2012, The Regents of the University of California
|
||||||
through the Lawrence Berkeley National Laboratory and the
|
through the Lawrence Berkeley National Laboratory and the
|
||||||
International Computer Science Institute. All rights reserved.
|
International Computer Science Institute. All rights reserved.
|
||||||
|
|
||||||
|
|
16
INSTALL
16
INSTALL
|
@ -14,10 +14,11 @@ before you begin:
|
||||||
|
|
||||||
* OpenSSL (headers and libraries) http://www.openssl.org
|
* OpenSSL (headers and libraries) http://www.openssl.org
|
||||||
|
|
||||||
* Libmagic For identifying file types (e.g., in FTP transfers).
|
* SWIG http://www.swig.org
|
||||||
|
|
||||||
* Libz For decompressing HTTP bodies by the HTTP analyzer, and for
|
* Libmagic
|
||||||
compressed Bro-to-Bro communication.
|
|
||||||
|
* Libz
|
||||||
|
|
||||||
Bro can make uses of some optional libraries if they are found at
|
Bro can make uses of some optional libraries if they are found at
|
||||||
installation time:
|
installation time:
|
||||||
|
@ -27,11 +28,13 @@ installation time:
|
||||||
Bro also needs the following tools, but on most systems they will
|
Bro also needs the following tools, but on most systems they will
|
||||||
already come preinstalled:
|
already come preinstalled:
|
||||||
|
|
||||||
|
* Bash (For Bro Control).
|
||||||
* BIND8 (headers and libraries)
|
* BIND8 (headers and libraries)
|
||||||
* Bison (GNU Parser Generator)
|
* Bison (GNU Parser Generator)
|
||||||
* Flex (Fast Lexical Analyzer)
|
* Flex (Fast Lexical Analyzer)
|
||||||
* Perl (Used only during the Bro build process)
|
* Perl (Used only during the Bro build process)
|
||||||
|
|
||||||
|
|
||||||
Installation
|
Installation
|
||||||
============
|
============
|
||||||
|
|
||||||
|
@ -64,13 +67,16 @@ except for ``aux/bro-aux`` will also be built and installed by doing
|
||||||
``--disable-*`` options that can be given to the configure script to
|
``--disable-*`` options that can be given to the configure script to
|
||||||
turn off unwanted auxiliary projects.
|
turn off unwanted auxiliary projects.
|
||||||
|
|
||||||
|
OpenBSD users, please see our `FAQ
|
||||||
|
<http://www.bro-ids.org/documentation/faq.html>` if you are having
|
||||||
|
problems installing Bro.
|
||||||
|
|
||||||
Running Bro
|
Running Bro
|
||||||
===========
|
===========
|
||||||
|
|
||||||
Bro is a complex program and it takes a bit of time to get familiar
|
Bro is a complex program and it takes a bit of time to get familiar
|
||||||
with it. A good place for newcomers to start is the
|
with it. A good place for newcomers to start is the Quickstart Guide
|
||||||
:doc:`quick start guide <quickstart>`.
|
at http://www.bro-ids.org/documentation/quickstart.html.
|
||||||
|
|
||||||
For developers that wish to run Bro directly from the ``build/``
|
For developers that wish to run Bro directly from the ``build/``
|
||||||
directory (i.e., without performing ``make install``), they will have
|
directory (i.e., without performing ``make install``), they will have
|
||||||
|
|
7
Makefile
7
Makefile
|
@ -2,7 +2,7 @@
|
||||||
# A simple static wrapper for a number of standard Makefile targets,
|
# A simple static wrapper for a number of standard Makefile targets,
|
||||||
# mostly just forwarding to build/Makefile. This is provided only for
|
# mostly just forwarding to build/Makefile. This is provided only for
|
||||||
# convenience and supports only a subset of what CMake's Makefile
|
# convenience and supports only a subset of what CMake's Makefile
|
||||||
# to offer. For more, execute that one directly.
|
# offers. For more, execute that one directly.
|
||||||
#
|
#
|
||||||
|
|
||||||
BUILD=build
|
BUILD=build
|
||||||
|
@ -14,7 +14,7 @@ HAVE_MODULES=git submodule | grep -v cmake >/dev/null
|
||||||
all: configured
|
all: configured
|
||||||
$(MAKE) -C $(BUILD) $@
|
$(MAKE) -C $(BUILD) $@
|
||||||
|
|
||||||
install: configured
|
install: configured all
|
||||||
$(MAKE) -C $(BUILD) $@
|
$(MAKE) -C $(BUILD) $@
|
||||||
|
|
||||||
install-aux: configured
|
install-aux: configured
|
||||||
|
@ -60,6 +60,9 @@ bindist:
|
||||||
distclean:
|
distclean:
|
||||||
rm -rf $(BUILD)
|
rm -rf $(BUILD)
|
||||||
|
|
||||||
|
test:
|
||||||
|
@(cd testing && make )
|
||||||
|
|
||||||
configured:
|
configured:
|
||||||
@test -d $(BUILD) || ( echo "Error: No build/ directory found. Did you run configure?" && exit 1 )
|
@test -d $(BUILD) || ( echo "Error: No build/ directory found. Did you run configure?" && exit 1 )
|
||||||
@test -e $(BUILD)/Makefile || ( echo "Error: No build/Makefile found. Did you run configure?" && exit 1 )
|
@test -e $(BUILD)/Makefile || ( echo "Error: No build/Makefile found. Did you run configure?" && exit 1 )
|
||||||
|
|
64
NEWS
Normal file
64
NEWS
Normal file
|
@ -0,0 +1,64 @@
|
||||||
|
|
||||||
|
Release Notes
|
||||||
|
=============
|
||||||
|
|
||||||
|
This document summarizes the most important changes in the current Bro
|
||||||
|
release. For a complete list of changes, see the ``CHANGES`` file.
|
||||||
|
|
||||||
|
Bro 2.0
|
||||||
|
-------
|
||||||
|
|
||||||
|
As the version number jump suggests, Bro 2.0 is a major upgrade and
|
||||||
|
lots of things have changed. We have assembled a separate upgrade
|
||||||
|
guide with the most important changes compared to Bro 1.5 at
|
||||||
|
http://www.bro-ids.org/documentation/upgrade.html. You can find
|
||||||
|
the offline version of that document in ``doc/upgrade.rst.``.
|
||||||
|
|
||||||
|
Compared to the earlier 2.0 Beta version, the major changes in the
|
||||||
|
final release are:
|
||||||
|
|
||||||
|
* The default scripts now come with complete reference
|
||||||
|
documentation. See
|
||||||
|
http://www.bro-ids.org/documentation/index.html.
|
||||||
|
|
||||||
|
* libz and libmagic are now required dependencies.
|
||||||
|
|
||||||
|
* Reduced snaplen default from 65535 to old default of 8192. The
|
||||||
|
large value was introducing performance problems on many
|
||||||
|
systems.
|
||||||
|
|
||||||
|
* Replaced the --snaplen/-l command line option with a
|
||||||
|
scripting-layer option called "snaplen". The new option can also
|
||||||
|
be redefined on the command line, e.g. ``bro -i eth0
|
||||||
|
snaplen=65535``.
|
||||||
|
|
||||||
|
* Reintroduced the BRO_LOG_SUFFIX environment variable that the
|
||||||
|
ASCII logger now respects to add a suffix to the log files it
|
||||||
|
creates.
|
||||||
|
|
||||||
|
* The ASCII logs now include further header information, and
|
||||||
|
fields set to an empty value are now logged as ``(empty)`` by
|
||||||
|
default (instead of ``-``, which is already used for fields that
|
||||||
|
are not set at all).
|
||||||
|
|
||||||
|
* Some NOTICES were renamed, and the signatures of some SSL events
|
||||||
|
have changed.
|
||||||
|
|
||||||
|
* bro-cut got some new capabilities:
|
||||||
|
|
||||||
|
- If no field names are given on the command line, we now pass
|
||||||
|
through all fields.
|
||||||
|
|
||||||
|
- New options -u/-U for time output in UTC.
|
||||||
|
|
||||||
|
- New option -F to give output field separator.
|
||||||
|
|
||||||
|
* Broccoli supports more types internally, allowing to send
|
||||||
|
complex records.
|
||||||
|
|
||||||
|
* Many smaller bug fixes, portability improvements, and general
|
||||||
|
polishing across all modules.
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
10
README
10
README
|
@ -4,13 +4,15 @@ Bro Network Security Monitor
|
||||||
|
|
||||||
Bro is a powerful framework for network analysis and security
|
Bro is a powerful framework for network analysis and security
|
||||||
monitoring. Please see the INSTALL file for installation instructions
|
monitoring. Please see the INSTALL file for installation instructions
|
||||||
and pointers for getting started. For more documentation, research
|
and pointers for getting started. NEWS contains release notes for the
|
||||||
publications, and community contact information, see Bro's home page:
|
current version, and CHANGES has the complete history of changes.
|
||||||
|
Please see COPYING for licensing information.
|
||||||
|
|
||||||
|
For more documentation, research publications, and community contact
|
||||||
|
information, please see Bro's home page:
|
||||||
|
|
||||||
http://www.bro-ids.org
|
http://www.bro-ids.org
|
||||||
|
|
||||||
Please see COPYING for licensing information.
|
|
||||||
|
|
||||||
On behalf of the Bro Development Team,
|
On behalf of the Bro Development Team,
|
||||||
|
|
||||||
Vern Paxson & Robin Sommer,
|
Vern Paxson & Robin Sommer,
|
||||||
|
|
2
VERSION
2
VERSION
|
@ -1 +1 @@
|
||||||
2.0-beta-98
|
2.0
|
||||||
|
|
|
@ -1 +1 @@
|
||||||
Subproject commit 82bd9613fb869e44f5f0d7929bdd9a88cde84077
|
Subproject commit aa1aa85ddcf524ffcfcf9efa5277bfac341871f7
|
|
@ -1 +1 @@
|
||||||
Subproject commit 4d387ce660468b44df99d4c87d6016ae4ed2fdc4
|
Subproject commit 1ac5080de06bb29f2bc9fc73bd7dbbf945c7dc81
|
|
@ -1 +1 @@
|
||||||
Subproject commit 05d51b55dd5a8430151cff74f7876bd73385d5cd
|
Subproject commit 722ca1f266d8cf2f79a89e24b6b0a15174365ccd
|
|
@ -1 +1 @@
|
||||||
Subproject commit 406c689c88997c4459eaf36bd2c65bce2e847c2d
|
Subproject commit 84636789bf48376d7278b31e0bd7be19ff72c566
|
|
@ -1 +1 @@
|
||||||
Subproject commit 38890e851416fa9fc827a1d36f06c4cb9f7d4e69
|
Subproject commit 5350e4652b44ce1fbd9fffe1228d097fb04247cd
|
2
cmake
2
cmake
|
@ -1 +1 @@
|
||||||
Subproject commit f0f7958639bb921985c1f58f1186da4b49b5d54d
|
Subproject commit ca4ed1a237215765ce9a7f2bc4b57b56958039ef
|
14
config.h.in
14
config.h.in
|
@ -17,9 +17,6 @@
|
||||||
/* We are on a Linux system */
|
/* We are on a Linux system */
|
||||||
#cmakedefine HAVE_LINUX
|
#cmakedefine HAVE_LINUX
|
||||||
|
|
||||||
/* Define if you have the <magic.h> header file. */
|
|
||||||
#cmakedefine HAVE_MAGIC_H
|
|
||||||
|
|
||||||
/* Define if you have the `mallinfo' function. */
|
/* Define if you have the `mallinfo' function. */
|
||||||
#cmakedefine HAVE_MALLINFO
|
#cmakedefine HAVE_MALLINFO
|
||||||
|
|
||||||
|
@ -35,8 +32,8 @@
|
||||||
/* Define if you have the <net/ethernet.h> header file. */
|
/* Define if you have the <net/ethernet.h> header file. */
|
||||||
#cmakedefine HAVE_NET_ETHERNET_H
|
#cmakedefine HAVE_NET_ETHERNET_H
|
||||||
|
|
||||||
/* We are on a OpenBSD system */
|
/* Define if you have the <net/ethertypes.h> header file. */
|
||||||
#cmakedefine HAVE_OPENBSD
|
#cmakedefine HAVE_NET_ETHERTYPES_H
|
||||||
|
|
||||||
/* have os-proto.h */
|
/* have os-proto.h */
|
||||||
#cmakedefine HAVE_OS_PROTO_H
|
#cmakedefine HAVE_OS_PROTO_H
|
||||||
|
@ -148,3 +145,10 @@
|
||||||
|
|
||||||
/* Define u_int8_t */
|
/* Define u_int8_t */
|
||||||
#cmakedefine u_int8_t @u_int8_t@
|
#cmakedefine u_int8_t @u_int8_t@
|
||||||
|
|
||||||
|
/* OpenBSD's bpf.h may not declare this data link type, but it's supposed to be
|
||||||
|
used consistently for the same purpose on all platforms. */
|
||||||
|
#cmakedefine HAVE_DLT_PPP_SERIAL
|
||||||
|
#ifndef HAVE_DLT_PPP_SERIAL
|
||||||
|
#define DLT_PPP_SERIAL @DLT_PPP_SERIAL@
|
||||||
|
#endif
|
||||||
|
|
1
doc/.gitignore
vendored
1
doc/.gitignore
vendored
|
@ -1 +1,2 @@
|
||||||
html
|
html
|
||||||
|
*.pyc
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
Documentation
|
Documentation
|
||||||
=============
|
=============
|
||||||
|
|
||||||
This directory contains Bro documentation in reStructured text format
|
This directory contains Bro documentation in reStructuredText format
|
||||||
(see http://docutils.sourceforge.net/rst.html).
|
(see http://docutils.sourceforge.net/rst.html).
|
||||||
|
|
||||||
It is the root of a Sphinx source tree and can be modified to add more
|
It is the root of a Sphinx source tree and can be modified to add more
|
||||||
|
|
1
doc/_static/960.css
vendored
Normal file
1
doc/_static/960.css
vendored
Normal file
File diff suppressed because one or more lines are too long
513
doc/_static/basic.css
vendored
Normal file
513
doc/_static/basic.css
vendored
Normal file
|
@ -0,0 +1,513 @@
|
||||||
|
/*
|
||||||
|
* basic.css
|
||||||
|
* ~~~~~~~~~
|
||||||
|
*
|
||||||
|
* Sphinx stylesheet -- basic theme.
|
||||||
|
*
|
||||||
|
* :copyright: Copyright 2007-2011 by the Sphinx team, see AUTHORS.
|
||||||
|
* :license: BSD, see LICENSE for details.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* -- main layout ----------------------------------------------------------- */
|
||||||
|
|
||||||
|
div.clearer {
|
||||||
|
clear: both;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- relbar ---------------------------------------------------------------- */
|
||||||
|
|
||||||
|
div.related {
|
||||||
|
width: 100%;
|
||||||
|
font-size: 90%;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.related h3 {
|
||||||
|
display: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.related ul {
|
||||||
|
margin: 0;
|
||||||
|
padding: 0 0 0 10px;
|
||||||
|
list-style: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.related li {
|
||||||
|
display: inline;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.related li.right {
|
||||||
|
float: right;
|
||||||
|
margin-right: 5px;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- sidebar --------------------------------------------------------------- */
|
||||||
|
|
||||||
|
div.sphinxsidebarwrapper {
|
||||||
|
padding: 10px 5px 0 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar {
|
||||||
|
float: left;
|
||||||
|
width: 230px;
|
||||||
|
margin-left: -100%;
|
||||||
|
font-size: 90%;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar ul {
|
||||||
|
list-style: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar ul ul,
|
||||||
|
div.sphinxsidebar ul.want-points {
|
||||||
|
margin-left: 20px;
|
||||||
|
list-style: square;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar ul ul {
|
||||||
|
margin-top: 0;
|
||||||
|
margin-bottom: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar form {
|
||||||
|
margin-top: 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar input {
|
||||||
|
border: 1px solid #98dbcc;
|
||||||
|
font-family: sans-serif;
|
||||||
|
font-size: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar input[type="text"] {
|
||||||
|
width: 170px;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar input[type="submit"] {
|
||||||
|
width: 30px;
|
||||||
|
}
|
||||||
|
|
||||||
|
img {
|
||||||
|
border: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- search page ----------------------------------------------------------- */
|
||||||
|
|
||||||
|
ul.search {
|
||||||
|
margin: 10px 0 0 20px;
|
||||||
|
padding: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul.search li {
|
||||||
|
padding: 5px 0 5px 20px;
|
||||||
|
background-image: url(file.png);
|
||||||
|
background-repeat: no-repeat;
|
||||||
|
background-position: 0 7px;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul.search li a {
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul.search li div.context {
|
||||||
|
color: #888;
|
||||||
|
margin: 2px 0 0 30px;
|
||||||
|
text-align: left;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul.keywordmatches li.goodmatch a {
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- index page ------------------------------------------------------------ */
|
||||||
|
|
||||||
|
table.contentstable {
|
||||||
|
width: 90%;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.contentstable p.biglink {
|
||||||
|
line-height: 150%;
|
||||||
|
}
|
||||||
|
|
||||||
|
a.biglink {
|
||||||
|
font-size: 1.3em;
|
||||||
|
}
|
||||||
|
|
||||||
|
span.linkdescr {
|
||||||
|
font-style: italic;
|
||||||
|
padding-top: 5px;
|
||||||
|
font-size: 90%;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- general index --------------------------------------------------------- */
|
||||||
|
|
||||||
|
table.indextable {
|
||||||
|
width: 100%;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.indextable td {
|
||||||
|
text-align: left;
|
||||||
|
vertical-align: top;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.indextable dl, table.indextable dd {
|
||||||
|
margin-top: 0;
|
||||||
|
margin-bottom: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.indextable tr.pcap {
|
||||||
|
height: 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.indextable tr.cap {
|
||||||
|
margin-top: 10px;
|
||||||
|
background-color: #f2f2f2;
|
||||||
|
}
|
||||||
|
|
||||||
|
img.toggler {
|
||||||
|
margin-right: 3px;
|
||||||
|
margin-top: 3px;
|
||||||
|
cursor: pointer;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.modindex-jumpbox {
|
||||||
|
border-top: 1px solid #ddd;
|
||||||
|
border-bottom: 1px solid #ddd;
|
||||||
|
margin: 1em 0 1em 0;
|
||||||
|
padding: 0.4em;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.genindex-jumpbox {
|
||||||
|
border-top: 1px solid #ddd;
|
||||||
|
border-bottom: 1px solid #ddd;
|
||||||
|
margin: 1em 0 1em 0;
|
||||||
|
padding: 0.4em;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- general body styles --------------------------------------------------- */
|
||||||
|
|
||||||
|
a.headerlink {
|
||||||
|
visibility: hidden;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.body p.caption {
|
||||||
|
text-align: inherit;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.body td {
|
||||||
|
text-align: left;
|
||||||
|
}
|
||||||
|
|
||||||
|
.field-list ul {
|
||||||
|
padding-left: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.first {
|
||||||
|
margin-top: 0 !important;
|
||||||
|
}
|
||||||
|
|
||||||
|
p.rubric {
|
||||||
|
margin-top: 30px;
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
img.align-left, .figure.align-left, object.align-left {
|
||||||
|
clear: left;
|
||||||
|
float: left;
|
||||||
|
margin-right: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
img.align-right, .figure.align-right, object.align-right {
|
||||||
|
clear: right;
|
||||||
|
float: right;
|
||||||
|
margin-left: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
img.align-center, .figure.align-center, object.align-center {
|
||||||
|
display: block;
|
||||||
|
margin-left: auto;
|
||||||
|
margin-right: auto;
|
||||||
|
}
|
||||||
|
|
||||||
|
.align-left {
|
||||||
|
text-align: left;
|
||||||
|
}
|
||||||
|
|
||||||
|
.align-center {
|
||||||
|
text-align: center;
|
||||||
|
}
|
||||||
|
|
||||||
|
.align-right {
|
||||||
|
text-align: right;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- sidebars -------------------------------------------------------------- */
|
||||||
|
|
||||||
|
div.sidebar {
|
||||||
|
margin: 0 0 0.5em 1em;
|
||||||
|
border: 1px solid #ddb;
|
||||||
|
padding: 7px 7px 0 7px;
|
||||||
|
background-color: #ffe;
|
||||||
|
width: 40%;
|
||||||
|
float: right;
|
||||||
|
}
|
||||||
|
|
||||||
|
p.sidebar-title {
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- topics ---------------------------------------------------------------- */
|
||||||
|
|
||||||
|
div.topic {
|
||||||
|
border: 1px solid #ccc;
|
||||||
|
padding: 7px 7px 0 7px;
|
||||||
|
margin: 10px 0 10px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
p.topic-title {
|
||||||
|
font-size: 1.1em;
|
||||||
|
font-weight: bold;
|
||||||
|
margin-top: 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- admonitions ----------------------------------------------------------- */
|
||||||
|
|
||||||
|
div.admonition {
|
||||||
|
margin-top: 10px;
|
||||||
|
margin-bottom: 10px;
|
||||||
|
padding: 7px;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.admonition dt {
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.admonition dl {
|
||||||
|
margin-bottom: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
p.admonition-title {
|
||||||
|
margin: 0px 10px 5px 0px;
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.body p.centered {
|
||||||
|
text-align: center;
|
||||||
|
margin-top: 25px;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- tables ---------------------------------------------------------------- */
|
||||||
|
|
||||||
|
table.field-list td, table.field-list th {
|
||||||
|
border: 0 !important;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.footnote td, table.footnote th {
|
||||||
|
border: 0 !important;
|
||||||
|
}
|
||||||
|
|
||||||
|
th {
|
||||||
|
text-align: left;
|
||||||
|
padding-right: 5px;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.citation {
|
||||||
|
border-left: solid 1px gray;
|
||||||
|
margin-left: 1px;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.citation td {
|
||||||
|
border-bottom: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- other body styles ----------------------------------------------------- */
|
||||||
|
|
||||||
|
ol.arabic {
|
||||||
|
list-style: decimal;
|
||||||
|
}
|
||||||
|
|
||||||
|
ol.loweralpha {
|
||||||
|
list-style: lower-alpha;
|
||||||
|
}
|
||||||
|
|
||||||
|
ol.upperalpha {
|
||||||
|
list-style: upper-alpha;
|
||||||
|
}
|
||||||
|
|
||||||
|
ol.lowerroman {
|
||||||
|
list-style: lower-roman;
|
||||||
|
}
|
||||||
|
|
||||||
|
ol.upperroman {
|
||||||
|
list-style: upper-roman;
|
||||||
|
}
|
||||||
|
|
||||||
|
dd p {
|
||||||
|
margin-top: 0px;
|
||||||
|
}
|
||||||
|
|
||||||
|
dd ul, dd table {
|
||||||
|
margin-bottom: 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
dd {
|
||||||
|
margin-top: 3px;
|
||||||
|
margin-bottom: 10px;
|
||||||
|
margin-left: 30px;
|
||||||
|
}
|
||||||
|
|
||||||
|
dt:target, .highlighted {
|
||||||
|
background-color: #fbe54e;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl.glossary dt {
|
||||||
|
font-weight: bold;
|
||||||
|
font-size: 1.1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.field-list ul {
|
||||||
|
margin: 0;
|
||||||
|
padding-left: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.field-list p {
|
||||||
|
margin: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
.refcount {
|
||||||
|
color: #060;
|
||||||
|
}
|
||||||
|
|
||||||
|
.optional {
|
||||||
|
font-size: 1.3em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.versionmodified {
|
||||||
|
font-style: italic;
|
||||||
|
}
|
||||||
|
|
||||||
|
.system-message {
|
||||||
|
background-color: #fda;
|
||||||
|
padding: 5px;
|
||||||
|
border: 3px solid red;
|
||||||
|
}
|
||||||
|
|
||||||
|
.footnote:target {
|
||||||
|
background-color: #ffa;
|
||||||
|
}
|
||||||
|
|
||||||
|
.line-block {
|
||||||
|
display: block;
|
||||||
|
margin-top: 1em;
|
||||||
|
margin-bottom: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.line-block .line-block {
|
||||||
|
margin-top: 0;
|
||||||
|
margin-bottom: 0;
|
||||||
|
margin-left: 1.5em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.guilabel, .menuselection {
|
||||||
|
font-family: sans-serif;
|
||||||
|
}
|
||||||
|
|
||||||
|
.accelerator {
|
||||||
|
text-decoration: underline;
|
||||||
|
}
|
||||||
|
|
||||||
|
.classifier {
|
||||||
|
font-style: oblique;
|
||||||
|
}
|
||||||
|
|
||||||
|
abbr, acronym {
|
||||||
|
border-bottom: dotted 1px;
|
||||||
|
cursor: help;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- code displays --------------------------------------------------------- */
|
||||||
|
|
||||||
|
pre {
|
||||||
|
overflow: auto;
|
||||||
|
overflow-y: hidden; /* fixes display issues on Chrome browsers */
|
||||||
|
}
|
||||||
|
|
||||||
|
td.linenos pre {
|
||||||
|
padding: 5px 0px;
|
||||||
|
border: 0;
|
||||||
|
background-color: transparent;
|
||||||
|
color: #aaa;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.highlighttable {
|
||||||
|
margin-left: 0.5em;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.highlighttable td {
|
||||||
|
padding: 0 0.5em 0 0.5em;
|
||||||
|
}
|
||||||
|
|
||||||
|
tt.descname {
|
||||||
|
background-color: transparent;
|
||||||
|
font-weight: bold;
|
||||||
|
# font-size: 1.2em;
|
||||||
|
}
|
||||||
|
|
||||||
|
tt.descclassname {
|
||||||
|
background-color: transparent;
|
||||||
|
}
|
||||||
|
|
||||||
|
tt.xref, a tt {
|
||||||
|
background-color: transparent;
|
||||||
|
# font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
h1 tt, h2 tt, h3 tt, h4 tt, h5 tt, h6 tt {
|
||||||
|
background-color: transparent;
|
||||||
|
}
|
||||||
|
|
||||||
|
.viewcode-link {
|
||||||
|
float: right;
|
||||||
|
}
|
||||||
|
|
||||||
|
.viewcode-back {
|
||||||
|
float: right;
|
||||||
|
font-family: sans-serif;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.viewcode-block:target {
|
||||||
|
margin: -1px -10px;
|
||||||
|
padding: 0 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- math display ---------------------------------------------------------- */
|
||||||
|
|
||||||
|
img.math {
|
||||||
|
vertical-align: middle;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.body div.math p {
|
||||||
|
text-align: center;
|
||||||
|
}
|
||||||
|
|
||||||
|
span.eqno {
|
||||||
|
float: right;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* -- printout stylesheet --------------------------------------------------- */
|
||||||
|
|
||||||
|
@media print {
|
||||||
|
div.document,
|
||||||
|
div.documentwrapper,
|
||||||
|
div.bodywrapper {
|
||||||
|
margin: 0 !important;
|
||||||
|
width: 100%;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.sphinxsidebar,
|
||||||
|
div.related,
|
||||||
|
div.footer,
|
||||||
|
#top-link {
|
||||||
|
display: none;
|
||||||
|
}
|
||||||
|
}
|
149
doc/_static/broxygen-extra.css
vendored
149
doc/_static/broxygen-extra.css
vendored
|
@ -1,5 +1,15 @@
|
||||||
.highlight {
|
|
||||||
background-color: #ffffff;
|
a.toc-backref {
|
||||||
|
color: #333;
|
||||||
|
}
|
||||||
|
|
||||||
|
h1, h2, h3, h4, h5, h6,
|
||||||
|
h1 a, h2 a, h3 a, h4 a, h5 a, h6 a {
|
||||||
|
padding:0 0 0px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul {
|
||||||
|
padding-bottom: 0px;
|
||||||
}
|
}
|
||||||
|
|
||||||
h1 {
|
h1 {
|
||||||
|
@ -13,3 +23,138 @@ h1 {
|
||||||
color: #000;
|
color: #000;
|
||||||
border-bottom: 0px;
|
border-bottom: 0px;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
th.field-name
|
||||||
|
{
|
||||||
|
white-space:nowrap;
|
||||||
|
}
|
||||||
|
|
||||||
|
h2 {
|
||||||
|
margin-top: 50px;
|
||||||
|
padding-bottom: 5px;
|
||||||
|
margin-bottom: 30px;
|
||||||
|
border-bottom: 1px solid;
|
||||||
|
border-color: #aaa;
|
||||||
|
font-style: normal;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.section h3 {
|
||||||
|
font-style: normal;
|
||||||
|
}
|
||||||
|
|
||||||
|
h3 {
|
||||||
|
font-size: 20px;
|
||||||
|
margin-top: 40px;
|
||||||
|
margin-bottom: 0¡px;
|
||||||
|
font-weight: bold;
|
||||||
|
font-style: normal;
|
||||||
|
}
|
||||||
|
|
||||||
|
h3.widgettitle {
|
||||||
|
font-style: normal;
|
||||||
|
}
|
||||||
|
|
||||||
|
h4 {
|
||||||
|
font-size:18px;
|
||||||
|
font-style: normal;
|
||||||
|
margin-bottom: 0em;
|
||||||
|
margin-top: 40px;
|
||||||
|
font-style: italic;
|
||||||
|
}
|
||||||
|
|
||||||
|
h5 {
|
||||||
|
font-size:16px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h6 {
|
||||||
|
font-size:15px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.toc-backref {
|
||||||
|
color: #333;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents ul {
|
||||||
|
padding-bottom: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl.namespace {
|
||||||
|
display: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl dt {
|
||||||
|
font-weight: normal;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.docutils tbody {
|
||||||
|
margin: 1em 1em 1em 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.docutils td {
|
||||||
|
padding: 5pt 5pt 5pt 5pt;
|
||||||
|
font-size: 14px;
|
||||||
|
border-left: 0;
|
||||||
|
border-right: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl pre {
|
||||||
|
font-size: 14px;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.docutils th {
|
||||||
|
padding: 5pt 5pt 5pt 5pt;
|
||||||
|
font-size: 14px;
|
||||||
|
font-style: normal;
|
||||||
|
border-left: 0;
|
||||||
|
border-right: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.docutils tr:first-child td {
|
||||||
|
#border-top: 1px solid #aaa;
|
||||||
|
}
|
||||||
|
|
||||||
|
.download {
|
||||||
|
font-family:"Courier New", Courier, mono;
|
||||||
|
font-weight: normal;
|
||||||
|
}
|
||||||
|
|
||||||
|
dt:target, .highlighted {
|
||||||
|
background-color: #ccc;
|
||||||
|
}
|
||||||
|
|
||||||
|
p {
|
||||||
|
padding-bottom: 0px;
|
||||||
|
}
|
||||||
|
|
||||||
|
p.last {
|
||||||
|
margin-bottom: 0px;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl {
|
||||||
|
padding: 1em 1em 1em 1em;
|
||||||
|
background: #fffff0;
|
||||||
|
border: 1px solid #aaa;
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
dl {
|
||||||
|
margin-bottom: 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
table.docutils {
|
||||||
|
background: #fffff0;
|
||||||
|
border-collapse: collapse;
|
||||||
|
border: 1px solid #ddd;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl table.docutils {
|
||||||
|
border: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
table.docutils dl {
|
||||||
|
border: 1px dashed #666;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
0
doc/_static/broxygen-extra.js
vendored
Normal file
0
doc/_static/broxygen-extra.js
vendored
Normal file
437
doc/_static/broxygen.css
vendored
Normal file
437
doc/_static/broxygen.css
vendored
Normal file
|
@ -0,0 +1,437 @@
|
||||||
|
/* Automatically generated. Do not edit. */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#bro-main, #bro-standalone-main {
|
||||||
|
padding: 0 0 0 0;
|
||||||
|
position:relative;
|
||||||
|
z-index:1;
|
||||||
|
}
|
||||||
|
|
||||||
|
#bro-main {
|
||||||
|
margin-bottom: 2em;
|
||||||
|
}
|
||||||
|
|
||||||
|
#bro-standalone-main {
|
||||||
|
margin-bottom: 0em;
|
||||||
|
padding-left: 50px;
|
||||||
|
padding-right: 50px;
|
||||||
|
}
|
||||||
|
|
||||||
|
#bro-outer {
|
||||||
|
color: #333;
|
||||||
|
background: #ffffff;
|
||||||
|
}
|
||||||
|
|
||||||
|
#bro-title {
|
||||||
|
font-weight: bold;
|
||||||
|
font-size: 32px;
|
||||||
|
line-height:32px;
|
||||||
|
text-align: center;
|
||||||
|
padding-top: 3px;
|
||||||
|
margin-bottom: 30px;
|
||||||
|
font-family: Palatino,'Palatino Linotype',Georgia,serif;;
|
||||||
|
color: #000;
|
||||||
|
}
|
||||||
|
|
||||||
|
.opening:first-letter {
|
||||||
|
font-size: 24px;
|
||||||
|
font-weight: bold;
|
||||||
|
letter-spacing: 0.05em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.opening {
|
||||||
|
font-size: 17px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.version {
|
||||||
|
text-align: right;
|
||||||
|
font-size: 12px;
|
||||||
|
color: #aaa;
|
||||||
|
line-height: 0;
|
||||||
|
height: 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
.git-info-version {
|
||||||
|
position: relative;
|
||||||
|
height: 2em;
|
||||||
|
top: -1em;
|
||||||
|
color: #ccc;
|
||||||
|
float: left;
|
||||||
|
font-size: 12px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.git-info-date {
|
||||||
|
position: relative;
|
||||||
|
height: 2em;
|
||||||
|
top: -1em;
|
||||||
|
color: #ccc;
|
||||||
|
float: right;
|
||||||
|
font-size: 12px;
|
||||||
|
}
|
||||||
|
|
||||||
|
body {
|
||||||
|
font-family:Arial, Helvetica, sans-serif;
|
||||||
|
font-size:15px;
|
||||||
|
line-height:22px;
|
||||||
|
color: #333;
|
||||||
|
margin: 0px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h1, h2, h3, h4, h5, h6,
|
||||||
|
h1 a, h2 a, h3 a, h4 a, h5 a, h6 a {
|
||||||
|
padding:0 0 20px 0;
|
||||||
|
font-weight:bold;
|
||||||
|
text-decoration:none;
|
||||||
|
}
|
||||||
|
|
||||||
|
div.section h3, div.section h4, div.section h5, div.section h6 {
|
||||||
|
font-style: italic;
|
||||||
|
}
|
||||||
|
|
||||||
|
h1, h2 {
|
||||||
|
font-size:27px;
|
||||||
|
letter-spacing:-1px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h3 {
|
||||||
|
margin-top: 1em;
|
||||||
|
font-size:18px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h4 {
|
||||||
|
font-size:16px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h5 {
|
||||||
|
font-size:15px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h6 {
|
||||||
|
font-size:12px;
|
||||||
|
}
|
||||||
|
|
||||||
|
p {
|
||||||
|
padding:0 0 20px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
hr {
|
||||||
|
background:none;
|
||||||
|
height:1px;
|
||||||
|
line-height:1px;
|
||||||
|
border:0;
|
||||||
|
margin:0 0 20px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul, ol {
|
||||||
|
margin:0 20px 20px 0;
|
||||||
|
padding-left:40px;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul.simple, ol.simple {
|
||||||
|
margin:0 0px 0px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
blockquote {
|
||||||
|
margin:0 0 0 40px;
|
||||||
|
}
|
||||||
|
|
||||||
|
strong, dfn {
|
||||||
|
font-weight:bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
em, dfn {
|
||||||
|
font-style:italic;
|
||||||
|
}
|
||||||
|
|
||||||
|
sup, sub {
|
||||||
|
line-height:0;
|
||||||
|
}
|
||||||
|
|
||||||
|
pre {
|
||||||
|
white-space:pre;
|
||||||
|
}
|
||||||
|
|
||||||
|
pre, code, tt {
|
||||||
|
font-family:"Courier New", Courier, mono;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl {
|
||||||
|
margin: 0 0 20px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
dl dt {
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
dd {
|
||||||
|
margin:0 0 20px 20px;
|
||||||
|
}
|
||||||
|
|
||||||
|
small {
|
||||||
|
font-size:75%;
|
||||||
|
}
|
||||||
|
|
||||||
|
a:link,
|
||||||
|
a:visited,
|
||||||
|
a:active
|
||||||
|
{
|
||||||
|
color: #2a85a7;
|
||||||
|
}
|
||||||
|
|
||||||
|
a:hover
|
||||||
|
{
|
||||||
|
color:#c24444;
|
||||||
|
}
|
||||||
|
|
||||||
|
h1, h2, h3, h4, h5, h6,
|
||||||
|
h1 a, h2 a, h3 a, h4 a, h5 a, h6 a
|
||||||
|
{
|
||||||
|
color: #333;
|
||||||
|
}
|
||||||
|
|
||||||
|
hr {
|
||||||
|
border-bottom:1px solid #ddd;
|
||||||
|
}
|
||||||
|
|
||||||
|
pre {
|
||||||
|
color: #333;
|
||||||
|
background: #FFFAE2;
|
||||||
|
padding: 7px 5px 3px 5px;
|
||||||
|
margin-bottom: 25px;
|
||||||
|
margin-top: 0px;
|
||||||
|
}
|
||||||
|
|
||||||
|
ul {
|
||||||
|
padding-bottom: 5px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h1, h2 {
|
||||||
|
margin-top: 30px;
|
||||||
|
}
|
||||||
|
|
||||||
|
h1 {
|
||||||
|
margin-bottom: 50px;
|
||||||
|
margin-bottom: 20px;
|
||||||
|
padding-bottom: 5px;
|
||||||
|
border-bottom: 1px solid;
|
||||||
|
border-color: #aaa;
|
||||||
|
}
|
||||||
|
|
||||||
|
h2 {
|
||||||
|
font-size: 24px;
|
||||||
|
}
|
||||||
|
|
||||||
|
pre {
|
||||||
|
-moz-box-shadow:0 0 6px #ddd;
|
||||||
|
-webkit-box-shadow:0 0 6px #ddd;
|
||||||
|
box-shadow:0 0 6px #ddd;
|
||||||
|
}
|
||||||
|
|
||||||
|
a {
|
||||||
|
text-decoration:none;
|
||||||
|
}
|
||||||
|
|
||||||
|
p {
|
||||||
|
padding-bottom: 15px;
|
||||||
|
}
|
||||||
|
|
||||||
|
p, dd, li {
|
||||||
|
text-align: justify;
|
||||||
|
}
|
||||||
|
|
||||||
|
li {
|
||||||
|
margin-bottom: 5px;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#footer .widget_links ul a,
|
||||||
|
#footer .widget_links ol a
|
||||||
|
{
|
||||||
|
color: #ddd;
|
||||||
|
}
|
||||||
|
|
||||||
|
#footer .widget_links ul a:hover,
|
||||||
|
#footer .widget_links ol a:hover
|
||||||
|
{
|
||||||
|
color:#c24444;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#footer .widget li {
|
||||||
|
padding-bottom:10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
#footer .widget_links li {
|
||||||
|
padding-bottom:1px;
|
||||||
|
}
|
||||||
|
|
||||||
|
#footer .widget li:last-child {
|
||||||
|
padding-bottom:0;
|
||||||
|
}
|
||||||
|
|
||||||
|
#footer .widgettitle {
|
||||||
|
color: #ddd;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
.widget {
|
||||||
|
margin:0 0 40px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widget, .widgettitle {
|
||||||
|
font-size:12px;
|
||||||
|
line-height:18px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widgettitle {
|
||||||
|
font-weight:bold;
|
||||||
|
text-transform:uppercase;
|
||||||
|
padding:0 0 10px 0;
|
||||||
|
margin:0 0 20px 0;
|
||||||
|
line-height:100%;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widget UL, .widget OL {
|
||||||
|
list-style-type:none;
|
||||||
|
margin:0;
|
||||||
|
padding:0;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widget p {
|
||||||
|
padding:0;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widget li {
|
||||||
|
padding-bottom:10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widget a {
|
||||||
|
text-decoration:none;
|
||||||
|
}
|
||||||
|
|
||||||
|
#bro-main .widgettitle,
|
||||||
|
{
|
||||||
|
color: #333;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
.widget img.left {
|
||||||
|
padding:5px 10px 10px 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widget img.right {
|
||||||
|
padding:5px 0 10px 10px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.ads .widgettitle {
|
||||||
|
margin-right:16px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widget {
|
||||||
|
margin-left: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widgettitle {
|
||||||
|
color: #333;
|
||||||
|
}
|
||||||
|
|
||||||
|
.widgettitle {
|
||||||
|
border-bottom:1px solid #ddd;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
.sidebar-toc ul li {
|
||||||
|
padding-bottom: 0px;
|
||||||
|
text-align: left;
|
||||||
|
list-style-type: square;
|
||||||
|
list-style-position: inside;
|
||||||
|
padding-left: 1em;
|
||||||
|
text-indent: -1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.sidebar-toc ul li li {
|
||||||
|
margin-left: 1em;
|
||||||
|
margin-bottom: 0px;
|
||||||
|
list-style-type: square;
|
||||||
|
}
|
||||||
|
|
||||||
|
.sidebar-toc ul li li a {
|
||||||
|
font-size: 8pt;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents {
|
||||||
|
padding: 10px;
|
||||||
|
background: #FFFAE2;
|
||||||
|
margin: 20px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.topic-title {
|
||||||
|
font-size: 20px;
|
||||||
|
font-weight: bold;
|
||||||
|
padding: 0px 0px 5px 0px;
|
||||||
|
text-align: center;
|
||||||
|
padding-top: .5em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents li {
|
||||||
|
margin-bottom: 0px;
|
||||||
|
list-style-type: square;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents ul ul li {
|
||||||
|
margin-left: 0px;
|
||||||
|
padding-left: 0px;
|
||||||
|
padding-top: 0em;
|
||||||
|
font-size: 90%;
|
||||||
|
list-style-type: square;
|
||||||
|
font-weight: normal;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents ul ul ul li {
|
||||||
|
list-style-type: none;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents ul ul ul ul li {
|
||||||
|
display:none;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents ul li {
|
||||||
|
padding-top: 1em;
|
||||||
|
list-style-type: none;
|
||||||
|
font-weight: bold;
|
||||||
|
}
|
||||||
|
|
||||||
|
.contents ul {
|
||||||
|
margin-left: 0px;
|
||||||
|
padding-left: 2em;
|
||||||
|
margin: 0px 0px 0px 0px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.note, .warning, .error {
|
||||||
|
margin-left: 2em;
|
||||||
|
margin-right: 2em;
|
||||||
|
margin-top: 1.5em;
|
||||||
|
margin-bottom: 1.5em;
|
||||||
|
padding: 0.5em 1em 0.5em 1em;
|
||||||
|
overflow: auto;
|
||||||
|
border-left: solid 3px #aaa;
|
||||||
|
font-size: 15px;
|
||||||
|
color: #333;
|
||||||
|
}
|
||||||
|
|
||||||
|
.admonition p {
|
||||||
|
margin-left: 1em;
|
||||||
|
}
|
||||||
|
|
||||||
|
.admonition-title {
|
||||||
|
font-size: 16px;
|
||||||
|
font-weight: bold;
|
||||||
|
color: #000;
|
||||||
|
padding-bottom: 0em;
|
||||||
|
margin-bottom: .5em;
|
||||||
|
margin-top: 0em;
|
||||||
|
}
|
309
doc/_static/default.css_t
vendored
309
doc/_static/default.css_t
vendored
|
@ -1,309 +0,0 @@
|
||||||
/*
|
|
||||||
* default.css_t
|
|
||||||
* ~~~~~~~~~~~~~
|
|
||||||
*
|
|
||||||
* Sphinx stylesheet -- default theme.
|
|
||||||
*
|
|
||||||
* :copyright: Copyright 2007-2011 by the Sphinx team, see AUTHORS.
|
|
||||||
* :license: BSD, see LICENSE for details.
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
|
|
||||||
@import url("basic.css");
|
|
||||||
|
|
||||||
/* -- page layout ----------------------------------------------------------- */
|
|
||||||
|
|
||||||
body {
|
|
||||||
font-family: {{ theme_bodyfont }};
|
|
||||||
font-size: 100%;
|
|
||||||
background-color: {{ theme_footerbgcolor }};
|
|
||||||
color: #000;
|
|
||||||
margin: 0;
|
|
||||||
padding: 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.document {
|
|
||||||
background-color: {{ theme_sidebarbgcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.documentwrapper {
|
|
||||||
float: left;
|
|
||||||
width: 100%;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.bodywrapper {
|
|
||||||
margin: 0 0 0 {{ theme_sidebarwidth|toint }}px;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.body {
|
|
||||||
background-color: {{ theme_bgcolor }};
|
|
||||||
color: {{ theme_textcolor }};
|
|
||||||
padding: 0 20px 30px 20px;
|
|
||||||
}
|
|
||||||
|
|
||||||
{%- if theme_rightsidebar|tobool %}
|
|
||||||
div.bodywrapper {
|
|
||||||
margin: 0 {{ theme_sidebarwidth|toint }}px 0 0;
|
|
||||||
}
|
|
||||||
{%- endif %}
|
|
||||||
|
|
||||||
div.footer {
|
|
||||||
color: {{ theme_footertextcolor }};
|
|
||||||
background-color: {{ theme_footerbgcolor }};
|
|
||||||
width: 100%;
|
|
||||||
padding: 9px 0 9px 0;
|
|
||||||
text-align: center;
|
|
||||||
font-size: 75%;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.footer a {
|
|
||||||
color: {{ theme_footertextcolor }};
|
|
||||||
text-decoration: underline;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.related {
|
|
||||||
background-color: {{ theme_relbarbgcolor }};
|
|
||||||
line-height: 30px;
|
|
||||||
color: {{ theme_relbartextcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.related a {
|
|
||||||
color: {{ theme_relbarlinkcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar {
|
|
||||||
{%- if theme_stickysidebar|tobool %}
|
|
||||||
top: 30px;
|
|
||||||
bottom: 0;
|
|
||||||
margin: 0;
|
|
||||||
position: fixed;
|
|
||||||
overflow: auto;
|
|
||||||
height: auto;
|
|
||||||
{%- endif %}
|
|
||||||
{%- if theme_rightsidebar|tobool %}
|
|
||||||
float: right;
|
|
||||||
{%- if theme_stickysidebar|tobool %}
|
|
||||||
right: 0;
|
|
||||||
{%- endif %}
|
|
||||||
{%- endif %}
|
|
||||||
}
|
|
||||||
|
|
||||||
{%- if theme_stickysidebar|tobool %}
|
|
||||||
/* this is nice, but it it leads to hidden headings when jumping
|
|
||||||
to an anchor */
|
|
||||||
/*
|
|
||||||
div.related {
|
|
||||||
position: fixed;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.documentwrapper {
|
|
||||||
margin-top: 30px;
|
|
||||||
}
|
|
||||||
*/
|
|
||||||
{%- endif %}
|
|
||||||
|
|
||||||
div.sphinxsidebar h3 {
|
|
||||||
font-family: {{ theme_bodyfont }};
|
|
||||||
color: {{ theme_sidebartextcolor }};
|
|
||||||
font-size: 1.4em;
|
|
||||||
font-weight: normal;
|
|
||||||
margin: 0;
|
|
||||||
padding: 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar h3 a {
|
|
||||||
color: {{ theme_sidebartextcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar h4 {
|
|
||||||
font-family: {{ theme_bodyfont }};
|
|
||||||
color: {{ theme_sidebartextcolor }};
|
|
||||||
font-size: 1.3em;
|
|
||||||
font-weight: normal;
|
|
||||||
margin: 5px 0 0 0;
|
|
||||||
padding: 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar p {
|
|
||||||
color: {{ theme_sidebartextcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar p.topless {
|
|
||||||
margin: 5px 10px 10px 10px;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar ul {
|
|
||||||
margin: 10px;
|
|
||||||
padding: 0;
|
|
||||||
color: {{ theme_sidebartextcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar a {
|
|
||||||
color: {{ theme_sidebarlinkcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.sphinxsidebar input {
|
|
||||||
border: 1px solid {{ theme_sidebarlinkcolor }};
|
|
||||||
font-family: sans-serif;
|
|
||||||
font-size: 1em;
|
|
||||||
}
|
|
||||||
|
|
||||||
{% if theme_collapsiblesidebar|tobool %}
|
|
||||||
/* for collapsible sidebar */
|
|
||||||
div#sidebarbutton {
|
|
||||||
background-color: {{ theme_sidebarbtncolor }};
|
|
||||||
}
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
/* -- hyperlink styles ------------------------------------------------------ */
|
|
||||||
|
|
||||||
a {
|
|
||||||
color: {{ theme_linkcolor }};
|
|
||||||
text-decoration: none;
|
|
||||||
}
|
|
||||||
|
|
||||||
a:visited {
|
|
||||||
color: {{ theme_visitedlinkcolor }};
|
|
||||||
text-decoration: none;
|
|
||||||
}
|
|
||||||
|
|
||||||
{% if theme_externalrefs|tobool %}
|
|
||||||
a.external {
|
|
||||||
text-decoration: none;
|
|
||||||
border-bottom: 1px dashed {{ theme_linkcolor }};
|
|
||||||
}
|
|
||||||
|
|
||||||
a.external:hover {
|
|
||||||
text-decoration: none;
|
|
||||||
border-bottom: none;
|
|
||||||
}
|
|
||||||
|
|
||||||
a.external:visited {
|
|
||||||
text-decoration: none;
|
|
||||||
border-bottom: 1px dashed {{ theme_visitedlinkcolor }};
|
|
||||||
}
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
/* -- body styles ----------------------------------------------------------- */
|
|
||||||
|
|
||||||
div.body h1,
|
|
||||||
div.body h2,
|
|
||||||
div.body h3,
|
|
||||||
div.body h4,
|
|
||||||
div.body h5,
|
|
||||||
div.body h6 {
|
|
||||||
font-family: {{ theme_bodyfont }};
|
|
||||||
background-color: #ffffff;
|
|
||||||
font-weight: normal;
|
|
||||||
color: {{ theme_headtextcolor }};
|
|
||||||
border-bottom: 1px solid #aaa;
|
|
||||||
margin: 20px -20px 10px -20px;
|
|
||||||
padding: 3px 0 3px 10px;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.body h1 {
|
|
||||||
font-family: {{ theme_headfont }};
|
|
||||||
text-align: center;
|
|
||||||
border-bottom: none;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.body h1 { margin-top: 0; font-size: 200%; }
|
|
||||||
div.body h2 { font-size: 160%; }
|
|
||||||
div.body h3 { font-size: 140%; }
|
|
||||||
div.body h4 { font-size: 120%; }
|
|
||||||
div.body h5 { font-size: 110%; }
|
|
||||||
div.body h6 { font-size: 100%; }
|
|
||||||
|
|
||||||
a.headerlink {
|
|
||||||
color: {{ theme_headlinkcolor }};
|
|
||||||
font-size: 0.8em;
|
|
||||||
padding: 0 4px 0 4px;
|
|
||||||
text-decoration: none;
|
|
||||||
}
|
|
||||||
|
|
||||||
a.headerlink:hover {
|
|
||||||
background-color: {{ theme_headlinkcolor }};
|
|
||||||
color: white;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.admonition p.admonition-title + p {
|
|
||||||
display: inline;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.admonition p {
|
|
||||||
margin-bottom: 5px;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.admonition pre {
|
|
||||||
margin-bottom: 5px;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.admonition ul, div.admonition ol {
|
|
||||||
margin-bottom: 5px;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.note {
|
|
||||||
background-color: #eee;
|
|
||||||
border: 1px solid #ccc;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.seealso {
|
|
||||||
background-color: #ffc;
|
|
||||||
border: 1px solid #ff6;
|
|
||||||
}
|
|
||||||
|
|
||||||
div.warning {
|
|
||||||
background-color: #ffe4e4;
|
|
||||||
border: 1px solid #f66;
|
|
||||||
}
|
|
||||||
|
|
||||||
p.admonition-title {
|
|
||||||
display: inline;
|
|
||||||
}
|
|
||||||
|
|
||||||
p.admonition-title:after {
|
|
||||||
content: ":";
|
|
||||||
}
|
|
||||||
|
|
||||||
pre {
|
|
||||||
padding: 5px;
|
|
||||||
background-color: {{ theme_codebgcolor }};
|
|
||||||
color: {{ theme_codetextcolor }};
|
|
||||||
line-height: 120%;
|
|
||||||
border: 1px solid #ac9;
|
|
||||||
border-left: none;
|
|
||||||
border-right: none;
|
|
||||||
}
|
|
||||||
|
|
||||||
tt {
|
|
||||||
background-color: #ecf0f3;
|
|
||||||
padding: 0 1px 0 1px;
|
|
||||||
font-size: 0.95em;
|
|
||||||
}
|
|
||||||
|
|
||||||
th {
|
|
||||||
background-color: #ede;
|
|
||||||
}
|
|
||||||
|
|
||||||
.warning tt {
|
|
||||||
background: #efc2c2;
|
|
||||||
}
|
|
||||||
|
|
||||||
.note tt {
|
|
||||||
background: #d6d6d6;
|
|
||||||
}
|
|
||||||
|
|
||||||
.viewcode-back {
|
|
||||||
font-family: {{ theme_bodyfont }};
|
|
||||||
}
|
|
||||||
|
|
||||||
div.viewcode-block:target {
|
|
||||||
background-color: #f4debf;
|
|
||||||
border-top: 1px solid #ac9;
|
|
||||||
border-bottom: 1px solid #ac9;
|
|
||||||
}
|
|
||||||
|
|
||||||
th.field-name
|
|
||||||
{
|
|
||||||
white-space:nowrap;
|
|
||||||
}
|
|
3
doc/_static/download.js
vendored
3
doc/_static/download.js
vendored
|
@ -1,3 +0,0 @@
|
||||||
$(document).ready(function() {
|
|
||||||
$('.docutils.download').removeClass('download');
|
|
||||||
});
|
|
58
doc/_static/pygments.css
vendored
Normal file
58
doc/_static/pygments.css
vendored
Normal file
|
@ -0,0 +1,58 @@
|
||||||
|
.hll { background-color: #ffffcc }
|
||||||
|
.c { color: #aaaaaa; font-style: italic } /* Comment */
|
||||||
|
.err { color: #F00000; background-color: #F0A0A0 } /* Error */
|
||||||
|
.k { color: #0000aa } /* Keyword */
|
||||||
|
.cm { color: #aaaaaa; font-style: italic } /* Comment.Multiline */
|
||||||
|
.cp { color: #4c8317 } /* Comment.Preproc */
|
||||||
|
.c1 { color: #aaaaaa; font-style: italic } /* Comment.Single */
|
||||||
|
.cs { color: #0000aa; font-style: italic } /* Comment.Special */
|
||||||
|
.gd { color: #aa0000 } /* Generic.Deleted */
|
||||||
|
.ge { font-style: italic } /* Generic.Emph */
|
||||||
|
.gr { color: #aa0000 } /* Generic.Error */
|
||||||
|
.gh { color: #000080; font-weight: bold } /* Generic.Heading */
|
||||||
|
.gi { color: #00aa00 } /* Generic.Inserted */
|
||||||
|
.go { color: #888888 } /* Generic.Output */
|
||||||
|
.gp { color: #555555 } /* Generic.Prompt */
|
||||||
|
.gs { font-weight: bold } /* Generic.Strong */
|
||||||
|
.gu { color: #800080; font-weight: bold } /* Generic.Subheading */
|
||||||
|
.gt { color: #aa0000 } /* Generic.Traceback */
|
||||||
|
.kc { color: #0000aa } /* Keyword.Constant */
|
||||||
|
.kd { color: #0000aa } /* Keyword.Declaration */
|
||||||
|
.kn { color: #0000aa } /* Keyword.Namespace */
|
||||||
|
.kp { color: #0000aa } /* Keyword.Pseudo */
|
||||||
|
.kr { color: #0000aa } /* Keyword.Reserved */
|
||||||
|
.kt { color: #00aaaa } /* Keyword.Type */
|
||||||
|
.m { color: #009999 } /* Literal.Number */
|
||||||
|
.s { color: #aa5500 } /* Literal.String */
|
||||||
|
.na { color: #1e90ff } /* Name.Attribute */
|
||||||
|
.nb { color: #00aaaa } /* Name.Builtin */
|
||||||
|
.nc { color: #00aa00; text-decoration: underline } /* Name.Class */
|
||||||
|
.no { color: #aa0000 } /* Name.Constant */
|
||||||
|
.nd { color: #888888 } /* Name.Decorator */
|
||||||
|
.ni { color: #800000; font-weight: bold } /* Name.Entity */
|
||||||
|
.nf { color: #00aa00 } /* Name.Function */
|
||||||
|
.nn { color: #00aaaa; text-decoration: underline } /* Name.Namespace */
|
||||||
|
.nt { color: #1e90ff; font-weight: bold } /* Name.Tag */
|
||||||
|
.nv { color: #aa0000 } /* Name.Variable */
|
||||||
|
.ow { color: #0000aa } /* Operator.Word */
|
||||||
|
.w { color: #bbbbbb } /* Text.Whitespace */
|
||||||
|
.mf { color: #009999 } /* Literal.Number.Float */
|
||||||
|
.mh { color: #009999 } /* Literal.Number.Hex */
|
||||||
|
.mi { color: #009999 } /* Literal.Number.Integer */
|
||||||
|
.mo { color: #009999 } /* Literal.Number.Oct */
|
||||||
|
.sb { color: #aa5500 } /* Literal.String.Backtick */
|
||||||
|
.sc { color: #aa5500 } /* Literal.String.Char */
|
||||||
|
.sd { color: #aa5500 } /* Literal.String.Doc */
|
||||||
|
.s2 { color: #aa5500 } /* Literal.String.Double */
|
||||||
|
.se { color: #aa5500 } /* Literal.String.Escape */
|
||||||
|
.sh { color: #aa5500 } /* Literal.String.Heredoc */
|
||||||
|
.si { color: #aa5500 } /* Literal.String.Interpol */
|
||||||
|
.sx { color: #aa5500 } /* Literal.String.Other */
|
||||||
|
.sr { color: #009999 } /* Literal.String.Regex */
|
||||||
|
.s1 { color: #aa5500 } /* Literal.String.Single */
|
||||||
|
.ss { color: #0000aa } /* Literal.String.Symbol */
|
||||||
|
.bp { color: #00aaaa } /* Name.Builtin.Pseudo */
|
||||||
|
.vc { color: #aa0000 } /* Name.Variable.Class */
|
||||||
|
.vg { color: #aa0000 } /* Name.Variable.Global */
|
||||||
|
.vi { color: #aa0000 } /* Name.Variable.Instance */
|
||||||
|
.il { color: #009999 } /* Literal.Number.Integer.Long */
|
64
doc/_static/showhide.js
vendored
64
doc/_static/showhide.js
vendored
|
@ -1,64 +0,0 @@
|
||||||
// make literal blocks corresponding to identifier initial values
|
|
||||||
// hidden by default
|
|
||||||
$(document).ready(function() {
|
|
||||||
|
|
||||||
var showText='(Show Value)';
|
|
||||||
var hideText='(Hide Value)';
|
|
||||||
|
|
||||||
var is_visible = false;
|
|
||||||
|
|
||||||
// select field-list tables that come before a literal block
|
|
||||||
tables = $('.highlight-python').prev('table.docutils.field-list');
|
|
||||||
|
|
||||||
tables.find('th.field-name').filter(function(index) {
|
|
||||||
return $(this).html() == "Default :";
|
|
||||||
}).next().append('<a href="#" class="toggleLink">'+showText+'</a>');
|
|
||||||
|
|
||||||
// hide all literal blocks that follow a field-list table
|
|
||||||
tables.next('.highlight-python').hide();
|
|
||||||
|
|
||||||
// register handler for clicking a "toggle" link
|
|
||||||
$('a.toggleLink').click(function() {
|
|
||||||
is_visible = !is_visible;
|
|
||||||
|
|
||||||
$(this).html( (!is_visible) ? showText : hideText);
|
|
||||||
|
|
||||||
// the link is inside a <table><tbody><tr><td> and the next
|
|
||||||
// literal block after the table is the literal block that we want
|
|
||||||
// to show/hide
|
|
||||||
$(this).parent().parent().parent().parent().next('.highlight-python').slideToggle('fast');
|
|
||||||
|
|
||||||
// override default link behavior
|
|
||||||
return false;
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
// make "Private Interface" sections hidden by default
|
|
||||||
$(document).ready(function() {
|
|
||||||
|
|
||||||
var showText='Show Private Interface (for internal use)';
|
|
||||||
var hideText='Hide Private Interface';
|
|
||||||
|
|
||||||
var is_visible = false;
|
|
||||||
|
|
||||||
// insert show/hide links
|
|
||||||
$('#private-interface').children(":first-child").after('<a href="#" class="privateToggle">'+showText+'</a>');
|
|
||||||
|
|
||||||
// wrap all sub-sections in a new div that can be hidden/shown
|
|
||||||
$('#private-interface').children(".section").wrapAll('<div class="private" />');
|
|
||||||
|
|
||||||
// hide the given class
|
|
||||||
$('.private').hide();
|
|
||||||
|
|
||||||
// register handler for clicking a "toggle" link
|
|
||||||
$('a.privateToggle').click(function() {
|
|
||||||
is_visible = !is_visible;
|
|
||||||
|
|
||||||
$(this).html( (!is_visible) ? showText : hideText);
|
|
||||||
|
|
||||||
$('.private').slideToggle('fast');
|
|
||||||
|
|
||||||
// override default link behavior
|
|
||||||
return false;
|
|
||||||
});
|
|
||||||
});
|
|
23
doc/_templates/layout.html
vendored
23
doc/_templates/layout.html
vendored
|
@ -1,11 +1,12 @@
|
||||||
{% extends "!layout.html" %}
|
{% extends "!layout.html" %}
|
||||||
|
|
||||||
{% block extrahead %}
|
{% block extrahead %}
|
||||||
<link rel="stylesheet" type="text/css" href="http://www.bro-ids.org/css/bro-ids.css" />
|
<link rel="stylesheet" type="text/css" href="{{ pathto('_static/broxygen.css', 1) }}"></script>
|
||||||
<link rel="stylesheet" type="text/css" href="http://www.bro-ids.org/css/960.css" />
|
<link rel="stylesheet" type="text/css" href="{{ pathto('_static/960.css', 1) }}"></script>
|
||||||
<link rel="stylesheet" type="text/css" href="http://www.bro-ids.org/css/pygments.css" />
|
<link rel="stylesheet" type="text/css" href="{{ pathto('_static/pygments.css', 1) }}"></script>
|
||||||
<link rel="stylesheet" type="text/css" href="{{ pathto('_static/broxygen-extra.css', 1) }}"></script>
|
<link rel="stylesheet" type="text/css" href="{{ pathto('_static/broxygen-extra.css', 1) }}"></script>
|
||||||
<script type="text/javascript" src="{{ pathto('_static/download.js', 1) }}"></script>
|
|
||||||
|
<script type="text/javascript" src="{{ pathto('_static/broxygen-extra.js', 1) }}"></script>
|
||||||
{% endblock %}
|
{% endblock %}
|
||||||
|
|
||||||
{% block header %}
|
{% block header %}
|
||||||
|
@ -47,6 +48,7 @@
|
||||||
Table of Contents
|
Table of Contents
|
||||||
</h3>
|
</h3>
|
||||||
<p>
|
<p>
|
||||||
|
<!-- <ul id="sidebar-toc"></ul> -->
|
||||||
<ul>{{toc}}</ul>
|
<ul>{{toc}}</ul>
|
||||||
</p>
|
</p>
|
||||||
</div>
|
</div>
|
||||||
|
@ -73,6 +75,19 @@
|
||||||
</div>
|
</div>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
|
|
||||||
|
{%- if pagename != "search" %}
|
||||||
|
<div id="searchbox" style="display: none" class="widget">
|
||||||
|
<h3 class="widgettitle">{{ _('Search') }}</h3>
|
||||||
|
<form class="search" action="{{ pathto('search') }}" method="get">
|
||||||
|
<input type="text" name="q" />
|
||||||
|
<input type="submit" value="{{ _('Search') }}" />
|
||||||
|
<input type="hidden" name="check_keywords" value="yes" />
|
||||||
|
<input type="hidden" name="area" value="default" />
|
||||||
|
</form>
|
||||||
|
</div>
|
||||||
|
<script type="text/javascript">$('#searchbox').show(0);</script>
|
||||||
|
{%- endif %}
|
||||||
|
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
|
|
@ -24,7 +24,7 @@ sys.path.insert(0, os.path.abspath('sphinx-sources/ext'))
|
||||||
|
|
||||||
# Add any Sphinx extension module names here, as strings. They can be extensions
|
# Add any Sphinx extension module names here, as strings. They can be extensions
|
||||||
# coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
|
# coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
|
||||||
extensions = ['bro', 'rst_directive', 'sphinx.ext.todo']
|
extensions = ['bro', 'rst_directive', 'sphinx.ext.todo', 'adapt-toc']
|
||||||
|
|
||||||
# Add any paths that contain templates here, relative to this directory.
|
# Add any paths that contain templates here, relative to this directory.
|
||||||
templates_path = ['sphinx-sources/_templates', 'sphinx-sources/_static']
|
templates_path = ['sphinx-sources/_templates', 'sphinx-sources/_static']
|
||||||
|
@ -40,7 +40,7 @@ master_doc = 'index'
|
||||||
|
|
||||||
# General information about the project.
|
# General information about the project.
|
||||||
project = u'Bro'
|
project = u'Bro'
|
||||||
copyright = u'2011, The Bro Project'
|
copyright = u'2012, The Bro Project'
|
||||||
|
|
||||||
# The version info for the project you're documenting, acts as replacement for
|
# The version info for the project you're documenting, acts as replacement for
|
||||||
# |version| and |release|, also used in various other places throughout the
|
# |version| and |release|, also used in various other places throughout the
|
||||||
|
@ -169,6 +169,7 @@ html_sidebars = {
|
||||||
# Output file base name for HTML help builder.
|
# Output file base name for HTML help builder.
|
||||||
htmlhelp_basename = 'Broxygen'
|
htmlhelp_basename = 'Broxygen'
|
||||||
|
|
||||||
|
html_add_permalinks = None
|
||||||
|
|
||||||
# -- Options for LaTeX output --------------------------------------------------
|
# -- Options for LaTeX output --------------------------------------------------
|
||||||
|
|
||||||
|
@ -208,7 +209,6 @@ latex_documents = [
|
||||||
# If false, no module index is generated.
|
# If false, no module index is generated.
|
||||||
#latex_domain_indices = True
|
#latex_domain_indices = True
|
||||||
|
|
||||||
|
|
||||||
# -- Options for manual page output --------------------------------------------
|
# -- Options for manual page output --------------------------------------------
|
||||||
|
|
||||||
# One entry per manual page. List of tuples
|
# One entry per manual page. List of tuples
|
||||||
|
|
29
doc/ext/adapt-toc.py
Normal file
29
doc/ext/adapt-toc.py
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
|
||||||
|
import sys
|
||||||
|
import re
|
||||||
|
|
||||||
|
# Removes the first TOC level, which is just the page title.
|
||||||
|
def process_html_toc(app, pagename, templatename, context, doctree):
|
||||||
|
|
||||||
|
if not "toc" in context:
|
||||||
|
return
|
||||||
|
|
||||||
|
toc = context["toc"]
|
||||||
|
|
||||||
|
lines = toc.strip().split("\n")
|
||||||
|
lines = lines[2:-2]
|
||||||
|
|
||||||
|
toc = "\n".join(lines)
|
||||||
|
toc = "<ul>" + toc
|
||||||
|
|
||||||
|
context["toc"] = toc
|
||||||
|
|
||||||
|
# print >>sys.stderr, pagename
|
||||||
|
# print >>sys.stderr, context["toc"]
|
||||||
|
# print >>sys.stderr, "-----"
|
||||||
|
# print >>sys.stderr, toc
|
||||||
|
# print >>sys.stderr, "===="
|
||||||
|
|
||||||
|
def setup(app):
|
||||||
|
app.connect('html-page-context', process_html_toc)
|
||||||
|
|
|
@ -257,6 +257,9 @@ class BroDomain(Domain):
|
||||||
objects[objtype, target],
|
objects[objtype, target],
|
||||||
objtype + '-' + target,
|
objtype + '-' + target,
|
||||||
contnode, target + ' ' + objtype)
|
contnode, target + ' ' + objtype)
|
||||||
|
else:
|
||||||
|
self.env.warn(fromdocname,
|
||||||
|
'unknown target for ":bro:%s:`%s`"' % (typ, target))
|
||||||
|
|
||||||
def get_objects(self):
|
def get_objects(self):
|
||||||
for (typ, name), docname in self.data['objects'].iteritems():
|
for (typ, name), docname in self.data['objects'].iteritems():
|
||||||
|
|
64
doc/faq.rst
64
doc/faq.rst
|
@ -28,6 +28,23 @@ Here are some pointers to more information:
|
||||||
Lothar Braun et. al evaluates packet capture performance on
|
Lothar Braun et. al evaluates packet capture performance on
|
||||||
commodity hardware
|
commodity hardware
|
||||||
|
|
||||||
|
Are there any gotchas regarding interface configuration for live capture? Or why might I be seeing abnormally large packets much greater than interface MTU?
|
||||||
|
-------------------------------------------------------------------------------------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
Some NICs offload the reassembly of traffic into "superpackets" so that
|
||||||
|
fewer packets are then passed up the stack (e.g. "TCP segmentation
|
||||||
|
offload", or "generic segmentation offload"). The result is that the
|
||||||
|
capturing application will observe packets much larger than the MTU size
|
||||||
|
of the interface they were captured from and may also interfere with the
|
||||||
|
maximum packet capture length, ``snaplen``, so it's a good idea to disable
|
||||||
|
an interface's offloading features.
|
||||||
|
|
||||||
|
You can use the ``ethtool`` program on Linux to view and disable
|
||||||
|
offloading features of an interface. See this page for more explicit
|
||||||
|
directions:
|
||||||
|
|
||||||
|
http://securityonion.blogspot.com/2011/10/when-is-full-packet-capture-not-full.html
|
||||||
|
|
||||||
What does an error message like ``internal error: NB-DNS error`` mean?
|
What does an error message like ``internal error: NB-DNS error`` mean?
|
||||||
---------------------------------------------------------------------------------------------------------------------------------
|
---------------------------------------------------------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
@ -35,6 +52,19 @@ That often means that DNS is not set up correctly on the system
|
||||||
running Bro. Try verifying from the command line that DNS lookups
|
running Bro. Try verifying from the command line that DNS lookups
|
||||||
work, e.g., ``host www.google.com``.
|
work, e.g., ``host www.google.com``.
|
||||||
|
|
||||||
|
I am using OpenBSD and having problems installing Bro?
|
||||||
|
------------------------------------------------------
|
||||||
|
|
||||||
|
One potential issue is that the top-level Makefile may not work with
|
||||||
|
OpenBSD's default make program, in which case you can either install
|
||||||
|
the ``gmake`` package and use it instead or first change into the
|
||||||
|
``build/`` directory before doing either ``make`` or ``make install``
|
||||||
|
such that the CMake-generated Makefile's are used directly.
|
||||||
|
|
||||||
|
Generally, please note that we do not regularly test OpenBSD builds.
|
||||||
|
We appreciate any patches that improve Bro's support for this
|
||||||
|
platform.
|
||||||
|
|
||||||
|
|
||||||
Usage
|
Usage
|
||||||
=====
|
=====
|
||||||
|
@ -42,34 +72,30 @@ Usage
|
||||||
How can I identify backscatter?
|
How can I identify backscatter?
|
||||||
-------------------------------
|
-------------------------------
|
||||||
|
|
||||||
Identifying backscatter via connections labeled as ``OTH`` is not
|
Identifying backscatter via connections labeled as ``OTH`` is not a reliable
|
||||||
a reliable means to detect backscatter. Use rather the following
|
means to detect backscatter. Backscatter is however visible by interpreting
|
||||||
procedure:
|
the contents of the ``history`` field in the ``conn.log`` file. The basic idea
|
||||||
|
is to watch for connections that never had an initial ``SYN`` but started
|
||||||
* Enable connection history via ``redef record_state_history=T`` to
|
instead with a ``SYN-ACK`` or ``RST`` (though this latter generally is just
|
||||||
track all control/data packet types in connection logs.
|
discarded). Here are some history fields which provide backscatter examples:
|
||||||
|
``hAFf``, ``r``. Refer to the conn protocol analysis scripts to interpret the
|
||||||
* Backscatter is now visible in terms of connections that never had an
|
individual character meanings in the history field.
|
||||||
initial ``SYN`` but started instead with a ``SYN-ACK`` or ``RST``
|
|
||||||
(though this latter generally is just discarded).
|
|
||||||
|
|
||||||
Is there help for understanding Bro's resource consumption?
|
Is there help for understanding Bro's resource consumption?
|
||||||
-----------------------------------------------------------
|
-----------------------------------------------------------
|
||||||
|
|
||||||
There are two scripts that collect statistics on resource usage:
|
There are two scripts that collect statistics on resource usage:
|
||||||
``stats.bro`` and ``profiling.bro``. The former is quite lightweight,
|
``misc/stats.bro`` and ``misc/profiling.bro``. The former is quite
|
||||||
while the latter should only be used for debugging. Furthermore,
|
lightweight, while the latter should only be used for debugging.
|
||||||
there's also ``print-globals.bro``, which prints the size of all
|
|
||||||
global script variable at termination.
|
|
||||||
|
|
||||||
How can I capture packets as an unprivileged user?
|
How can I capture packets as an unprivileged user?
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
|
|
||||||
Normally, unprivileged users cannot capture packets from a network
|
Normally, unprivileged users cannot capture packets from a network interface,
|
||||||
interface, which means they would not be able to use Bro to read/analyze
|
which means they would not be able to use Bro to read/analyze live traffic.
|
||||||
live traffic. However, there are ways to enable packet capture
|
However, there are operating system specific ways to enable packet capture
|
||||||
permission for non-root users, which is worth doing in the context of
|
permission for non-root users, which is worth doing in the context of using
|
||||||
using Bro to monitor live traffic
|
Bro to monitor live traffic.
|
||||||
|
|
||||||
With Linux Capabilities
|
With Linux Capabilities
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
|
@ -41,10 +41,10 @@ Script Reference
|
||||||
.. toctree::
|
.. toctree::
|
||||||
:maxdepth: 1
|
:maxdepth: 1
|
||||||
|
|
||||||
scripts/builtins
|
|
||||||
scripts/bifs
|
|
||||||
scripts/packages
|
scripts/packages
|
||||||
scripts/index
|
scripts/index
|
||||||
|
scripts/builtins
|
||||||
|
scripts/bifs
|
||||||
|
|
||||||
Other Bro Components
|
Other Bro Components
|
||||||
--------------------
|
--------------------
|
||||||
|
|
|
@ -163,7 +163,7 @@ quickly ...
|
||||||
.. sidebar:: A More Generic Path Function
|
.. sidebar:: A More Generic Path Function
|
||||||
|
|
||||||
The ``split_log`` method has one draw-back: it can be used
|
The ``split_log`` method has one draw-back: it can be used
|
||||||
only with the :bro:enum:`Conn::Log` stream as the record type is hardcoded
|
only with the :bro:enum:`Conn::LOG` stream as the record type is hardcoded
|
||||||
into its argument list. However, Bro allows to do a more generic
|
into its argument list. However, Bro allows to do a more generic
|
||||||
variant:
|
variant:
|
||||||
|
|
||||||
|
|
|
@ -31,19 +31,19 @@ See the `bro downloads page`_ for currently supported/targeted platforms.
|
||||||
|
|
||||||
* RPM
|
* RPM
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo yum localinstall Bro-all*.rpm
|
sudo yum localinstall Bro-*.rpm
|
||||||
|
|
||||||
* DEB
|
* DEB
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo gdebi Bro-all-*.deb
|
sudo gdebi Bro-*.deb
|
||||||
|
|
||||||
* MacOS Disk Image with Installer
|
* MacOS Disk Image with Installer
|
||||||
|
|
||||||
Just open the ``Bro-all-*.dmg`` and then run the ``.pkg`` installer.
|
Just open the ``Bro-*.dmg`` and then run the ``.pkg`` installer.
|
||||||
Everything installed by the package will go into ``/opt/bro``.
|
Everything installed by the package will go into ``/opt/bro``.
|
||||||
|
|
||||||
The primary install prefix for binary packages is ``/opt/bro``.
|
The primary install prefix for binary packages is ``/opt/bro``.
|
||||||
|
@ -56,15 +56,17 @@ Building From Source
|
||||||
Required Dependencies
|
Required Dependencies
|
||||||
~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
The following dependencies are required to build Bro:
|
||||||
|
|
||||||
* RPM/RedHat-based Linux:
|
* RPM/RedHat-based Linux:
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo yum install cmake make gcc gcc-c++ flex bison libpcap-devel openssl-devel python-devel swig zlib-devel file-devel
|
sudo yum install cmake make gcc gcc-c++ flex bison libpcap-devel openssl-devel python-devel swig zlib-devel file-devel
|
||||||
|
|
||||||
* DEB/Debian-based Linux:
|
* DEB/Debian-based Linux:
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo apt-get install cmake make gcc g++ flex bison libpcap-dev libssl-dev python-dev swig zlib1g-dev libmagic-dev
|
sudo apt-get install cmake make gcc g++ flex bison libpcap-dev libssl-dev python-dev swig zlib1g-dev libmagic-dev
|
||||||
|
|
||||||
|
@ -73,9 +75,13 @@ Required Dependencies
|
||||||
Most required dependencies should come with a minimal FreeBSD install
|
Most required dependencies should come with a minimal FreeBSD install
|
||||||
except for the following.
|
except for the following.
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo pkg_add -r cmake swig bison python
|
sudo pkg_add -r bash cmake swig bison python
|
||||||
|
|
||||||
|
Note that ``bash`` needs to be in ``PATH``, which by default it is
|
||||||
|
not. The FreeBSD package installs the binary into
|
||||||
|
``/usr/local/bin``.
|
||||||
|
|
||||||
* Mac OS X
|
* Mac OS X
|
||||||
|
|
||||||
|
@ -99,19 +105,19 @@ sending emails.
|
||||||
|
|
||||||
* RPM/RedHat-based Linux:
|
* RPM/RedHat-based Linux:
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo yum install GeoIP-devel sendmail
|
sudo yum install GeoIP-devel sendmail
|
||||||
|
|
||||||
* DEB/Debian-based Linux:
|
* DEB/Debian-based Linux:
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo apt-get install libgeoip-dev sendmail
|
sudo apt-get install libgeoip-dev sendmail
|
||||||
|
|
||||||
* Ports-based FreeBSD
|
* Ports-based FreeBSD
|
||||||
|
|
||||||
.. console::
|
.. console::
|
||||||
|
|
||||||
sudo pkg_add -r GeoIP
|
sudo pkg_add -r GeoIP
|
||||||
|
|
||||||
|
|
|
@ -34,6 +34,7 @@ rest_target(${psd} base/frameworks/dpd/main.bro)
|
||||||
rest_target(${psd} base/frameworks/intel/main.bro)
|
rest_target(${psd} base/frameworks/intel/main.bro)
|
||||||
rest_target(${psd} base/frameworks/logging/main.bro)
|
rest_target(${psd} base/frameworks/logging/main.bro)
|
||||||
rest_target(${psd} base/frameworks/logging/postprocessors/scp.bro)
|
rest_target(${psd} base/frameworks/logging/postprocessors/scp.bro)
|
||||||
|
rest_target(${psd} base/frameworks/logging/postprocessors/sftp.bro)
|
||||||
rest_target(${psd} base/frameworks/logging/writers/ascii.bro)
|
rest_target(${psd} base/frameworks/logging/writers/ascii.bro)
|
||||||
rest_target(${psd} base/frameworks/metrics/cluster.bro)
|
rest_target(${psd} base/frameworks/metrics/cluster.bro)
|
||||||
rest_target(${psd} base/frameworks/metrics/main.bro)
|
rest_target(${psd} base/frameworks/metrics/main.bro)
|
||||||
|
@ -102,6 +103,7 @@ rest_target(${psd} policy/misc/analysis-groups.bro)
|
||||||
rest_target(${psd} policy/misc/capture-loss.bro)
|
rest_target(${psd} policy/misc/capture-loss.bro)
|
||||||
rest_target(${psd} policy/misc/loaded-scripts.bro)
|
rest_target(${psd} policy/misc/loaded-scripts.bro)
|
||||||
rest_target(${psd} policy/misc/profiling.bro)
|
rest_target(${psd} policy/misc/profiling.bro)
|
||||||
|
rest_target(${psd} policy/misc/stats.bro)
|
||||||
rest_target(${psd} policy/misc/trim-trace-file.bro)
|
rest_target(${psd} policy/misc/trim-trace-file.bro)
|
||||||
rest_target(${psd} policy/protocols/conn/known-hosts.bro)
|
rest_target(${psd} policy/protocols/conn/known-hosts.bro)
|
||||||
rest_target(${psd} policy/protocols/conn/known-services.bro)
|
rest_target(${psd} policy/protocols/conn/known-services.bro)
|
||||||
|
|
|
@ -6,6 +6,11 @@ Types
|
||||||
|
|
||||||
The Bro scripting language supports the following built-in types.
|
The Bro scripting language supports the following built-in types.
|
||||||
|
|
||||||
|
.. bro:type:: void
|
||||||
|
|
||||||
|
An internal Bro type representing an absence of a type. Should
|
||||||
|
most often be seen as a possible function return type.
|
||||||
|
|
||||||
.. bro:type:: bool
|
.. bro:type:: bool
|
||||||
|
|
||||||
Reflects a value with one of two meanings: true or false. The two
|
Reflects a value with one of two meanings: true or false. The two
|
||||||
|
@ -621,6 +626,10 @@ scripting language supports the following built-in attributes.
|
||||||
|
|
||||||
Writes a record field to the associated log stream.
|
Writes a record field to the associated log stream.
|
||||||
|
|
||||||
|
.. bro:attr:: &error_handler
|
||||||
|
|
||||||
|
.. TODO: needs documented
|
||||||
|
|
||||||
.. bro:attr:: (&tracked)
|
.. bro:attr:: (&tracked)
|
||||||
|
|
||||||
.. TODO: needs documented or removed if it's not used anywhere.
|
.. TODO: needs documented or removed if it's not used anywhere.
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
##! This is an example script that demonstrates how to document. Comments
|
##! This is an example script that demonstrates documentation features.
|
||||||
##! of the form ``##!`` are for the script summary. The contents of
|
##! Comments of the form ``##!`` are for the script summary. The contents of
|
||||||
##! these comments are transferred directly into the auto-generated
|
##! these comments are transferred directly into the auto-generated
|
||||||
##! `reStructuredText <http://docutils.sourceforge.net/rst.html>`_
|
##! `reStructuredText <http://docutils.sourceforge.net/rst.html>`_
|
||||||
##! (reST) document's summary section.
|
##! (reST) document's summary section.
|
||||||
|
@ -22,8 +22,8 @@
|
||||||
# field comments, it's necessary to disambiguate the field with
|
# field comments, it's necessary to disambiguate the field with
|
||||||
# which a comment associates: e.g. "##<" can be used on the same line
|
# which a comment associates: e.g. "##<" can be used on the same line
|
||||||
# as a field to signify the comment relates to it and not the
|
# as a field to signify the comment relates to it and not the
|
||||||
# following field. "##<" is not meant for general use, just
|
# following field. "##<" can also be used more generally in any
|
||||||
# record/enum fields.
|
# variable declarations to associate with the last-declared identifier.
|
||||||
#
|
#
|
||||||
# Generally, the auto-doc comments (##) are associated with the
|
# Generally, the auto-doc comments (##) are associated with the
|
||||||
# next declaration/identifier found in the script, but the doc framework
|
# next declaration/identifier found in the script, but the doc framework
|
||||||
|
@ -151,7 +151,7 @@ export {
|
||||||
const an_option: set[addr, addr, string] &redef;
|
const an_option: set[addr, addr, string] &redef;
|
||||||
|
|
||||||
# default initialization will be self-documenting
|
# default initialization will be self-documenting
|
||||||
const option_with_init = 0.01 secs &redef;
|
const option_with_init = 0.01 secs &redef; ##< More docs can be added here.
|
||||||
|
|
||||||
############## state variables ############
|
############## state variables ############
|
||||||
# right now, I'm defining this as any global
|
# right now, I'm defining this as any global
|
||||||
|
@ -183,6 +183,7 @@ export {
|
||||||
|
|
||||||
## Summarize "an_event" here.
|
## Summarize "an_event" here.
|
||||||
## Give more details about "an_event" here.
|
## Give more details about "an_event" here.
|
||||||
|
## Example::an_event should not be confused as a parameter.
|
||||||
## name: describe the argument here
|
## name: describe the argument here
|
||||||
global an_event: event(name: string);
|
global an_event: event(name: string);
|
||||||
|
|
||||||
|
|
|
@ -168,10 +168,6 @@ New Default Settings
|
||||||
are loaded. See ``PacketFilter::all_packets`` for how to revert to old
|
are loaded. See ``PacketFilter::all_packets`` for how to revert to old
|
||||||
behavior.
|
behavior.
|
||||||
|
|
||||||
- By default, Bro now sets a libpcap snaplen of 65535. Depending on
|
|
||||||
the OS, this may have performance implications and you can use the
|
|
||||||
``--snaplen`` option to change the value.
|
|
||||||
|
|
||||||
API Changes
|
API Changes
|
||||||
-----------
|
-----------
|
||||||
|
|
||||||
|
|
|
@ -44,7 +44,7 @@ event bro_init() &priority=9
|
||||||
{
|
{
|
||||||
if ( n$node_type == WORKER && n$proxy == node )
|
if ( n$node_type == WORKER && n$proxy == node )
|
||||||
Communication::nodes[i] =
|
Communication::nodes[i] =
|
||||||
[$host=n$ip, $connect=F, $class=i, $events=worker2proxy_events];
|
[$host=n$ip, $connect=F, $class=i, $sync=T, $auth=T, $events=worker2proxy_events];
|
||||||
|
|
||||||
# accepts connections from the previous one.
|
# accepts connections from the previous one.
|
||||||
# (This is not ideal for setups with many proxies)
|
# (This is not ideal for setups with many proxies)
|
||||||
|
|
|
@ -1,43 +1,30 @@
|
||||||
##! This is a utility script that sends the current values of all &redef'able
|
##! The control framework provides the foundation for providing "commands"
|
||||||
##! consts to a remote Bro then sends the :bro:id:`configuration_update` event
|
##! that can be taken remotely at runtime to modify a running Bro instance
|
||||||
##! and terminates processing.
|
##! or collect information from the running instance.
|
||||||
##!
|
|
||||||
##! Intended to be used from the command line like this when starting a controller::
|
|
||||||
##!
|
|
||||||
##! bro <scripts> frameworks/control/controller Control::host=<host_addr> Control::port=<host_port> Control::cmd=<command> [Control::arg=<arg>]
|
|
||||||
##!
|
|
||||||
##! A controllee only needs to load the controllee script in addition
|
|
||||||
##! to the specific analysis scripts desired. It may also need a node
|
|
||||||
##! configured as a controller node in the communications nodes configuration::
|
|
||||||
##!
|
|
||||||
##! bro <scripts> frameworks/control/controllee
|
|
||||||
##!
|
|
||||||
##! To use the framework as a controllee, it only needs to be loaded and
|
|
||||||
##! the controlled node need to accept all events in the "Control::" namespace
|
|
||||||
##! from the host where the control actions will be performed from along with
|
|
||||||
##! using the "control" class.
|
|
||||||
|
|
||||||
module Control;
|
module Control;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
## This is the address of the host that will be controlled.
|
## The address of the host that will be controlled.
|
||||||
const host = 0.0.0.0 &redef;
|
const host = 0.0.0.0 &redef;
|
||||||
|
|
||||||
## This is the port of the host that will be controlled.
|
## The port of the host that will be controlled.
|
||||||
const host_port = 0/tcp &redef;
|
const host_port = 0/tcp &redef;
|
||||||
|
|
||||||
## This is the command that is being done. It's typically set on the
|
## The command that is being done. It's typically set on the
|
||||||
## command line and influences whether this instance starts up as a
|
## command line.
|
||||||
## controller or controllee.
|
|
||||||
const cmd = "" &redef;
|
const cmd = "" &redef;
|
||||||
|
|
||||||
## This can be used by commands that take an argument.
|
## This can be used by commands that take an argument.
|
||||||
const arg = "" &redef;
|
const arg = "" &redef;
|
||||||
|
|
||||||
|
## Events that need to be handled by controllers.
|
||||||
const controller_events = /Control::.*_request/ &redef;
|
const controller_events = /Control::.*_request/ &redef;
|
||||||
|
|
||||||
|
## Events that need to be handled by controllees.
|
||||||
const controllee_events = /Control::.*_response/ &redef;
|
const controllee_events = /Control::.*_response/ &redef;
|
||||||
|
|
||||||
## These are the commands that can be given on the command line for
|
## The commands that can currently be given on the command line for
|
||||||
## remote control.
|
## remote control.
|
||||||
const commands: set[string] = {
|
const commands: set[string] = {
|
||||||
"id_value",
|
"id_value",
|
||||||
|
@ -45,15 +32,15 @@ export {
|
||||||
"net_stats",
|
"net_stats",
|
||||||
"configuration_update",
|
"configuration_update",
|
||||||
"shutdown",
|
"shutdown",
|
||||||
};
|
} &redef;
|
||||||
|
|
||||||
## Variable IDs that are to be ignored by the update process.
|
## Variable IDs that are to be ignored by the update process.
|
||||||
const ignore_ids: set[string] = {
|
const ignore_ids: set[string] = { };
|
||||||
};
|
|
||||||
|
|
||||||
## Event for requesting the value of an ID (a variable).
|
## Event for requesting the value of an ID (a variable).
|
||||||
global id_value_request: event(id: string);
|
global id_value_request: event(id: string);
|
||||||
## Event for returning the value of an ID after an :bro:id:`id_request` event.
|
## Event for returning the value of an ID after an
|
||||||
|
## :bro:id:`Control::id_value_request` event.
|
||||||
global id_value_response: event(id: string, val: string);
|
global id_value_response: event(id: string, val: string);
|
||||||
|
|
||||||
## Requests the current communication status.
|
## Requests the current communication status.
|
||||||
|
@ -68,7 +55,8 @@ export {
|
||||||
|
|
||||||
## Inform the remote Bro instance that it's configuration may have been updated.
|
## Inform the remote Bro instance that it's configuration may have been updated.
|
||||||
global configuration_update_request: event();
|
global configuration_update_request: event();
|
||||||
## This event is a wrapper and alias for the :bro:id:`configuration_update_request` event.
|
## This event is a wrapper and alias for the
|
||||||
|
## :bro:id:`Control::configuration_update_request` event.
|
||||||
## This event is also a primary hooking point for the control framework.
|
## This event is also a primary hooking point for the control framework.
|
||||||
global configuration_update: event();
|
global configuration_update: event();
|
||||||
## Message in response to a configuration update request.
|
## Message in response to a configuration update request.
|
||||||
|
|
|
@ -80,15 +80,15 @@ signature irc_server_reply {
|
||||||
tcp-state responder
|
tcp-state responder
|
||||||
}
|
}
|
||||||
|
|
||||||
signature irc_sig3 {
|
signature irc_server_to_server1 {
|
||||||
ip-proto == tcp
|
ip-proto == tcp
|
||||||
payload /(.*\x0a)*(\x20)*[Ss][Ee][Rr][Vv][Ee][Rr](\x20)+.+\x0a/
|
payload /(|.*[\r\n]) *[Ss][Ee][Rr][Vv][Ee][Rr] +[^ ]+ +[0-9]+ +:.+[\r\n]/
|
||||||
}
|
}
|
||||||
|
|
||||||
signature irc_sig4 {
|
signature irc_server_to_server2 {
|
||||||
ip-proto == tcp
|
ip-proto == tcp
|
||||||
payload /(.*\x0a)*(\x20)*[Ss][Ee][Rr][Vv][Ee][Rr](\x20)+.+\x0a/
|
payload /(|.*[\r\n]) *[Ss][Ee][Rr][Vv][Ee][Rr] +[^ ]+ +[0-9]+ +:.+[\r\n]/
|
||||||
requires-reverse-signature irc_sig3
|
requires-reverse-signature irc_server_to_server1
|
||||||
enable "irc"
|
enable "irc"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -11,7 +11,7 @@
|
||||||
# user_name
|
# user_name
|
||||||
# file_name
|
# file_name
|
||||||
# file_md5
|
# file_md5
|
||||||
# x509_cert - DER encoded, not PEM (ascii armored)
|
# x509_md5
|
||||||
|
|
||||||
# Example tags:
|
# Example tags:
|
||||||
# infrastructure
|
# infrastructure
|
||||||
|
@ -25,6 +25,7 @@
|
||||||
module Intel;
|
module Intel;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The intel logging stream identifier.
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
|
@ -33,71 +34,116 @@ export {
|
||||||
Detection,
|
Detection,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Record type used for logging information from the intelligence framework.
|
||||||
|
## Primarily for problems or oddities with inserting and querying data.
|
||||||
|
## This is important since the content of the intelligence framework can
|
||||||
|
## change quite dramatically during runtime and problems may be introduced
|
||||||
|
## into the data.
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## The current network time.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
|
## Represents the severity of the message.
|
||||||
## This value should be one of: "info", "warn", "error"
|
## This value should be one of: "info", "warn", "error"
|
||||||
level: string &log;
|
level: string &log;
|
||||||
|
## The message.
|
||||||
message: string &log;
|
message: string &log;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Record to represent metadata associated with a single piece of
|
||||||
|
## intelligence.
|
||||||
type MetaData: record {
|
type MetaData: record {
|
||||||
|
## A description for the data.
|
||||||
desc: string &optional;
|
desc: string &optional;
|
||||||
|
## A URL where more information may be found about the intelligence.
|
||||||
url: string &optional;
|
url: string &optional;
|
||||||
|
## The time at which the data was first declared to be intelligence.
|
||||||
first_seen: time &optional;
|
first_seen: time &optional;
|
||||||
|
## When this data was most recent inserted into the framework.
|
||||||
latest_seen: time &optional;
|
latest_seen: time &optional;
|
||||||
|
## Arbitrary text tags for the data.
|
||||||
tags: set[string];
|
tags: set[string];
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Record to represent a singular piece of intelligence.
|
||||||
type Item: record {
|
type Item: record {
|
||||||
|
## If the data is an IP address, this hold the address.
|
||||||
ip: addr &optional;
|
ip: addr &optional;
|
||||||
|
## If the data is textual, this holds the text.
|
||||||
str: string &optional;
|
str: string &optional;
|
||||||
|
## If the data is numeric, this holds the number.
|
||||||
num: int &optional;
|
num: int &optional;
|
||||||
|
## The subtype of the data for when either the $str or $num fields are
|
||||||
|
## given. If one of those fields are given, this field must be present.
|
||||||
subtype: string &optional;
|
subtype: string &optional;
|
||||||
|
|
||||||
|
## The next five fields are temporary until a better model for
|
||||||
|
## attaching metadata to an intelligence item is created.
|
||||||
desc: string &optional;
|
desc: string &optional;
|
||||||
url: string &optional;
|
url: string &optional;
|
||||||
first_seen: time &optional;
|
first_seen: time &optional;
|
||||||
latest_seen: time &optional;
|
latest_seen: time &optional;
|
||||||
tags: set[string];
|
tags: set[string];
|
||||||
|
|
||||||
## These single string tags are throw away until pybroccoli supports sets
|
## These single string tags are throw away until pybroccoli supports sets.
|
||||||
tag1: string &optional;
|
tag1: string &optional;
|
||||||
tag2: string &optional;
|
tag2: string &optional;
|
||||||
tag3: string &optional;
|
tag3: string &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Record model used for constructing queries against the intelligence
|
||||||
|
## framework.
|
||||||
type QueryItem: record {
|
type QueryItem: record {
|
||||||
|
## If an IP address is being queried for, this field should be given.
|
||||||
ip: addr &optional;
|
ip: addr &optional;
|
||||||
|
## If a string is being queried for, this field should be given.
|
||||||
str: string &optional;
|
str: string &optional;
|
||||||
|
## If numeric data is being queried for, this field should be given.
|
||||||
num: int &optional;
|
num: int &optional;
|
||||||
|
## If either a string or number is being queried for, this field should
|
||||||
|
## indicate the subtype of the data.
|
||||||
subtype: string &optional;
|
subtype: string &optional;
|
||||||
|
|
||||||
|
## A set of tags where if a single metadata record attached to an item
|
||||||
|
## has any one of the tags defined in this field, it will match.
|
||||||
or_tags: set[string] &optional;
|
or_tags: set[string] &optional;
|
||||||
|
## A set of tags where a single metadata record attached to an item
|
||||||
|
## must have all of the tags defined in this field.
|
||||||
and_tags: set[string] &optional;
|
and_tags: set[string] &optional;
|
||||||
|
|
||||||
## The predicate can be given when searching for a match. It will
|
## The predicate can be given when searching for a match. It will
|
||||||
## be tested against every :bro:type:`MetaData` item associated with
|
## be tested against every :bro:type:`Intel::MetaData` item associated
|
||||||
## the data being matched on. If it returns T a single time, the
|
## with the data being matched on. If it returns T a single time, the
|
||||||
## matcher will consider that the item has matched.
|
## matcher will consider that the item has matched. This field can
|
||||||
|
## be used for constructing arbitrarily complex queries that may not
|
||||||
|
## be possible with the $or_tags or $and_tags fields.
|
||||||
pred: function(meta: Intel::MetaData): bool &optional;
|
pred: function(meta: Intel::MetaData): bool &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Function to insert data into the intelligence framework.
|
||||||
|
##
|
||||||
|
## item: The data item.
|
||||||
|
##
|
||||||
|
## Returns: T if the data was successfully inserted into the framework,
|
||||||
|
## otherwise it returns F.
|
||||||
global insert: function(item: Item): bool;
|
global insert: function(item: Item): bool;
|
||||||
global insert_event: event(item: Item);
|
|
||||||
global matcher: function(item: QueryItem): bool;
|
|
||||||
|
|
||||||
type MetaDataStore: table[count] of MetaData;
|
## A wrapper for the :bro:id:`Intel::insert` function. This is primarily
|
||||||
type DataStore: record {
|
## used as the external API for inserting data into the intelligence
|
||||||
|
## using Broccoli.
|
||||||
|
global insert_event: event(item: Item);
|
||||||
|
|
||||||
|
## Function for matching data within the intelligence framework.
|
||||||
|
global matcher: function(item: QueryItem): bool;
|
||||||
|
}
|
||||||
|
|
||||||
|
type MetaDataStore: table[count] of MetaData;
|
||||||
|
type DataStore: record {
|
||||||
ip_data: table[addr] of MetaDataStore;
|
ip_data: table[addr] of MetaDataStore;
|
||||||
## The first string is the actual value and the second string is the subtype.
|
# The first string is the actual value and the second string is the subtype.
|
||||||
string_data: table[string, string] of MetaDataStore;
|
string_data: table[string, string] of MetaDataStore;
|
||||||
int_data: table[int, string] of MetaDataStore;
|
int_data: table[int, string] of MetaDataStore;
|
||||||
};
|
};
|
||||||
global data_store: DataStore;
|
global data_store: DataStore;
|
||||||
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
event bro_init()
|
event bro_init()
|
||||||
{
|
{
|
||||||
|
|
|
@ -1 +1,2 @@
|
||||||
@load ./scp
|
@load ./scp
|
||||||
|
@load ./sftp
|
||||||
|
|
|
@ -47,6 +47,10 @@ export {
|
||||||
## copy of the rotated-log to each destination in the set. This
|
## copy of the rotated-log to each destination in the set. This
|
||||||
## table can be modified at run-time.
|
## table can be modified at run-time.
|
||||||
global scp_destinations: table[Writer, string] of set[SCPDestination];
|
global scp_destinations: table[Writer, string] of set[SCPDestination];
|
||||||
|
|
||||||
|
## Default naming format for timestamps embedded into log filenames
|
||||||
|
## that use the SCP rotator.
|
||||||
|
const scp_rotation_date_format = "%Y-%m-%d-%H-%M-%S" &redef;
|
||||||
}
|
}
|
||||||
|
|
||||||
function scp_postprocessor(info: Log::RotationInfo): bool
|
function scp_postprocessor(info: Log::RotationInfo): bool
|
||||||
|
@ -56,7 +60,11 @@ function scp_postprocessor(info: Log::RotationInfo): bool
|
||||||
|
|
||||||
local command = "";
|
local command = "";
|
||||||
for ( d in scp_destinations[info$writer, info$path] )
|
for ( d in scp_destinations[info$writer, info$path] )
|
||||||
command += fmt("scp %s %s@%s:%s;", info$fname, d$user, d$host, d$path);
|
{
|
||||||
|
local dst = fmt("%s/%s.%s.log", d$path, info$path,
|
||||||
|
strftime(Log::scp_rotation_date_format, info$open));
|
||||||
|
command += fmt("scp %s %s@%s:%s;", info$fname, d$user, d$host, dst);
|
||||||
|
}
|
||||||
|
|
||||||
command += fmt("/bin/rm %s", info$fname);
|
command += fmt("/bin/rm %s", info$fname);
|
||||||
system(command);
|
system(command);
|
||||||
|
|
73
scripts/base/frameworks/logging/postprocessors/sftp.bro
Normal file
73
scripts/base/frameworks/logging/postprocessors/sftp.bro
Normal file
|
@ -0,0 +1,73 @@
|
||||||
|
##! This script defines a postprocessing function that can be applied
|
||||||
|
##! to a logging filter in order to automatically SFTP
|
||||||
|
##! a log stream (or a subset of it) to a remote host at configurable
|
||||||
|
##! rotation time intervals. Generally, to use this functionality
|
||||||
|
##! you must handle the :bro:id:`bro_init` event and do the following
|
||||||
|
##! in your handler:
|
||||||
|
##!
|
||||||
|
##! 1) Create a new :bro:type:`Log::Filter` record that defines a name/path,
|
||||||
|
##! rotation interval, and set the ``postprocessor`` to
|
||||||
|
##! :bro:id:`Log::sftp_postprocessor`.
|
||||||
|
##! 2) Add the filter to a logging stream using :bro:id:`Log::add_filter`.
|
||||||
|
##! 3) Add a table entry to :bro:id:`Log::sftp_destinations` for the filter's
|
||||||
|
##! writer/path pair which defines a set of :bro:type:`Log::SFTPDestination`
|
||||||
|
##! records.
|
||||||
|
|
||||||
|
module Log;
|
||||||
|
|
||||||
|
export {
|
||||||
|
## Securely transfers the rotated-log to all the remote hosts
|
||||||
|
## defined in :bro:id:`Log::sftp_destinations` and then deletes
|
||||||
|
## the local copy of the rotated-log. It's not active when
|
||||||
|
## reading from trace files.
|
||||||
|
##
|
||||||
|
## info: A record holding meta-information about the log file to be
|
||||||
|
## postprocessed.
|
||||||
|
##
|
||||||
|
## Returns: True if sftp system command was initiated or
|
||||||
|
## if no destination was configured for the log as described
|
||||||
|
## by *info*.
|
||||||
|
global sftp_postprocessor: function(info: Log::RotationInfo): bool;
|
||||||
|
|
||||||
|
## A container that describes the remote destination for the SFTP command,
|
||||||
|
## comprised of the username, host, and path at which to upload the file.
|
||||||
|
type SFTPDestination: record {
|
||||||
|
## The remote user to log in as. A trust mechanism should be
|
||||||
|
## pre-established.
|
||||||
|
user: string;
|
||||||
|
## The remote host to which to transfer logs.
|
||||||
|
host: string;
|
||||||
|
## The path/directory on the remote host to send logs.
|
||||||
|
path: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
## A table indexed by a particular log writer and filter path, that yields
|
||||||
|
## a set remote destinations. The :bro:id:`Log::sftp_postprocessor`
|
||||||
|
## function queries this table upon log rotation and performs a secure
|
||||||
|
## transfer of the rotated-log to each destination in the set. This
|
||||||
|
## table can be modified at run-time.
|
||||||
|
global sftp_destinations: table[Writer, string] of set[SFTPDestination];
|
||||||
|
|
||||||
|
## Default naming format for timestamps embedded into log filenames
|
||||||
|
## that use the SFTP rotator.
|
||||||
|
const sftp_rotation_date_format = "%Y-%m-%d-%H-%M-%S" &redef;
|
||||||
|
}
|
||||||
|
|
||||||
|
function sftp_postprocessor(info: Log::RotationInfo): bool
|
||||||
|
{
|
||||||
|
if ( reading_traces() || [info$writer, info$path] !in sftp_destinations )
|
||||||
|
return T;
|
||||||
|
|
||||||
|
local command = "";
|
||||||
|
for ( d in sftp_destinations[info$writer, info$path] )
|
||||||
|
{
|
||||||
|
local dst = fmt("%s/%s.%s.log", d$path, info$path,
|
||||||
|
strftime(Log::sftp_rotation_date_format, info$open));
|
||||||
|
command += fmt("echo put %s %s | sftp -b - %s@%s;", info$fname, dst,
|
||||||
|
d$user, d$host);
|
||||||
|
}
|
||||||
|
|
||||||
|
command += fmt("/bin/rm %s", info$fname);
|
||||||
|
system(command);
|
||||||
|
return T;
|
||||||
|
}
|
|
@ -21,8 +21,9 @@ export {
|
||||||
## Separator between set elements.
|
## Separator between set elements.
|
||||||
const set_separator = "," &redef;
|
const set_separator = "," &redef;
|
||||||
|
|
||||||
## String to use for empty fields.
|
## String to use for empty fields. This should be different from
|
||||||
const empty_field = "-" &redef;
|
## *unset_field* to make the output non-ambigious.
|
||||||
|
const empty_field = "(empty)" &redef;
|
||||||
|
|
||||||
## String to use for an unset &optional field.
|
## String to use for an unset &optional field.
|
||||||
const unset_field = "-" &redef;
|
const unset_field = "-" &redef;
|
||||||
|
|
|
@ -13,11 +13,11 @@
|
||||||
module Metrics;
|
module Metrics;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
## This value allows a user to decide how large of result groups the
|
## Allows a user to decide how large of result groups the
|
||||||
## workers should transmit values.
|
## workers should transmit values for cluster metric aggregation.
|
||||||
const cluster_send_in_groups_of = 50 &redef;
|
const cluster_send_in_groups_of = 50 &redef;
|
||||||
|
|
||||||
## This is the percent of the full threshold value that needs to be met
|
## The percent of the full threshold value that needs to be met
|
||||||
## on a single worker for that worker to send the value to its manager in
|
## on a single worker for that worker to send the value to its manager in
|
||||||
## order for it to request a global view for that value. There is no
|
## order for it to request a global view for that value. There is no
|
||||||
## requirement that the manager requests a global view for the index
|
## requirement that the manager requests a global view for the index
|
||||||
|
@ -25,11 +25,11 @@ export {
|
||||||
## recently.
|
## recently.
|
||||||
const cluster_request_global_view_percent = 0.1 &redef;
|
const cluster_request_global_view_percent = 0.1 &redef;
|
||||||
|
|
||||||
## This event is sent by the manager in a cluster to initiate the
|
## Event sent by the manager in a cluster to initiate the
|
||||||
## collection of metrics values for a filter.
|
## collection of metrics values for a filter.
|
||||||
global cluster_filter_request: event(uid: string, id: ID, filter_name: string);
|
global cluster_filter_request: event(uid: string, id: ID, filter_name: string);
|
||||||
|
|
||||||
## This event is sent by nodes that are collecting metrics after receiving
|
## Event sent by nodes that are collecting metrics after receiving
|
||||||
## a request for the metric filter from the manager.
|
## a request for the metric filter from the manager.
|
||||||
global cluster_filter_response: event(uid: string, id: ID, filter_name: string, data: MetricTable, done: bool);
|
global cluster_filter_response: event(uid: string, id: ID, filter_name: string, data: MetricTable, done: bool);
|
||||||
|
|
||||||
|
@ -40,12 +40,12 @@ export {
|
||||||
global cluster_index_request: event(uid: string, id: ID, filter_name: string, index: Index);
|
global cluster_index_request: event(uid: string, id: ID, filter_name: string, index: Index);
|
||||||
|
|
||||||
## This event is sent by nodes in response to a
|
## This event is sent by nodes in response to a
|
||||||
## :bro:id:`cluster_index_request` event.
|
## :bro:id:`Metrics::cluster_index_request` event.
|
||||||
global cluster_index_response: event(uid: string, id: ID, filter_name: string, index: Index, val: count);
|
global cluster_index_response: event(uid: string, id: ID, filter_name: string, index: Index, val: count);
|
||||||
|
|
||||||
## This is sent by workers to indicate that they crossed the percent of the
|
## This is sent by workers to indicate that they crossed the percent of the
|
||||||
## current threshold by the percentage defined globally in
|
## current threshold by the percentage defined globally in
|
||||||
## :bro:id:`cluster_request_global_view_percent`
|
## :bro:id:`Metrics::cluster_request_global_view_percent`
|
||||||
global cluster_index_intermediate_response: event(id: Metrics::ID, filter_name: string, index: Metrics::Index, val: count);
|
global cluster_index_intermediate_response: event(id: Metrics::ID, filter_name: string, index: Metrics::Index, val: count);
|
||||||
|
|
||||||
## This event is scheduled internally on workers to send result chunks.
|
## This event is scheduled internally on workers to send result chunks.
|
||||||
|
|
|
@ -1,13 +1,16 @@
|
||||||
##! This is the implementation of the metrics framework.
|
##! The metrics framework provides a way to count and measure data.
|
||||||
|
|
||||||
@load base/frameworks/notice
|
@load base/frameworks/notice
|
||||||
|
|
||||||
module Metrics;
|
module Metrics;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The metrics logging stream identifier.
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
|
## Identifiers for metrics to collect.
|
||||||
type ID: enum {
|
type ID: enum {
|
||||||
|
## Blank placeholder value.
|
||||||
NOTHING,
|
NOTHING,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -15,10 +18,13 @@ export {
|
||||||
## current value to the logging stream.
|
## current value to the logging stream.
|
||||||
const default_break_interval = 15mins &redef;
|
const default_break_interval = 15mins &redef;
|
||||||
|
|
||||||
## This is the interval for how often notices will happen after they have
|
## This is the interval for how often threshold based notices will happen
|
||||||
## already fired.
|
## after they have already fired.
|
||||||
const renotice_interval = 1hr &redef;
|
const renotice_interval = 1hr &redef;
|
||||||
|
|
||||||
|
## Represents a thing which is having metrics collected for it. An instance
|
||||||
|
## of this record type and a :bro:type:`Metrics::ID` together represent a
|
||||||
|
## single measurement.
|
||||||
type Index: record {
|
type Index: record {
|
||||||
## Host is the value to which this metric applies.
|
## Host is the value to which this metric applies.
|
||||||
host: addr &optional;
|
host: addr &optional;
|
||||||
|
@ -37,17 +43,30 @@ export {
|
||||||
network: subnet &optional;
|
network: subnet &optional;
|
||||||
} &log;
|
} &log;
|
||||||
|
|
||||||
|
## The record type that is used for logging metrics.
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## Timestamp at which the metric was "broken".
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
|
## What measurement the metric represents.
|
||||||
metric_id: ID &log;
|
metric_id: ID &log;
|
||||||
|
## The name of the filter being logged. :bro:type:`Metrics::ID` values
|
||||||
|
## can have multiple filters which represent different perspectives on
|
||||||
|
## the data so this is necessary to understand the value.
|
||||||
filter_name: string &log;
|
filter_name: string &log;
|
||||||
|
## What the metric value applies to.
|
||||||
index: Index &log;
|
index: Index &log;
|
||||||
|
## The simple numeric value of the metric.
|
||||||
value: count &log;
|
value: count &log;
|
||||||
};
|
};
|
||||||
|
|
||||||
# TODO: configure a metrics filter logging stream to log the current
|
# TODO: configure a metrics filter logging stream to log the current
|
||||||
# metrics configuration in case someone is looking through
|
# metrics configuration in case someone is looking through
|
||||||
# old logs and the configuration has changed since then.
|
# old logs and the configuration has changed since then.
|
||||||
|
|
||||||
|
## Filters define how the data from a metric is aggregated and handled.
|
||||||
|
## Filters can be used to set how often the measurements are cut or "broken"
|
||||||
|
## and logged or how the data within them is aggregated. It's also
|
||||||
|
## possible to disable logging and use filters for thresholding.
|
||||||
type Filter: record {
|
type Filter: record {
|
||||||
## The :bro:type:`Metrics::ID` that this filter applies to.
|
## The :bro:type:`Metrics::ID` that this filter applies to.
|
||||||
id: ID &optional;
|
id: ID &optional;
|
||||||
|
@ -62,7 +81,7 @@ export {
|
||||||
aggregation_mask: count &optional;
|
aggregation_mask: count &optional;
|
||||||
## This is essentially a mapping table between addresses and subnets.
|
## This is essentially a mapping table between addresses and subnets.
|
||||||
aggregation_table: table[subnet] of subnet &optional;
|
aggregation_table: table[subnet] of subnet &optional;
|
||||||
## The interval at which the metric should be "broken" and written
|
## The interval at which this filter should be "broken" and written
|
||||||
## to the logging stream. The counters are also reset to zero at
|
## to the logging stream. The counters are also reset to zero at
|
||||||
## this time so any threshold based detection needs to be set to a
|
## this time so any threshold based detection needs to be set to a
|
||||||
## number that should be expected to happen within this period.
|
## number that should be expected to happen within this period.
|
||||||
|
@ -79,7 +98,7 @@ export {
|
||||||
notice_threshold: count &optional;
|
notice_threshold: count &optional;
|
||||||
## A series of thresholds at which to generate notices.
|
## A series of thresholds at which to generate notices.
|
||||||
notice_thresholds: vector of count &optional;
|
notice_thresholds: vector of count &optional;
|
||||||
## How often this notice should be raised for this metric index. It
|
## How often this notice should be raised for this filter. It
|
||||||
## will be generated everytime it crosses a threshold, but if the
|
## will be generated everytime it crosses a threshold, but if the
|
||||||
## $break_interval is set to 5mins and this is set to 1hr the notice
|
## $break_interval is set to 5mins and this is set to 1hr the notice
|
||||||
## only be generated once per hour even if something crosses the
|
## only be generated once per hour even if something crosses the
|
||||||
|
@ -87,15 +106,43 @@ export {
|
||||||
notice_freq: interval &optional;
|
notice_freq: interval &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Function to associate a metric filter with a metric ID.
|
||||||
|
##
|
||||||
|
## id: The metric ID that the filter should be associated with.
|
||||||
|
##
|
||||||
|
## filter: The record representing the filter configuration.
|
||||||
global add_filter: function(id: ID, filter: Filter);
|
global add_filter: function(id: ID, filter: Filter);
|
||||||
|
|
||||||
|
## Add data into a :bro:type:`Metrics::ID`. This should be called when
|
||||||
|
## a script has measured some point value and is ready to increment the
|
||||||
|
## counters.
|
||||||
|
##
|
||||||
|
## id: The metric ID that the data represents.
|
||||||
|
##
|
||||||
|
## index: The metric index that the value is to be added to.
|
||||||
|
##
|
||||||
|
## increment: How much to increment the counter by.
|
||||||
global add_data: function(id: ID, index: Index, increment: count);
|
global add_data: function(id: ID, index: Index, increment: count);
|
||||||
|
|
||||||
|
## Helper function to represent a :bro:type:`Metrics::Index` value as
|
||||||
|
## a simple string
|
||||||
|
##
|
||||||
|
## index: The metric index that is to be converted into a string.
|
||||||
|
##
|
||||||
|
## Returns: A string reprentation of the metric index.
|
||||||
global index2str: function(index: Index): string;
|
global index2str: function(index: Index): string;
|
||||||
|
|
||||||
# This is the event that is used to "finish" metrics and adapt the metrics
|
## Event that is used to "finish" metrics and adapt the metrics
|
||||||
# framework for clustered or non-clustered usage.
|
## framework for clustered or non-clustered usage.
|
||||||
|
##
|
||||||
|
## ..note: This is primarily intended for internal use.
|
||||||
global log_it: event(filter: Filter);
|
global log_it: event(filter: Filter);
|
||||||
|
|
||||||
|
## Event to access metrics records as they are passed to the logging framework.
|
||||||
global log_metrics: event(rec: Info);
|
global log_metrics: event(rec: Info);
|
||||||
|
|
||||||
|
## Type to store a table of metrics values. Interal use only!
|
||||||
|
type MetricTable: table[Index] of count &default=0;
|
||||||
}
|
}
|
||||||
|
|
||||||
redef record Notice::Info += {
|
redef record Notice::Info += {
|
||||||
|
@ -105,7 +152,6 @@ redef record Notice::Info += {
|
||||||
global metric_filters: table[ID] of vector of Filter = table();
|
global metric_filters: table[ID] of vector of Filter = table();
|
||||||
global filter_store: table[ID, string] of Filter = table();
|
global filter_store: table[ID, string] of Filter = table();
|
||||||
|
|
||||||
type MetricTable: table[Index] of count &default=0;
|
|
||||||
# This is indexed by metric ID and stream filter name.
|
# This is indexed by metric ID and stream filter name.
|
||||||
global store: table[ID, string] of MetricTable = table() &default=table();
|
global store: table[ID, string] of MetricTable = table() &default=table();
|
||||||
|
|
||||||
|
|
|
@ -31,6 +31,7 @@ export {
|
||||||
## Add a helper to the notice policy for looking up GeoIP data.
|
## Add a helper to the notice policy for looking up GeoIP data.
|
||||||
redef Notice::policy += {
|
redef Notice::policy += {
|
||||||
[$pred(n: Notice::Info) = { return (n$note in Notice::lookup_location_types); },
|
[$pred(n: Notice::Info) = { return (n$note in Notice::lookup_location_types); },
|
||||||
|
$action = ACTION_ADD_GEODATA,
|
||||||
$priority = 10],
|
$priority = 10],
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -11,8 +11,8 @@ module Notice;
|
||||||
export {
|
export {
|
||||||
redef enum Action += {
|
redef enum Action += {
|
||||||
## Indicate that the generated email should be addressed to the
|
## Indicate that the generated email should be addressed to the
|
||||||
## appropriate email addresses as found in the
|
## appropriate email addresses as found by the
|
||||||
## :bro:id:`Site::addr_to_emails` variable based on the relevant
|
## :bro:id:`Site::get_emails` function based on the relevant
|
||||||
## address or addresses indicated in the notice.
|
## address or addresses indicated in the notice.
|
||||||
ACTION_EMAIL_ADMIN
|
ACTION_EMAIL_ADMIN
|
||||||
};
|
};
|
||||||
|
|
|
@ -7,7 +7,7 @@ module Notice;
|
||||||
export {
|
export {
|
||||||
redef enum Action += {
|
redef enum Action += {
|
||||||
## Indicates that the notice should be sent to the pager email address
|
## Indicates that the notice should be sent to the pager email address
|
||||||
## configured in the :bro:id:`mail_page_dest` variable.
|
## configured in the :bro:id:`Notice::mail_page_dest` variable.
|
||||||
ACTION_PAGE
|
ACTION_PAGE
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -14,14 +14,17 @@ export {
|
||||||
## Address to send the pretty-printed reports to. Default if not set is
|
## Address to send the pretty-printed reports to. Default if not set is
|
||||||
## :bro:id:`Notice::mail_dest`.
|
## :bro:id:`Notice::mail_dest`.
|
||||||
const mail_dest_pretty_printed = "" &redef;
|
const mail_dest_pretty_printed = "" &redef;
|
||||||
|
|
||||||
## If an address from one of these networks is reported, we mark
|
## If an address from one of these networks is reported, we mark
|
||||||
## the entry with an addition quote symbol (that is, ">"). Many MUAs
|
## the entry with an additional quote symbol (i.e., ">"). Many MUAs
|
||||||
## then highlight such lines differently.
|
## then highlight such lines differently.
|
||||||
global flag_nets: set[subnet] &redef;
|
global flag_nets: set[subnet] &redef;
|
||||||
|
|
||||||
## Function that renders a single alarm. Can be overidden.
|
## Function that renders a single alarm. Can be overidden.
|
||||||
global pretty_print_alarm: function(out: file, n: Info) &redef;
|
global pretty_print_alarm: function(out: file, n: Info) &redef;
|
||||||
|
|
||||||
|
## Force generating mail file, even if reading from traces or no mail
|
||||||
|
## destination is defined. This is mainly for testing.
|
||||||
|
global force_email_summaries = F &redef;
|
||||||
}
|
}
|
||||||
|
|
||||||
# We maintain an old-style file recording the pretty-printed alarms.
|
# We maintain an old-style file recording the pretty-printed alarms.
|
||||||
|
@ -32,6 +35,9 @@ global pp_alarms_open: bool = F;
|
||||||
# Returns True if pretty-printed alarm summaries are activated.
|
# Returns True if pretty-printed alarm summaries are activated.
|
||||||
function want_pp() : bool
|
function want_pp() : bool
|
||||||
{
|
{
|
||||||
|
if ( force_email_summaries )
|
||||||
|
return T;
|
||||||
|
|
||||||
return (pretty_print_alarms && ! reading_traces()
|
return (pretty_print_alarms && ! reading_traces()
|
||||||
&& (mail_dest != "" || mail_dest_pretty_printed != ""));
|
&& (mail_dest != "" || mail_dest_pretty_printed != ""));
|
||||||
}
|
}
|
||||||
|
@ -44,34 +50,45 @@ function pp_open()
|
||||||
|
|
||||||
pp_alarms_open = T;
|
pp_alarms_open = T;
|
||||||
pp_alarms = open(pp_alarms_name);
|
pp_alarms = open(pp_alarms_name);
|
||||||
|
|
||||||
local dest = mail_dest_pretty_printed != "" ? mail_dest_pretty_printed
|
|
||||||
: mail_dest;
|
|
||||||
|
|
||||||
local headers = email_headers("Alarm summary", dest);
|
|
||||||
write_file(pp_alarms, headers + "\n");
|
|
||||||
}
|
}
|
||||||
|
|
||||||
# Closes and mails out the current output file.
|
# Closes and mails out the current output file.
|
||||||
function pp_send()
|
function pp_send(rinfo: Log::RotationInfo)
|
||||||
{
|
{
|
||||||
if ( ! pp_alarms_open )
|
if ( ! pp_alarms_open )
|
||||||
return;
|
return;
|
||||||
|
|
||||||
write_file(pp_alarms, "\n\n--\n[Automatically generated]\n\n");
|
write_file(pp_alarms, "\n\n--\n[Automatically generated]\n\n");
|
||||||
close(pp_alarms);
|
close(pp_alarms);
|
||||||
|
|
||||||
system(fmt("/bin/cat %s | %s -t -oi && /bin/rm %s",
|
|
||||||
pp_alarms_name, sendmail, pp_alarms_name));
|
|
||||||
|
|
||||||
pp_alarms_open = F;
|
pp_alarms_open = F;
|
||||||
|
|
||||||
|
local from = strftime("%H:%M:%S", rinfo$open);
|
||||||
|
local to = strftime("%H:%M:%S", rinfo$close);
|
||||||
|
local subject = fmt("Alarm summary from %s-%s", from, to);
|
||||||
|
local dest = mail_dest_pretty_printed != "" ? mail_dest_pretty_printed
|
||||||
|
: mail_dest;
|
||||||
|
|
||||||
|
if ( dest == "" )
|
||||||
|
# No mail destination configured, just leave the file alone. This is mainly for
|
||||||
|
# testing.
|
||||||
|
return;
|
||||||
|
|
||||||
|
local headers = email_headers(subject, dest);
|
||||||
|
|
||||||
|
local header_name = pp_alarms_name + ".tmp";
|
||||||
|
local header = open(header_name);
|
||||||
|
write_file(header, headers + "\n");
|
||||||
|
close(header);
|
||||||
|
|
||||||
|
system(fmt("/bin/cat %s %s | %s -t -oi && /bin/rm -f %s %s",
|
||||||
|
header_name, pp_alarms_name, sendmail, header_name, pp_alarms_name));
|
||||||
}
|
}
|
||||||
|
|
||||||
# Postprocessor function that triggers the email.
|
# Postprocessor function that triggers the email.
|
||||||
function pp_postprocessor(info: Log::RotationInfo): bool
|
function pp_postprocessor(info: Log::RotationInfo): bool
|
||||||
{
|
{
|
||||||
if ( want_pp() )
|
if ( want_pp() )
|
||||||
pp_send();
|
pp_send(info);
|
||||||
|
|
||||||
return T;
|
return T;
|
||||||
}
|
}
|
||||||
|
@ -93,7 +110,7 @@ event notice(n: Notice::Info) &priority=-5
|
||||||
if ( ! want_pp() )
|
if ( ! want_pp() )
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if ( ACTION_LOG !in n$actions )
|
if ( ACTION_ALARM !in n$actions )
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if ( ! pp_alarms_open )
|
if ( ! pp_alarms_open )
|
||||||
|
@ -154,30 +171,24 @@ function pretty_print_alarm(out: file, n: Info)
|
||||||
|
|
||||||
if ( n?$id )
|
if ( n?$id )
|
||||||
{
|
{
|
||||||
orig_p = fmt(":%s", n$id$orig_p);
|
h1 = n$id$orig_h;
|
||||||
resp_p = fmt(":%s", n$id$resp_p);
|
h2 = n$id$resp_h;
|
||||||
|
who = fmt("%s:%s -> %s:%s", h1, n$id$orig_p, h2, n$id$resp_p);
|
||||||
}
|
}
|
||||||
|
else if ( n?$src && n?$dst )
|
||||||
if ( n?$src && n?$dst )
|
|
||||||
{
|
{
|
||||||
h1 = n$src;
|
h1 = n$src;
|
||||||
h2 = n$dst;
|
h2 = n$dst;
|
||||||
who = fmt("%s%s -> %s%s", h1, orig_p, h2, resp_p);
|
who = fmt("%s -> %s", h1, h2);
|
||||||
|
}
|
||||||
|
else if ( n?$src )
|
||||||
|
{
|
||||||
|
h1 = n$src;
|
||||||
|
who = fmt("%s%s", h1, (n?$p ? fmt(":%s", n$p) : ""));
|
||||||
|
}
|
||||||
|
|
||||||
if ( n?$uid )
|
if ( n?$uid )
|
||||||
who = fmt("%s (uid %s)", who, n$uid );
|
who = fmt("%s (uid %s)", who, n$uid );
|
||||||
}
|
|
||||||
|
|
||||||
else if ( n?$src )
|
|
||||||
{
|
|
||||||
local p = "";
|
|
||||||
|
|
||||||
if ( n?$p )
|
|
||||||
p = fmt(":%s", n$p);
|
|
||||||
|
|
||||||
h1 = n$src;
|
|
||||||
who = fmt("%s%s", h1, p);
|
|
||||||
}
|
|
||||||
|
|
||||||
local flag = (h1 in flag_nets || h2 in flag_nets);
|
local flag = (h1 in flag_nets || h2 in flag_nets);
|
||||||
|
|
||||||
|
@ -191,6 +202,12 @@ function pretty_print_alarm(out: file, n: Info)
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if ( reading_traces() )
|
||||||
|
{
|
||||||
|
do_msg(out, n, line1, line2, line3, h1, "<skipped>", h2, "<skipped>");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
when ( local h1name = lookup_addr(h1) )
|
when ( local h1name = lookup_addr(h1) )
|
||||||
{
|
{
|
||||||
if ( h2 == 0.0.0.0 )
|
if ( h2 == 0.0.0.0 )
|
||||||
|
|
|
@ -1,8 +1,18 @@
|
||||||
|
##! Loading this script extends the :bro:enum:`Notice::ACTION_EMAIL` action
|
||||||
|
##! by appending to the email the hostnames associated with
|
||||||
|
##! :bro:type:`Notice::Info`'s *src* and *dst* fields as determined by a
|
||||||
|
##! DNS lookup.
|
||||||
|
|
||||||
@load ../main
|
@load ../main
|
||||||
|
|
||||||
module Notice;
|
module Notice;
|
||||||
|
|
||||||
# This probably doesn't actually work due to the async lookup_addr.
|
# We have to store references to the notices here because the when statement
|
||||||
|
# clones the frame which doesn't give us access to modify values outside
|
||||||
|
# of it's execution scope. (we get a clone of the notice instead of a
|
||||||
|
# reference to the original notice)
|
||||||
|
global tmp_notice_storage: table[string] of Notice::Info &create_expire=max_email_delay+10secs;
|
||||||
|
|
||||||
event Notice::notice(n: Notice::Info) &priority=10
|
event Notice::notice(n: Notice::Info) &priority=10
|
||||||
{
|
{
|
||||||
if ( ! n?$src && ! n?$dst )
|
if ( ! n?$src && ! n?$dst )
|
||||||
|
@ -12,21 +22,31 @@ event Notice::notice(n: Notice::Info) &priority=10
|
||||||
if ( ACTION_EMAIL !in n$actions )
|
if ( ACTION_EMAIL !in n$actions )
|
||||||
return;
|
return;
|
||||||
|
|
||||||
|
# I'm not recovering gracefully from the when statements because I want
|
||||||
|
# the notice framework to detect that something has exceeded the maximum
|
||||||
|
# allowed email delay and tell the user.
|
||||||
|
local uid = unique_id("");
|
||||||
|
tmp_notice_storage[uid] = n;
|
||||||
|
|
||||||
local output = "";
|
local output = "";
|
||||||
if ( n?$src )
|
if ( n?$src )
|
||||||
{
|
{
|
||||||
|
add n$email_delay_tokens["hostnames-src"];
|
||||||
when ( local src_name = lookup_addr(n$src) )
|
when ( local src_name = lookup_addr(n$src) )
|
||||||
{
|
{
|
||||||
output = string_cat("orig_h/src hostname: ", src_name, "\n");
|
output = string_cat("orig/src hostname: ", src_name, "\n");
|
||||||
n$email_body_sections[|n$email_body_sections|] = output;
|
tmp_notice_storage[uid]$email_body_sections[|tmp_notice_storage[uid]$email_body_sections|] = output;
|
||||||
|
delete tmp_notice_storage[uid]$email_delay_tokens["hostnames-src"];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if ( n?$dst )
|
if ( n?$dst )
|
||||||
{
|
{
|
||||||
|
add n$email_delay_tokens["hostnames-dst"];
|
||||||
when ( local dst_name = lookup_addr(n$dst) )
|
when ( local dst_name = lookup_addr(n$dst) )
|
||||||
{
|
{
|
||||||
output = string_cat("resp_h/dst hostname: ", dst_name, "\n");
|
output = string_cat("resp/dst hostname: ", dst_name, "\n");
|
||||||
n$email_body_sections[|n$email_body_sections|] = output;
|
tmp_notice_storage[uid]$email_body_sections[|tmp_notice_storage[uid]$email_body_sections|] = output;
|
||||||
|
delete tmp_notice_storage[uid]$email_delay_tokens["hostnames-dst"];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -74,7 +74,11 @@ export {
|
||||||
## the notice policy.
|
## the notice policy.
|
||||||
iconn: icmp_conn &optional;
|
iconn: icmp_conn &optional;
|
||||||
|
|
||||||
## The type of the notice.
|
## The transport protocol. Filled automatically when either conn, iconn
|
||||||
|
## or p is specified.
|
||||||
|
proto: transport_proto &log &optional;
|
||||||
|
|
||||||
|
## The :bro:type:`Notice::Type` of the notice.
|
||||||
note: Type &log;
|
note: Type &log;
|
||||||
## The human readable message for the notice.
|
## The human readable message for the notice.
|
||||||
msg: string &log &optional;
|
msg: string &log &optional;
|
||||||
|
@ -106,7 +110,13 @@ export {
|
||||||
## expand on notices that are being emailed. The normal way to add text
|
## expand on notices that are being emailed. The normal way to add text
|
||||||
## is to extend the vector by handling the :bro:id:`Notice::notice`
|
## is to extend the vector by handling the :bro:id:`Notice::notice`
|
||||||
## event and modifying the notice in place.
|
## event and modifying the notice in place.
|
||||||
email_body_sections: vector of string &default=vector();
|
email_body_sections: vector of string &optional;
|
||||||
|
|
||||||
|
## Adding a string "token" to this set will cause the notice framework's
|
||||||
|
## built-in emailing functionality to delay sending the email until
|
||||||
|
## either the token has been removed or the email has been delayed
|
||||||
|
## for :bro:id:`Notice::max_email_delay`.
|
||||||
|
email_delay_tokens: set[string] &optional;
|
||||||
|
|
||||||
## This field is to be provided when a notice is generated for the
|
## This field is to be provided when a notice is generated for the
|
||||||
## purpose of deduplicating notices. The identifier string should
|
## purpose of deduplicating notices. The identifier string should
|
||||||
|
@ -215,6 +225,8 @@ export {
|
||||||
const reply_to = "" &redef;
|
const reply_to = "" &redef;
|
||||||
## Text string prefixed to the subject of all emails sent out.
|
## Text string prefixed to the subject of all emails sent out.
|
||||||
const mail_subject_prefix = "[Bro]" &redef;
|
const mail_subject_prefix = "[Bro]" &redef;
|
||||||
|
## The maximum amount of time a plugin can delay email from being sent.
|
||||||
|
const max_email_delay = 15secs &redef;
|
||||||
|
|
||||||
## A log postprocessing function that implements emailing the contents
|
## A log postprocessing function that implements emailing the contents
|
||||||
## of a log upon rotation to any configured :bro:id:`Notice::mail_dest`.
|
## of a log upon rotation to any configured :bro:id:`Notice::mail_dest`.
|
||||||
|
@ -390,11 +402,35 @@ function email_headers(subject_desc: string, dest: string): string
|
||||||
return header_text;
|
return header_text;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
event delay_sending_email(n: Notice::Info, dest: string, extend: bool)
|
||||||
|
{
|
||||||
|
email_notice_to(n, dest, extend);
|
||||||
|
}
|
||||||
|
|
||||||
function email_notice_to(n: Notice::Info, dest: string, extend: bool)
|
function email_notice_to(n: Notice::Info, dest: string, extend: bool)
|
||||||
{
|
{
|
||||||
if ( reading_traces() || dest == "" )
|
if ( reading_traces() || dest == "" )
|
||||||
return;
|
return;
|
||||||
|
|
||||||
|
if ( extend )
|
||||||
|
{
|
||||||
|
if ( |n$email_delay_tokens| > 0 )
|
||||||
|
{
|
||||||
|
# If we still are within the max_email_delay, keep delaying.
|
||||||
|
if ( n$ts + max_email_delay > network_time() )
|
||||||
|
{
|
||||||
|
schedule 1sec { delay_sending_email(n, dest, extend) };
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
event reporter_info(network_time(),
|
||||||
|
fmt("Notice email delay tokens weren't released in time (%s).", n$email_delay_tokens),
|
||||||
|
"");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
local email_text = email_headers(fmt("%s", n$note), dest);
|
local email_text = email_headers(fmt("%s", n$note), dest);
|
||||||
|
|
||||||
# First off, finish the headers and include the human readable messages
|
# First off, finish the headers and include the human readable messages
|
||||||
|
@ -420,9 +456,10 @@ function email_notice_to(n: Notice::Info, dest: string, extend: bool)
|
||||||
# Add the extended information if it's requested.
|
# Add the extended information if it's requested.
|
||||||
if ( extend )
|
if ( extend )
|
||||||
{
|
{
|
||||||
|
email_text = string_cat(email_text, "\nEmail Extensions\n");
|
||||||
|
email_text = string_cat(email_text, "----------------\n");
|
||||||
for ( i in n$email_body_sections )
|
for ( i in n$email_body_sections )
|
||||||
{
|
{
|
||||||
email_text = string_cat(email_text, "******************\n");
|
|
||||||
email_text = string_cat(email_text, n$email_body_sections[i], "\n");
|
email_text = string_cat(email_text, n$email_body_sections[i], "\n");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -502,8 +539,12 @@ function apply_policy(n: Notice::Info)
|
||||||
n$p = n$id$resp_p;
|
n$p = n$id$resp_p;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if ( n?$p )
|
||||||
|
n$proto = get_port_transport_proto(n$p);
|
||||||
|
|
||||||
if ( n?$iconn )
|
if ( n?$iconn )
|
||||||
{
|
{
|
||||||
|
n$proto = icmp;
|
||||||
if ( ! n?$src )
|
if ( ! n?$src )
|
||||||
n$src = n$iconn$orig_h;
|
n$src = n$iconn$orig_h;
|
||||||
if ( ! n?$dst )
|
if ( ! n?$dst )
|
||||||
|
@ -519,6 +560,11 @@ function apply_policy(n: Notice::Info)
|
||||||
if ( ! n?$actions )
|
if ( ! n?$actions )
|
||||||
n$actions = set();
|
n$actions = set();
|
||||||
|
|
||||||
|
if ( ! n?$email_body_sections )
|
||||||
|
n$email_body_sections = vector();
|
||||||
|
if ( ! n?$email_delay_tokens )
|
||||||
|
n$email_delay_tokens = set();
|
||||||
|
|
||||||
if ( ! n?$policy_items )
|
if ( ! n?$policy_items )
|
||||||
n$policy_items = set();
|
n$policy_items = set();
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
##! This script provides the framework for software version detection and
|
##! This script provides the framework for software version detection and
|
||||||
##! parsing, but doesn't actually do any detection on it's own. It relys on
|
##! parsing but doesn't actually do any detection on it's own. It relys on
|
||||||
##! other protocol specific scripts to parse out software from the protocols
|
##! other protocol specific scripts to parse out software from the protocols
|
||||||
##! that they analyze. The entry point for providing new software detections
|
##! that they analyze. The entry point for providing new software detections
|
||||||
##! to this framework is through the :bro:id:`Software::found` function.
|
##! to this framework is through the :bro:id:`Software::found` function.
|
||||||
|
@ -10,39 +10,44 @@
|
||||||
module Software;
|
module Software;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The software logging stream identifier.
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
|
## Scripts detecting new types of software need to redef this enum to add
|
||||||
|
## their own specific software types which would then be used when they
|
||||||
|
## create :bro:type:`Software::Info` records.
|
||||||
type Type: enum {
|
type Type: enum {
|
||||||
|
## A placeholder type for when the type of software is not known.
|
||||||
UNKNOWN,
|
UNKNOWN,
|
||||||
OPERATING_SYSTEM,
|
|
||||||
DATABASE_SERVER,
|
|
||||||
# There are a number of ways to detect printers on the
|
|
||||||
# network, we just need to codify them in a script and move
|
|
||||||
# this out of here. It isn't currently used for anything.
|
|
||||||
PRINTER,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## A structure to represent the numeric version of software.
|
||||||
type Version: record {
|
type Version: record {
|
||||||
major: count &optional; ##< Major version number
|
## Major version number
|
||||||
minor: count &optional; ##< Minor version number
|
major: count &optional;
|
||||||
minor2: count &optional; ##< Minor subversion number
|
## Minor version number
|
||||||
addl: string &optional; ##< Additional version string (e.g. "beta42")
|
minor: count &optional;
|
||||||
|
## Minor subversion number
|
||||||
|
minor2: count &optional;
|
||||||
|
## Additional version string (e.g. "beta42")
|
||||||
|
addl: string &optional;
|
||||||
} &log;
|
} &log;
|
||||||
|
|
||||||
|
## The record type that is used for representing and logging software.
|
||||||
type Info: record {
|
type Info: record {
|
||||||
## The time at which the software was first detected.
|
## The time at which the software was detected.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
## The IP address detected running the software.
|
## The IP address detected running the software.
|
||||||
host: addr &log;
|
host: addr &log;
|
||||||
## The type of software detected (e.g. WEB_SERVER)
|
## The type of software detected (e.g. :bro:enum:`HTTP::SERVER`).
|
||||||
software_type: Type &log &default=UNKNOWN;
|
software_type: Type &log &default=UNKNOWN;
|
||||||
## Name of the software (e.g. Apache)
|
## Name of the software (e.g. Apache).
|
||||||
name: string &log;
|
name: string &log;
|
||||||
## Version of the software
|
## Version of the software.
|
||||||
version: Version &log;
|
version: Version &log;
|
||||||
## The full unparsed version string found because the version parsing
|
## The full unparsed version string found because the version parsing
|
||||||
## doesn't work 100% reliably and this acts as a fall back in the logs.
|
## doesn't always work reliably in all cases and this acts as a
|
||||||
|
## fallback in the logs.
|
||||||
unparsed_version: string &log &optional;
|
unparsed_version: string &log &optional;
|
||||||
|
|
||||||
## This can indicate that this software being detected should
|
## This can indicate that this software being detected should
|
||||||
|
@ -55,37 +60,48 @@ export {
|
||||||
force_log: bool &default=F;
|
force_log: bool &default=F;
|
||||||
};
|
};
|
||||||
|
|
||||||
## The hosts whose software should be detected and tracked.
|
## Hosts whose software should be detected and tracked.
|
||||||
## Choices are: LOCAL_HOSTS, REMOTE_HOSTS, ALL_HOSTS, NO_HOSTS
|
## Choices are: LOCAL_HOSTS, REMOTE_HOSTS, ALL_HOSTS, NO_HOSTS
|
||||||
const asset_tracking = LOCAL_HOSTS &redef;
|
const asset_tracking = LOCAL_HOSTS &redef;
|
||||||
|
|
||||||
|
|
||||||
## Other scripts should call this function when they detect software.
|
## Other scripts should call this function when they detect software.
|
||||||
## unparsed_version: This is the full string from which the
|
## unparsed_version: This is the full string from which the
|
||||||
## :bro:type:`Software::Info` was extracted.
|
## :bro:type:`Software::Info` was extracted.
|
||||||
|
##
|
||||||
|
## id: The connection id where the software was discovered.
|
||||||
|
##
|
||||||
|
## info: A record representing the software discovered.
|
||||||
|
##
|
||||||
## Returns: T if the software was logged, F otherwise.
|
## Returns: T if the software was logged, F otherwise.
|
||||||
global found: function(id: conn_id, info: Software::Info): bool;
|
global found: function(id: conn_id, info: Software::Info): bool;
|
||||||
|
|
||||||
## This function can take many software version strings and parse them
|
## Take many common software version strings and parse them
|
||||||
## into a sensible :bro:type:`Software::Version` record. There are
|
## into a sensible :bro:type:`Software::Version` record. There are
|
||||||
## still many cases where scripts may have to have their own specific
|
## still many cases where scripts may have to have their own specific
|
||||||
## version parsing though.
|
## version parsing though.
|
||||||
|
##
|
||||||
|
## unparsed_version: The raw version string.
|
||||||
|
##
|
||||||
|
## host: The host where the software was discovered.
|
||||||
|
##
|
||||||
|
## software_type: The type of software.
|
||||||
|
##
|
||||||
|
## Returns: A complete record ready for the :bro:id:`Software::found` function.
|
||||||
global parse: function(unparsed_version: string,
|
global parse: function(unparsed_version: string,
|
||||||
host: addr,
|
host: addr,
|
||||||
software_type: Type): Info;
|
software_type: Type): Info;
|
||||||
|
|
||||||
## Compare two versions.
|
## Compare two version records.
|
||||||
|
##
|
||||||
## Returns: -1 for v1 < v2, 0 for v1 == v2, 1 for v1 > v2.
|
## Returns: -1 for v1 < v2, 0 for v1 == v2, 1 for v1 > v2.
|
||||||
## If the numerical version numbers match, the addl string
|
## If the numerical version numbers match, the addl string
|
||||||
## is compared lexicographically.
|
## is compared lexicographically.
|
||||||
global cmp_versions: function(v1: Version, v2: Version): int;
|
global cmp_versions: function(v1: Version, v2: Version): int;
|
||||||
|
|
||||||
## This type represents a set of software. It's used by the
|
## Type to represent a collection of :bro:type:`Software::Info` records.
|
||||||
## :bro:id:`tracked` variable to store all known pieces of software
|
## It's indexed with the name of a piece of software such as "Firefox"
|
||||||
## for a particular host. It's indexed with the name of a piece of
|
## and it yields a :bro:type:`Software::Info` record with more information
|
||||||
## software such as "Firefox" and it yields a
|
## about the software.
|
||||||
## :bro:type:`Software::Info` record with more information about the
|
|
||||||
## software.
|
|
||||||
type SoftwareSet: table[string] of Info;
|
type SoftwareSet: table[string] of Info;
|
||||||
|
|
||||||
## The set of software associated with an address. Data expires from
|
## The set of software associated with an address. Data expires from
|
||||||
|
|
File diff suppressed because it is too large
Load diff
|
@ -1,10 +1,13 @@
|
||||||
##! This script can be used to extract either the originator's data or the
|
##! This script can be used to extract either the originator's data or the
|
||||||
##! responders data or both. By default nothing is extracted, and in order
|
##! responders data or both. By default nothing is extracted, and in order
|
||||||
##! to actually extract data the ``c$extract_orig`` and/or the
|
##! to actually extract data the ``c$extract_orig`` and/or the
|
||||||
##! ``c$extract_resp`` variable must be set to T. One way to achieve this
|
##! ``c$extract_resp`` variable must be set to ``T``. One way to achieve this
|
||||||
##! would be to handle the connection_established event elsewhere and set the
|
##! would be to handle the :bro:id:`connection_established` event elsewhere
|
||||||
##! extract_orig and extract_resp options there. However, there may be trouble
|
##! and set the ``extract_orig`` and ``extract_resp`` options there.
|
||||||
##! with the timing due the event queue delay.
|
##! However, there may be trouble with the timing due to event queue delay.
|
||||||
|
##!
|
||||||
|
##! .. note::
|
||||||
|
##!
|
||||||
##! This script does not work well in a cluster context unless it has a
|
##! This script does not work well in a cluster context unless it has a
|
||||||
##! remotely mounted disk to write the content files to.
|
##! remotely mounted disk to write the content files to.
|
||||||
|
|
||||||
|
@ -13,11 +16,12 @@
|
||||||
module Conn;
|
module Conn;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
## The prefix given to files as they are opened on disk.
|
## The prefix given to files containing extracted connections as they are
|
||||||
|
## opened on disk.
|
||||||
const extraction_prefix = "contents" &redef;
|
const extraction_prefix = "contents" &redef;
|
||||||
|
|
||||||
## If this variable is set to T, then all contents of all files will be
|
## If this variable is set to ``T``, then all contents of all connections
|
||||||
## extracted.
|
## will be extracted.
|
||||||
const default_extract = F &redef;
|
const default_extract = F &redef;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -4,7 +4,7 @@
|
||||||
module Conn;
|
module Conn;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
## Define inactivty timeouts by the service detected being used over
|
## Define inactivity timeouts by the service detected being used over
|
||||||
## the connection.
|
## the connection.
|
||||||
const analyzer_inactivity_timeouts: table[AnalyzerTag] of interval = {
|
const analyzer_inactivity_timeouts: table[AnalyzerTag] of interval = {
|
||||||
# For interactive services, allow longer periods of inactivity.
|
# For interactive services, allow longer periods of inactivity.
|
||||||
|
|
|
@ -1,17 +1,33 @@
|
||||||
|
##! This script manages the tracking/logging of general information regarding
|
||||||
|
##! TCP, UDP, and ICMP traffic. For UDP and ICMP, "connections" are to
|
||||||
|
##! be interpreted using flow semantics (sequence of packets from a source
|
||||||
|
##! host/post to a destination host/port). Further, ICMP "ports" are to
|
||||||
|
##! be interpreted as the source port meaning the ICMP message type and
|
||||||
|
##! the destination port being the ICMP message code.
|
||||||
|
|
||||||
@load base/utils/site
|
@load base/utils/site
|
||||||
|
|
||||||
module Conn;
|
module Conn;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The connection logging stream identifier.
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
|
## The record type which contains column fields of the connection log.
|
||||||
type Info: record {
|
type Info: record {
|
||||||
## This is the time of the first packet.
|
## This is the time of the first packet.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
|
## A unique identifier of a connection.
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
|
## The connection's 4-tuple of endpoint addresses/ports.
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
|
## The transport layer protocol of the connection.
|
||||||
proto: transport_proto &log;
|
proto: transport_proto &log;
|
||||||
|
## An identification of an application protocol being sent over the
|
||||||
|
## the connection.
|
||||||
service: string &log &optional;
|
service: string &log &optional;
|
||||||
|
## How long the connection lasted. For 3-way or 4-way connection
|
||||||
|
## tear-downs, this will not include the final ACK.
|
||||||
duration: interval &log &optional;
|
duration: interval &log &optional;
|
||||||
## The number of payload bytes the originator sent. For TCP
|
## The number of payload bytes the originator sent. For TCP
|
||||||
## this is taken from sequence numbers and might be inaccurate
|
## this is taken from sequence numbers and might be inaccurate
|
||||||
|
@ -51,8 +67,8 @@ export {
|
||||||
## have been completed prior to the packet loss.
|
## have been completed prior to the packet loss.
|
||||||
missed_bytes: count &log &default=0;
|
missed_bytes: count &log &default=0;
|
||||||
|
|
||||||
## Records the state history of (TCP) connections as
|
## Records the state history of connections as a string of letters.
|
||||||
## a string of letters.
|
## For TCP connections the meaning of those letters is:
|
||||||
##
|
##
|
||||||
## ====== ====================================================
|
## ====== ====================================================
|
||||||
## Letter Meaning
|
## Letter Meaning
|
||||||
|
@ -71,7 +87,8 @@ export {
|
||||||
## originator and lower case then means the responder.
|
## originator and lower case then means the responder.
|
||||||
## Also, there is compression. We only record one "d" in each direction,
|
## Also, there is compression. We only record one "d" in each direction,
|
||||||
## for instance. I.e., we just record that data went in that direction.
|
## for instance. I.e., we just record that data went in that direction.
|
||||||
## This history is not meant to encode how much data that happened to be.
|
## This history is not meant to encode how much data that happened to
|
||||||
|
## be.
|
||||||
history: string &log &optional;
|
history: string &log &optional;
|
||||||
## Number of packets the originator sent.
|
## Number of packets the originator sent.
|
||||||
## Only set if :bro:id:`use_conn_size_analyzer` = T
|
## Only set if :bro:id:`use_conn_size_analyzer` = T
|
||||||
|
@ -86,6 +103,8 @@ export {
|
||||||
resp_ip_bytes: count &log &optional;
|
resp_ip_bytes: count &log &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Event that can be handled to access the :bro:type:`Conn::Info`
|
||||||
|
## record as it is sent on to the logging framework.
|
||||||
global log_conn: event(rec: Info);
|
global log_conn: event(rec: Info);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -4,9 +4,9 @@
|
||||||
module DNS;
|
module DNS;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
const PTR = 12;
|
const PTR = 12; ##< RR TYPE value for a domain name pointer.
|
||||||
const EDNS = 41;
|
const EDNS = 41; ##< An OPT RR TYPE value described by EDNS.
|
||||||
const ANY = 255;
|
const ANY = 255; ##< A QTYPE value describing a request for all records.
|
||||||
|
|
||||||
## Mapping of DNS query type codes to human readable string representation.
|
## Mapping of DNS query type codes to human readable string representation.
|
||||||
const query_types = {
|
const query_types = {
|
||||||
|
@ -29,50 +29,43 @@ export {
|
||||||
[ANY] = "*",
|
[ANY] = "*",
|
||||||
} &default = function(n: count): string { return fmt("query-%d", n); };
|
} &default = function(n: count): string { return fmt("query-%d", n); };
|
||||||
|
|
||||||
const code_types = {
|
|
||||||
[0] = "X0",
|
|
||||||
[1] = "Xfmt",
|
|
||||||
[2] = "Xsrv",
|
|
||||||
[3] = "Xnam",
|
|
||||||
[4] = "Ximp",
|
|
||||||
[5] = "X[",
|
|
||||||
} &default="?";
|
|
||||||
|
|
||||||
## Errors used for non-TSIG/EDNS types.
|
## Errors used for non-TSIG/EDNS types.
|
||||||
const base_errors = {
|
const base_errors = {
|
||||||
[0] = "NOERROR", ##< No Error
|
[0] = "NOERROR", # No Error
|
||||||
[1] = "FORMERR", ##< Format Error
|
[1] = "FORMERR", # Format Error
|
||||||
[2] = "SERVFAIL", ##< Server Failure
|
[2] = "SERVFAIL", # Server Failure
|
||||||
[3] = "NXDOMAIN", ##< Non-Existent Domain
|
[3] = "NXDOMAIN", # Non-Existent Domain
|
||||||
[4] = "NOTIMP", ##< Not Implemented
|
[4] = "NOTIMP", # Not Implemented
|
||||||
[5] = "REFUSED", ##< Query Refused
|
[5] = "REFUSED", # Query Refused
|
||||||
[6] = "YXDOMAIN", ##< Name Exists when it should not
|
[6] = "YXDOMAIN", # Name Exists when it should not
|
||||||
[7] = "YXRRSET", ##< RR Set Exists when it should not
|
[7] = "YXRRSET", # RR Set Exists when it should not
|
||||||
[8] = "NXRRSet", ##< RR Set that should exist does not
|
[8] = "NXRRSet", # RR Set that should exist does not
|
||||||
[9] = "NOTAUTH", ##< Server Not Authoritative for zone
|
[9] = "NOTAUTH", # Server Not Authoritative for zone
|
||||||
[10] = "NOTZONE", ##< Name not contained in zone
|
[10] = "NOTZONE", # Name not contained in zone
|
||||||
[11] = "unassigned-11", ##< available for assignment
|
[11] = "unassigned-11", # available for assignment
|
||||||
[12] = "unassigned-12", ##< available for assignment
|
[12] = "unassigned-12", # available for assignment
|
||||||
[13] = "unassigned-13", ##< available for assignment
|
[13] = "unassigned-13", # available for assignment
|
||||||
[14] = "unassigned-14", ##< available for assignment
|
[14] = "unassigned-14", # available for assignment
|
||||||
[15] = "unassigned-15", ##< available for assignment
|
[15] = "unassigned-15", # available for assignment
|
||||||
[16] = "BADVERS", ##< for EDNS, collision w/ TSIG
|
[16] = "BADVERS", # for EDNS, collision w/ TSIG
|
||||||
[17] = "BADKEY", ##< Key not recognized
|
[17] = "BADKEY", # Key not recognized
|
||||||
[18] = "BADTIME", ##< Signature out of time window
|
[18] = "BADTIME", # Signature out of time window
|
||||||
[19] = "BADMODE", ##< Bad TKEY Mode
|
[19] = "BADMODE", # Bad TKEY Mode
|
||||||
[20] = "BADNAME", ##< Duplicate key name
|
[20] = "BADNAME", # Duplicate key name
|
||||||
[21] = "BADALG", ##< Algorithm not supported
|
[21] = "BADALG", # Algorithm not supported
|
||||||
[22] = "BADTRUNC", ##< draft-ietf-dnsext-tsig-sha-05.txt
|
[22] = "BADTRUNC", # draft-ietf-dnsext-tsig-sha-05.txt
|
||||||
[3842] = "BADSIG", ##< 16 <= number collision with EDNS(16);
|
[3842] = "BADSIG", # 16 <= number collision with EDNS(16);
|
||||||
##< this is a translation from TSIG(16)
|
# this is a translation from TSIG(16)
|
||||||
} &default = function(n: count): string { return fmt("rcode-%d", n); };
|
} &default = function(n: count): string { return fmt("rcode-%d", n); };
|
||||||
|
|
||||||
# This deciphers EDNS Z field values.
|
## This deciphers EDNS Z field values.
|
||||||
const edns_zfield = {
|
const edns_zfield = {
|
||||||
[0] = "NOVALUE", # regular entry
|
[0] = "NOVALUE", # regular entry
|
||||||
[32768] = "DNS_SEC_OK", # accepts DNS Sec RRs
|
[32768] = "DNS_SEC_OK", # accepts DNS Sec RRs
|
||||||
} &default="?";
|
} &default="?";
|
||||||
|
|
||||||
|
## Possible values of the CLASS field in resource records or QCLASS field
|
||||||
|
## in query messages.
|
||||||
const classes = {
|
const classes = {
|
||||||
[1] = "C_INTERNET",
|
[1] = "C_INTERNET",
|
||||||
[2] = "C_CSNET",
|
[2] = "C_CSNET",
|
||||||
|
|
|
@ -1,38 +1,80 @@
|
||||||
|
##! Base DNS analysis script which tracks and logs DNS queries along with
|
||||||
|
##! their responses.
|
||||||
|
|
||||||
@load ./consts
|
@load ./consts
|
||||||
|
|
||||||
module DNS;
|
module DNS;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The DNS logging stream identifier.
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
|
## The record type which contains the column fields of the DNS log.
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## The earliest time at which a DNS protocol message over the
|
||||||
|
## associated connection is observed.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
|
## A unique identifier of the connection over which DNS messages
|
||||||
|
## are being transferred.
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
|
## The connection's 4-tuple of endpoint addresses/ports.
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
|
## The transport layer protocol of the connection.
|
||||||
proto: transport_proto &log;
|
proto: transport_proto &log;
|
||||||
|
## A 16 bit identifier assigned by the program that generated the
|
||||||
|
## DNS query. Also used in responses to match up replies to
|
||||||
|
## outstanding queries.
|
||||||
trans_id: count &log &optional;
|
trans_id: count &log &optional;
|
||||||
|
## The domain name that is the subject of the DNS query.
|
||||||
query: string &log &optional;
|
query: string &log &optional;
|
||||||
|
## The QCLASS value specifying the class of the query.
|
||||||
qclass: count &log &optional;
|
qclass: count &log &optional;
|
||||||
|
## A descriptive name for the class of the query.
|
||||||
qclass_name: string &log &optional;
|
qclass_name: string &log &optional;
|
||||||
|
## A QTYPE value specifying the type of the query.
|
||||||
qtype: count &log &optional;
|
qtype: count &log &optional;
|
||||||
|
## A descriptive name for the type of the query.
|
||||||
qtype_name: string &log &optional;
|
qtype_name: string &log &optional;
|
||||||
|
## The response code value in DNS response messages.
|
||||||
rcode: count &log &optional;
|
rcode: count &log &optional;
|
||||||
|
## A descriptive name for the response code value.
|
||||||
rcode_name: string &log &optional;
|
rcode_name: string &log &optional;
|
||||||
|
## Whether the message is a query (F) or response (T).
|
||||||
QR: bool &log &default=F;
|
QR: bool &log &default=F;
|
||||||
|
## The Authoritative Answer bit for response messages specifies that
|
||||||
|
## the responding name server is an authority for the domain name
|
||||||
|
## in the question section.
|
||||||
AA: bool &log &default=F;
|
AA: bool &log &default=F;
|
||||||
|
## The Truncation bit specifies that the message was truncated.
|
||||||
TC: bool &log &default=F;
|
TC: bool &log &default=F;
|
||||||
|
## The Recursion Desired bit indicates to a name server to recursively
|
||||||
|
## purse the query.
|
||||||
RD: bool &log &default=F;
|
RD: bool &log &default=F;
|
||||||
|
## The Recursion Available bit in a response message indicates if
|
||||||
|
## the name server supports recursive queries.
|
||||||
RA: bool &log &default=F;
|
RA: bool &log &default=F;
|
||||||
|
## A reserved field that is currently supposed to be zero in all
|
||||||
|
## queries and responses.
|
||||||
Z: count &log &default=0;
|
Z: count &log &default=0;
|
||||||
TTL: interval &log &optional;
|
## The set of resource descriptions in answer of the query.
|
||||||
answers: set[string] &log &optional;
|
answers: vector of string &log &optional;
|
||||||
|
## The caching intervals of the associated RRs described by the
|
||||||
|
## ``answers`` field.
|
||||||
|
TTLs: vector of interval &log &optional;
|
||||||
|
|
||||||
## This value indicates if this request/response pair is ready to be logged.
|
## This value indicates if this request/response pair is ready to be
|
||||||
|
## logged.
|
||||||
ready: bool &default=F;
|
ready: bool &default=F;
|
||||||
|
## The total number of resource records in a reply message's answer
|
||||||
|
## section.
|
||||||
total_answers: count &optional;
|
total_answers: count &optional;
|
||||||
|
## The total number of resource records in a reply message's answer,
|
||||||
|
## authority, and additional sections.
|
||||||
total_replies: count &optional;
|
total_replies: count &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## A record type which tracks the status of DNS queries for a given
|
||||||
|
## :bro:type:`connection`.
|
||||||
type State: record {
|
type State: record {
|
||||||
## Indexed by query id, returns Info record corresponding to
|
## Indexed by query id, returns Info record corresponding to
|
||||||
## query/response which haven't completed yet.
|
## query/response which haven't completed yet.
|
||||||
|
@ -44,11 +86,21 @@ export {
|
||||||
finished_answers: set[count] &optional;
|
finished_answers: set[count] &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## An event that can be handled to access the :bro:type:`DNS::Info`
|
||||||
|
## record as it is sent to the logging framework.
|
||||||
global log_dns: event(rec: Info);
|
global log_dns: event(rec: Info);
|
||||||
|
|
||||||
## This is called by the specific dns_*_reply events with a "reply" which
|
## This is called by the specific dns_*_reply events with a "reply" which
|
||||||
## may not represent the full data available from the resource record, but
|
## may not represent the full data available from the resource record, but
|
||||||
## it's generally considered a summarization of the response(s).
|
## it's generally considered a summarization of the response(s).
|
||||||
|
##
|
||||||
|
## c: The connection record for which to fill in DNS reply data.
|
||||||
|
##
|
||||||
|
## msg: The DNS message header information for the response.
|
||||||
|
##
|
||||||
|
## ans: The general information of a RR response.
|
||||||
|
##
|
||||||
|
## reply: The specific response information according to RR type/class.
|
||||||
global do_reply: event(c: connection, msg: dns_msg, ans: dns_answer, reply: string);
|
global do_reply: event(c: connection, msg: dns_msg, ans: dns_answer, reply: string);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -102,7 +154,13 @@ function new_session(c: connection, trans_id: count): Info
|
||||||
function set_session(c: connection, msg: dns_msg, is_query: bool)
|
function set_session(c: connection, msg: dns_msg, is_query: bool)
|
||||||
{
|
{
|
||||||
if ( ! c?$dns_state || msg$id !in c$dns_state$pending )
|
if ( ! c?$dns_state || msg$id !in c$dns_state$pending )
|
||||||
|
{
|
||||||
c$dns_state$pending[msg$id] = new_session(c, msg$id);
|
c$dns_state$pending[msg$id] = new_session(c, msg$id);
|
||||||
|
# Try deleting this transaction id from the set of finished answers.
|
||||||
|
# Sometimes hosts will reuse ports and transaction ids and this should
|
||||||
|
# be considered to be a legit scenario (although bad practice).
|
||||||
|
delete c$dns_state$finished_answers[msg$id];
|
||||||
|
}
|
||||||
|
|
||||||
c$dns = c$dns_state$pending[msg$id];
|
c$dns = c$dns_state$pending[msg$id];
|
||||||
|
|
||||||
|
@ -134,20 +192,23 @@ event DNS::do_reply(c: connection, msg: dns_msg, ans: dns_answer, reply: string)
|
||||||
{
|
{
|
||||||
set_session(c, msg, F);
|
set_session(c, msg, F);
|
||||||
|
|
||||||
c$dns$AA = msg$AA;
|
|
||||||
c$dns$RA = msg$RA;
|
|
||||||
c$dns$TTL = ans$TTL;
|
|
||||||
|
|
||||||
if ( ans$answer_type == DNS_ANS )
|
if ( ans$answer_type == DNS_ANS )
|
||||||
{
|
{
|
||||||
|
c$dns$AA = msg$AA;
|
||||||
|
c$dns$RA = msg$RA;
|
||||||
|
|
||||||
if ( msg$id in c$dns_state$finished_answers )
|
if ( msg$id in c$dns_state$finished_answers )
|
||||||
event conn_weird("dns_reply_seen_after_done", c, "");
|
event conn_weird("dns_reply_seen_after_done", c, "");
|
||||||
|
|
||||||
if ( reply != "" )
|
if ( reply != "" )
|
||||||
{
|
{
|
||||||
if ( ! c$dns?$answers )
|
if ( ! c$dns?$answers )
|
||||||
c$dns$answers = set();
|
c$dns$answers = vector();
|
||||||
add c$dns$answers[reply];
|
c$dns$answers[|c$dns$answers|] = reply;
|
||||||
|
|
||||||
|
if ( ! c$dns?$TTLs )
|
||||||
|
c$dns$TTLs = vector();
|
||||||
|
c$dns$TTLs[|c$dns$TTLs|] = ans$TTL;
|
||||||
}
|
}
|
||||||
|
|
||||||
if ( c$dns?$answers && |c$dns$answers| == c$dns$total_answers )
|
if ( c$dns?$answers && |c$dns$answers| == c$dns$total_answers )
|
||||||
|
@ -164,7 +225,6 @@ event DNS::do_reply(c: connection, msg: dns_msg, ans: dns_answer, reply: string)
|
||||||
if ( c$dns$ready )
|
if ( c$dns$ready )
|
||||||
{
|
{
|
||||||
Log::write(DNS::LOG, c$dns);
|
Log::write(DNS::LOG, c$dns);
|
||||||
add c$dns_state$finished_answers[c$dns$trans_id];
|
|
||||||
# This record is logged and no longer pending.
|
# This record is logged and no longer pending.
|
||||||
delete c$dns_state$pending[c$dns$trans_id];
|
delete c$dns_state$pending[c$dns$trans_id];
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
##! File extraction for FTP.
|
##! File extraction support for FTP.
|
||||||
|
|
||||||
@load ./main
|
@load ./main
|
||||||
@load base/utils/files
|
@load base/utils/files
|
||||||
|
@ -6,7 +6,7 @@
|
||||||
module FTP;
|
module FTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
## Pattern of file mime types to extract from FTP entity bodies.
|
## Pattern of file mime types to extract from FTP transfers.
|
||||||
const extract_file_types = /NO_DEFAULT/ &redef;
|
const extract_file_types = /NO_DEFAULT/ &redef;
|
||||||
|
|
||||||
## The on-disk prefix for files to be extracted from FTP-data transfers.
|
## The on-disk prefix for files to be extracted from FTP-data transfers.
|
||||||
|
@ -14,10 +14,15 @@ export {
|
||||||
}
|
}
|
||||||
|
|
||||||
redef record Info += {
|
redef record Info += {
|
||||||
## The file handle for the file to be extracted
|
## On disk file where it was extracted to.
|
||||||
extraction_file: file &log &optional;
|
extraction_file: file &log &optional;
|
||||||
|
|
||||||
|
## Indicates if the current command/response pair should attempt to
|
||||||
|
## extract the file if a file was transferred.
|
||||||
extract_file: bool &default=F;
|
extract_file: bool &default=F;
|
||||||
|
|
||||||
|
## Internal tracking of the total number of files extracted during this
|
||||||
|
## session.
|
||||||
num_extracted_files: count &default=0;
|
num_extracted_files: count &default=0;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -33,7 +38,6 @@ event file_transferred(c: connection, prefix: string, descr: string,
|
||||||
if ( extract_file_types in s$mime_type )
|
if ( extract_file_types in s$mime_type )
|
||||||
{
|
{
|
||||||
s$extract_file = T;
|
s$extract_file = T;
|
||||||
add s$tags["extracted_file"];
|
|
||||||
++s$num_extracted_files;
|
++s$num_extracted_files;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,10 +2,6 @@
|
||||||
##! along with metadata. For example, if files are transferred, the argument
|
##! along with metadata. For example, if files are transferred, the argument
|
||||||
##! will take on the full path that the client is at along with the requested
|
##! will take on the full path that the client is at along with the requested
|
||||||
##! file name.
|
##! file name.
|
||||||
##!
|
|
||||||
##! TODO:
|
|
||||||
##!
|
|
||||||
##! * Handle encrypted sessions correctly (get an example?)
|
|
||||||
|
|
||||||
@load ./utils-commands
|
@load ./utils-commands
|
||||||
@load base/utils/paths
|
@load base/utils/paths
|
||||||
|
@ -14,38 +10,64 @@
|
||||||
module FTP;
|
module FTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The FTP protocol logging stream identifier.
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
|
## List of commands that should have their command/response pairs logged.
|
||||||
|
const logged_commands = {
|
||||||
|
"APPE", "DELE", "RETR", "STOR", "STOU", "ACCT"
|
||||||
|
} &redef;
|
||||||
|
|
||||||
## This setting changes if passwords used in FTP sessions are captured or not.
|
## This setting changes if passwords used in FTP sessions are captured or not.
|
||||||
const default_capture_password = F &redef;
|
const default_capture_password = F &redef;
|
||||||
|
|
||||||
|
## User IDs that can be considered "anonymous".
|
||||||
|
const guest_ids = { "anonymous", "ftp", "guest" } &redef;
|
||||||
|
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## Time when the command was sent.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
|
## User name for the current FTP session.
|
||||||
user: string &log &default="<unknown>";
|
user: string &log &default="<unknown>";
|
||||||
|
## Password for the current FTP session if captured.
|
||||||
password: string &log &optional;
|
password: string &log &optional;
|
||||||
|
## Command given by the client.
|
||||||
command: string &log &optional;
|
command: string &log &optional;
|
||||||
|
## Argument for the command if one is given.
|
||||||
arg: string &log &optional;
|
arg: string &log &optional;
|
||||||
|
|
||||||
|
## Libmagic "sniffed" file type if the command indicates a file transfer.
|
||||||
mime_type: string &log &optional;
|
mime_type: string &log &optional;
|
||||||
|
## Libmagic "sniffed" file description if the command indicates a file transfer.
|
||||||
mime_desc: string &log &optional;
|
mime_desc: string &log &optional;
|
||||||
|
## Size of the file if the command indicates a file transfer.
|
||||||
file_size: count &log &optional;
|
file_size: count &log &optional;
|
||||||
|
|
||||||
|
## Reply code from the server in response to the command.
|
||||||
reply_code: count &log &optional;
|
reply_code: count &log &optional;
|
||||||
|
## Reply message from the server in response to the command.
|
||||||
reply_msg: string &log &optional;
|
reply_msg: string &log &optional;
|
||||||
|
## Arbitrary tags that may indicate a particular attribute of this command.
|
||||||
tags: set[string] &log &default=set();
|
tags: set[string] &log &default=set();
|
||||||
|
|
||||||
## By setting the CWD to '/.', we can indicate that unless something
|
## Current working directory that this session is in. By making
|
||||||
|
## the default value '/.', we can indicate that unless something
|
||||||
## more concrete is discovered that the existing but unknown
|
## more concrete is discovered that the existing but unknown
|
||||||
## directory is ok to use.
|
## directory is ok to use.
|
||||||
cwd: string &default="/.";
|
cwd: string &default="/.";
|
||||||
|
|
||||||
|
## Command that is currently waiting for a response.
|
||||||
cmdarg: CmdArg &optional;
|
cmdarg: CmdArg &optional;
|
||||||
|
## Queue for commands that have been sent but not yet responded to
|
||||||
|
## are tracked here.
|
||||||
pending_commands: PendingCmds;
|
pending_commands: PendingCmds;
|
||||||
|
|
||||||
## This indicates if the session is in active or passive mode.
|
## Indicates if the session is in active or passive mode.
|
||||||
passive: bool &default=F;
|
passive: bool &default=F;
|
||||||
|
|
||||||
## This determines if the password will be captured for this request.
|
## Determines if the password will be captured for this request.
|
||||||
capture_password: bool &default=default_capture_password;
|
capture_password: bool &default=default_capture_password;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -57,21 +79,11 @@ export {
|
||||||
z: count;
|
z: count;
|
||||||
};
|
};
|
||||||
|
|
||||||
# TODO: add this back in some form. raise a notice again?
|
## Parse FTP reply codes into the three constituent single digit values.
|
||||||
#const excessive_filename_len = 250 &redef;
|
|
||||||
#const excessive_filename_trunc_len = 32 &redef;
|
|
||||||
|
|
||||||
## These are user IDs that can be considered "anonymous".
|
|
||||||
const guest_ids = { "anonymous", "ftp", "guest" } &redef;
|
|
||||||
|
|
||||||
## The list of commands that should have their command/response pairs logged.
|
|
||||||
const logged_commands = {
|
|
||||||
"APPE", "DELE", "RETR", "STOR", "STOU", "ACCT"
|
|
||||||
} &redef;
|
|
||||||
|
|
||||||
## This function splits FTP reply codes into the three constituent
|
|
||||||
global parse_ftp_reply_code: function(code: count): ReplyCode;
|
global parse_ftp_reply_code: function(code: count): ReplyCode;
|
||||||
|
|
||||||
|
## Event that can be handled to access the :bro:type:`FTP::Info`
|
||||||
|
## record as it is sent on to the logging framework.
|
||||||
global log_ftp: event(rec: Info);
|
global log_ftp: event(rec: Info);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -2,14 +2,22 @@ module FTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
type CmdArg: record {
|
type CmdArg: record {
|
||||||
|
## Time when the command was sent.
|
||||||
ts: time;
|
ts: time;
|
||||||
|
## Command.
|
||||||
cmd: string &default="<unknown>";
|
cmd: string &default="<unknown>";
|
||||||
|
## Argument for the command if one was given.
|
||||||
arg: string &default="";
|
arg: string &default="";
|
||||||
|
## Counter to track how many commands have been executed.
|
||||||
seq: count &default=0;
|
seq: count &default=0;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Structure for tracking pending commands in the event that the client
|
||||||
|
## sends a large number of commands before the server has a chance to
|
||||||
|
## reply.
|
||||||
type PendingCmds: table[count] of CmdArg;
|
type PendingCmds: table[count] of CmdArg;
|
||||||
|
|
||||||
|
## Possible response codes for a wide variety of FTP commands.
|
||||||
const cmd_reply_code: set[string, count] = {
|
const cmd_reply_code: set[string, count] = {
|
||||||
# According to RFC 959
|
# According to RFC 959
|
||||||
["<init>", [120, 220, 421]],
|
["<init>", [120, 220, 421]],
|
||||||
|
|
|
@ -8,29 +8,24 @@
|
||||||
module HTTP;
|
module HTTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
## Pattern of file mime types to extract from HTTP entity bodies.
|
## Pattern of file mime types to extract from HTTP response entity bodies.
|
||||||
const extract_file_types = /NO_DEFAULT/ &redef;
|
const extract_file_types = /NO_DEFAULT/ &redef;
|
||||||
|
|
||||||
## The on-disk prefix for files to be extracted from HTTP entity bodies.
|
## The on-disk prefix for files to be extracted from HTTP entity bodies.
|
||||||
const extraction_prefix = "http-item" &redef;
|
const extraction_prefix = "http-item" &redef;
|
||||||
|
|
||||||
redef record Info += {
|
redef record Info += {
|
||||||
## This field can be set per-connection to determine if the entity body
|
## On-disk file where the response body was extracted to.
|
||||||
## will be extracted. It must be set to T on or before the first
|
|
||||||
## entity_body_data event.
|
|
||||||
extracting_file: bool &default=F;
|
|
||||||
|
|
||||||
## This is the holder for the file handle as the file is being written
|
|
||||||
## to disk.
|
|
||||||
extraction_file: file &log &optional;
|
extraction_file: file &log &optional;
|
||||||
};
|
|
||||||
|
|
||||||
redef record State += {
|
## Indicates if the response body is to be extracted or not. Must be
|
||||||
entity_bodies: count &default=0;
|
## set before or by the first :bro:id:`http_entity_data` event for the
|
||||||
|
## content.
|
||||||
|
extract_file: bool &default=F;
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
event http_entity_data(c: connection, is_orig: bool, length: count, data: string) &priority=5
|
event http_entity_data(c: connection, is_orig: bool, length: count, data: string) &priority=-5
|
||||||
{
|
{
|
||||||
# Client body extraction is not currently supported in this script.
|
# Client body extraction is not currently supported in this script.
|
||||||
if ( is_orig )
|
if ( is_orig )
|
||||||
|
@ -41,8 +36,12 @@ event http_entity_data(c: connection, is_orig: bool, length: count, data: string
|
||||||
if ( c$http?$mime_type &&
|
if ( c$http?$mime_type &&
|
||||||
extract_file_types in c$http$mime_type )
|
extract_file_types in c$http$mime_type )
|
||||||
{
|
{
|
||||||
c$http$extracting_file = T;
|
c$http$extract_file = T;
|
||||||
local suffix = fmt("%s_%d.dat", is_orig ? "orig" : "resp", ++c$http_state$entity_bodies);
|
}
|
||||||
|
|
||||||
|
if ( c$http$extract_file )
|
||||||
|
{
|
||||||
|
local suffix = fmt("%s_%d.dat", is_orig ? "orig" : "resp", c$http_state$current_response);
|
||||||
local fname = generate_extraction_filename(extraction_prefix, c, suffix);
|
local fname = generate_extraction_filename(extraction_prefix, c, suffix);
|
||||||
|
|
||||||
c$http$extraction_file = open(fname);
|
c$http$extraction_file = open(fname);
|
||||||
|
@ -50,12 +49,12 @@ event http_entity_data(c: connection, is_orig: bool, length: count, data: string
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if ( c$http$extracting_file )
|
if ( c$http?$extraction_file )
|
||||||
print c$http$extraction_file, data;
|
print c$http$extraction_file, data;
|
||||||
}
|
}
|
||||||
|
|
||||||
event http_end_entity(c: connection, is_orig: bool)
|
event http_end_entity(c: connection, is_orig: bool)
|
||||||
{
|
{
|
||||||
if ( c$http$extracting_file )
|
if ( c$http?$extraction_file )
|
||||||
close(c$http$extraction_file);
|
close(c$http$extraction_file);
|
||||||
}
|
}
|
||||||
|
|
|
@ -11,7 +11,8 @@ export {
|
||||||
};
|
};
|
||||||
|
|
||||||
redef record Info += {
|
redef record Info += {
|
||||||
## The MD5 sum for a file transferred over HTTP will be stored here.
|
## MD5 sum for a file transferred over HTTP calculated from the
|
||||||
|
## response body.
|
||||||
md5: string &log &optional;
|
md5: string &log &optional;
|
||||||
|
|
||||||
## This value can be set per-transfer to determine per request
|
## This value can be set per-transfer to determine per request
|
||||||
|
@ -19,8 +20,8 @@ export {
|
||||||
## set to T at the time of or before the first chunk of body data.
|
## set to T at the time of or before the first chunk of body data.
|
||||||
calc_md5: bool &default=F;
|
calc_md5: bool &default=F;
|
||||||
|
|
||||||
## This boolean value indicates if an MD5 sum is currently being
|
## Indicates if an MD5 sum is being calculated for the current
|
||||||
## calculated for the current file transfer.
|
## request/response pair.
|
||||||
calculating_md5: bool &default=F;
|
calculating_md5: bool &default=F;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
##! This script is involved in the identification of file types in HTTP
|
##! Identification of file types in HTTP response bodies with file content sniffing.
|
||||||
##! response bodies.
|
|
||||||
|
|
||||||
@load base/frameworks/signatures
|
@load base/frameworks/signatures
|
||||||
@load base/frameworks/notice
|
@load base/frameworks/notice
|
||||||
|
@ -15,30 +14,32 @@ module HTTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
# This notice is thrown when the file extension doesn't
|
## Indicates when the file extension doesn't seem to match the file contents.
|
||||||
# seem to match the file contents.
|
|
||||||
Incorrect_File_Type,
|
Incorrect_File_Type,
|
||||||
};
|
};
|
||||||
|
|
||||||
redef record Info += {
|
redef record Info += {
|
||||||
## This will record the mime_type identified.
|
## Mime type of response body identified by content sniffing.
|
||||||
mime_type: string &log &optional;
|
mime_type: string &log &optional;
|
||||||
|
|
||||||
## This indicates that no data of the current file transfer has been
|
## Indicates that no data of the current file transfer has been
|
||||||
## seen yet. After the first :bro:id:`http_entity_data` event, it
|
## seen yet. After the first :bro:id:`http_entity_data` event, it
|
||||||
## will be set to T.
|
## will be set to F.
|
||||||
first_chunk: bool &default=T;
|
first_chunk: bool &default=T;
|
||||||
};
|
};
|
||||||
|
|
||||||
redef enum Tags += {
|
## Mapping between mime types and regular expressions for URLs
|
||||||
IDENTIFIED_FILE
|
## The :bro:enum:`HTTP::Incorrect_File_Type` notice is generated if the pattern
|
||||||
};
|
## doesn't match the mime type that was discovered.
|
||||||
|
|
||||||
# Create regexes that *should* in be in the urls for specifics mime types.
|
|
||||||
# Notices are thrown if the pattern doesn't match the url for the file type.
|
|
||||||
const mime_types_extensions: table[string] of pattern = {
|
const mime_types_extensions: table[string] of pattern = {
|
||||||
["application/x-dosexec"] = /\.([eE][xX][eE]|[dD][lL][lL])/,
|
["application/x-dosexec"] = /\.([eE][xX][eE]|[dD][lL][lL])/,
|
||||||
} &redef;
|
} &redef;
|
||||||
|
|
||||||
|
## A pattern for filtering out :bro:enum:`HTTP::Incorrect_File_Type` urls
|
||||||
|
## that are not noteworthy before a notice is created. Each
|
||||||
|
## pattern added should match the complete URL (the matched URLs include
|
||||||
|
## "http://" at the beginning).
|
||||||
|
const ignored_incorrect_file_type_urls = /^$/ &redef;
|
||||||
}
|
}
|
||||||
|
|
||||||
event signature_match(state: signature_state, msg: string, data: string) &priority=5
|
event signature_match(state: signature_state, msg: string, data: string) &priority=5
|
||||||
|
@ -59,6 +60,10 @@ event signature_match(state: signature_state, msg: string, data: string) &priori
|
||||||
c$http?$uri && mime_types_extensions[msg] !in c$http$uri )
|
c$http?$uri && mime_types_extensions[msg] !in c$http$uri )
|
||||||
{
|
{
|
||||||
local url = build_url_http(c$http);
|
local url = build_url_http(c$http);
|
||||||
|
|
||||||
|
if ( url == ignored_incorrect_file_type_urls )
|
||||||
|
return;
|
||||||
|
|
||||||
local message = fmt("%s %s %s", msg, c$http$method, url);
|
local message = fmt("%s %s %s", msg, c$http$method, url);
|
||||||
NOTICE([$note=Incorrect_File_Type,
|
NOTICE([$note=Incorrect_File_Type,
|
||||||
$msg=message,
|
$msg=message,
|
||||||
|
|
|
@ -1,3 +1,7 @@
|
||||||
|
##! Implements base functionality for HTTP analysis. The logging model is
|
||||||
|
##! to log request/response pairs and all relevant metadata together in
|
||||||
|
##! a single record.
|
||||||
|
|
||||||
@load base/utils/numbers
|
@load base/utils/numbers
|
||||||
@load base/utils/files
|
@load base/utils/files
|
||||||
|
|
||||||
|
@ -8,6 +12,7 @@ export {
|
||||||
|
|
||||||
## Indicate a type of attack or compromise in the record to be logged.
|
## Indicate a type of attack or compromise in the record to be logged.
|
||||||
type Tags: enum {
|
type Tags: enum {
|
||||||
|
## Placeholder.
|
||||||
EMPTY
|
EMPTY
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -15,64 +20,69 @@ export {
|
||||||
const default_capture_password = F &redef;
|
const default_capture_password = F &redef;
|
||||||
|
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## Timestamp for when the request happened.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
## This represents the pipelined depth into the connection of this
|
## Represents the pipelined depth into the connection of this
|
||||||
## request/response transaction.
|
## request/response transaction.
|
||||||
trans_depth: count &log;
|
trans_depth: count &log;
|
||||||
## The verb used in the HTTP request (GET, POST, HEAD, etc.).
|
## Verb used in the HTTP request (GET, POST, HEAD, etc.).
|
||||||
method: string &log &optional;
|
method: string &log &optional;
|
||||||
## The value of the HOST header.
|
## Value of the HOST header.
|
||||||
host: string &log &optional;
|
host: string &log &optional;
|
||||||
## The URI used in the request.
|
## URI used in the request.
|
||||||
uri: string &log &optional;
|
uri: string &log &optional;
|
||||||
## The value of the "referer" header. The comment is deliberately
|
## Value of the "referer" header. The comment is deliberately
|
||||||
## misspelled like the standard declares, but the name used here is
|
## misspelled like the standard declares, but the name used here is
|
||||||
## "referrer" spelled correctly.
|
## "referrer" spelled correctly.
|
||||||
referrer: string &log &optional;
|
referrer: string &log &optional;
|
||||||
## The value of the User-Agent header from the client.
|
## Value of the User-Agent header from the client.
|
||||||
user_agent: string &log &optional;
|
user_agent: string &log &optional;
|
||||||
## The actual uncompressed content size of the data transferred from
|
## Actual uncompressed content size of the data transferred from
|
||||||
## the client.
|
## the client.
|
||||||
request_body_len: count &log &default=0;
|
request_body_len: count &log &default=0;
|
||||||
## The actual uncompressed content size of the data transferred from
|
## Actual uncompressed content size of the data transferred from
|
||||||
## the server.
|
## the server.
|
||||||
response_body_len: count &log &default=0;
|
response_body_len: count &log &default=0;
|
||||||
## The status code returned by the server.
|
## Status code returned by the server.
|
||||||
status_code: count &log &optional;
|
status_code: count &log &optional;
|
||||||
## The status message returned by the server.
|
## Status message returned by the server.
|
||||||
status_msg: string &log &optional;
|
status_msg: string &log &optional;
|
||||||
## The last 1xx informational reply code returned by the server.
|
## Last seen 1xx informational reply code returned by the server.
|
||||||
info_code: count &log &optional;
|
info_code: count &log &optional;
|
||||||
## The last 1xx informational reply message returned by the server.
|
## Last seen 1xx informational reply message returned by the server.
|
||||||
info_msg: string &log &optional;
|
info_msg: string &log &optional;
|
||||||
## The filename given in the Content-Disposition header
|
## Filename given in the Content-Disposition header sent by the server.
|
||||||
## sent by the server.
|
|
||||||
filename: string &log &optional;
|
filename: string &log &optional;
|
||||||
## This is a set of indicators of various attributes discovered and
|
## A set of indicators of various attributes discovered and
|
||||||
## related to a particular request/response pair.
|
## related to a particular request/response pair.
|
||||||
tags: set[Tags] &log;
|
tags: set[Tags] &log;
|
||||||
|
|
||||||
## The username if basic-auth is performed for the request.
|
## Username if basic-auth is performed for the request.
|
||||||
username: string &log &optional;
|
username: string &log &optional;
|
||||||
## The password if basic-auth is performed for the request.
|
## Password if basic-auth is performed for the request.
|
||||||
password: string &log &optional;
|
password: string &log &optional;
|
||||||
|
|
||||||
## This determines if the password will be captured for this request.
|
## Determines if the password will be captured for this request.
|
||||||
capture_password: bool &default=default_capture_password;
|
capture_password: bool &default=default_capture_password;
|
||||||
|
|
||||||
## All of the headers that may indicate if the request was proxied.
|
## All of the headers that may indicate if the request was proxied.
|
||||||
proxied: set[string] &log &optional;
|
proxied: set[string] &log &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Structure to maintain state for an HTTP connection with multiple
|
||||||
|
## requests and responses.
|
||||||
type State: record {
|
type State: record {
|
||||||
|
## Pending requests.
|
||||||
pending: table[count] of Info;
|
pending: table[count] of Info;
|
||||||
current_response: count &default=0;
|
## Current request in the pending queue.
|
||||||
current_request: count &default=0;
|
current_request: count &default=0;
|
||||||
|
## Current response in the pending queue.
|
||||||
|
current_response: count &default=0;
|
||||||
};
|
};
|
||||||
|
|
||||||
## The list of HTTP headers typically used to indicate a proxied request.
|
## A list of HTTP headers typically used to indicate proxied requests.
|
||||||
const proxy_headers: set[string] = {
|
const proxy_headers: set[string] = {
|
||||||
"FORWARDED",
|
"FORWARDED",
|
||||||
"X-FORWARDED-FOR",
|
"X-FORWARDED-FOR",
|
||||||
|
@ -83,6 +93,8 @@ export {
|
||||||
"PROXY-CONNECTION",
|
"PROXY-CONNECTION",
|
||||||
} &redef;
|
} &redef;
|
||||||
|
|
||||||
|
## Event that can be handled to access the HTTP record as it is sent on
|
||||||
|
## to the logging framework.
|
||||||
global log_http: event(rec: Info);
|
global log_http: event(rec: Info);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -5,8 +5,31 @@
|
||||||
module HTTP;
|
module HTTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## Given a string containing a series of key-value pairs separated by "=",
|
||||||
|
## this function can be used to parse out all of the key names.
|
||||||
|
##
|
||||||
|
## data: The raw data, such as a URL or cookie value.
|
||||||
|
##
|
||||||
|
## kv_splitter: A regular expression representing the separator between
|
||||||
|
## key-value pairs.
|
||||||
|
##
|
||||||
|
## Returns: A vector of strings containing the keys.
|
||||||
global extract_keys: function(data: string, kv_splitter: pattern): string_vec;
|
global extract_keys: function(data: string, kv_splitter: pattern): string_vec;
|
||||||
|
|
||||||
|
## Creates a URL from an :bro:type:`HTTP::Info` record. This should handle
|
||||||
|
## edge cases such as proxied requests appropriately.
|
||||||
|
##
|
||||||
|
## rec: An :bro:type:`HTTP::Info` record.
|
||||||
|
##
|
||||||
|
## Returns: A URL, not prefixed by "http://".
|
||||||
global build_url: function(rec: Info): string;
|
global build_url: function(rec: Info): string;
|
||||||
|
|
||||||
|
## Creates a URL from an :bro:type:`HTTP::Info` record. This should handle
|
||||||
|
## edge cases such as proxied requests appropriately.
|
||||||
|
##
|
||||||
|
## rec: An :bro:type:`HTTP::Info` record.
|
||||||
|
##
|
||||||
|
## Returns: A URL prefixed with "http://".
|
||||||
global build_url_http: function(rec: Info): string;
|
global build_url_http: function(rec: Info): string;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -5,8 +5,9 @@
|
||||||
##! but that connection will actually be between B and C which could be
|
##! but that connection will actually be between B and C which could be
|
||||||
##! analyzed on a different worker.
|
##! analyzed on a different worker.
|
||||||
##!
|
##!
|
||||||
##! Example line from IRC server indicating that the DCC SEND is about to start:
|
|
||||||
##! PRIVMSG my_nick :^ADCC SEND whateverfile.zip 3640061780 1026 41709^A
|
# Example line from IRC server indicating that the DCC SEND is about to start:
|
||||||
|
# PRIVMSG my_nick :^ADCC SEND whateverfile.zip 3640061780 1026 41709^A
|
||||||
|
|
||||||
@load ./main
|
@load ./main
|
||||||
@load base/utils/files
|
@load base/utils/files
|
||||||
|
@ -14,23 +15,24 @@
|
||||||
module IRC;
|
module IRC;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Tag += { EXTRACTED_FILE };
|
|
||||||
|
|
||||||
## Pattern of file mime types to extract from IRC DCC file transfers.
|
## Pattern of file mime types to extract from IRC DCC file transfers.
|
||||||
const extract_file_types = /NO_DEFAULT/ &redef;
|
const extract_file_types = /NO_DEFAULT/ &redef;
|
||||||
|
|
||||||
## The on-disk prefix for files to be extracted from IRC DCC file transfers.
|
## On-disk prefix for files to be extracted from IRC DCC file transfers.
|
||||||
const extraction_prefix = "irc-dcc-item" &redef;
|
const extraction_prefix = "irc-dcc-item" &redef;
|
||||||
|
|
||||||
redef record Info += {
|
redef record Info += {
|
||||||
|
## DCC filename requested.
|
||||||
dcc_file_name: string &log &optional;
|
dcc_file_name: string &log &optional;
|
||||||
|
## Size of the DCC transfer as indicated by the sender.
|
||||||
dcc_file_size: count &log &optional;
|
dcc_file_size: count &log &optional;
|
||||||
|
## Sniffed mime type of the file.
|
||||||
dcc_mime_type: string &log &optional;
|
dcc_mime_type: string &log &optional;
|
||||||
|
|
||||||
## The file handle for the file to be extracted
|
## The file handle for the file to be extracted
|
||||||
extraction_file: file &log &optional;
|
extraction_file: file &log &optional;
|
||||||
|
|
||||||
## A boolean to indicate if the current file transfer should be extraced.
|
## A boolean to indicate if the current file transfer should be extracted.
|
||||||
extract_file: bool &default=F;
|
extract_file: bool &default=F;
|
||||||
|
|
||||||
## The count of the number of file that have been extracted during the session.
|
## The count of the number of file that have been extracted during the session.
|
||||||
|
@ -54,8 +56,10 @@ event file_transferred(c: connection, prefix: string, descr: string,
|
||||||
if ( extract_file_types == irc$dcc_mime_type )
|
if ( extract_file_types == irc$dcc_mime_type )
|
||||||
{
|
{
|
||||||
irc$extract_file = T;
|
irc$extract_file = T;
|
||||||
add irc$tags[EXTRACTED_FILE];
|
}
|
||||||
|
|
||||||
|
if ( irc$extract_file )
|
||||||
|
{
|
||||||
local suffix = fmt("%d.dat", ++irc$num_extracted_files);
|
local suffix = fmt("%d.dat", ++irc$num_extracted_files);
|
||||||
local fname = generate_extraction_filename(extraction_prefix, c, suffix);
|
local fname = generate_extraction_filename(extraction_prefix, c, suffix);
|
||||||
irc$extraction_file = open(fname);
|
irc$extraction_file = open(fname);
|
||||||
|
@ -76,7 +80,7 @@ event file_transferred(c: connection, prefix: string, descr: string,
|
||||||
Log::write(IRC::LOG, irc);
|
Log::write(IRC::LOG, irc);
|
||||||
irc$command = tmp;
|
irc$command = tmp;
|
||||||
|
|
||||||
if ( irc$extract_file && irc?$extraction_file )
|
if ( irc?$extraction_file )
|
||||||
set_contents_file(id, CONTENTS_RESP, irc$extraction_file);
|
set_contents_file(id, CONTENTS_RESP, irc$extraction_file);
|
||||||
|
|
||||||
# Delete these values in case another DCC transfer
|
# Delete these values in case another DCC transfer
|
||||||
|
|
|
@ -1,36 +1,38 @@
|
||||||
##! This is the script that implements the core IRC analysis support. It only
|
##! Implements the core IRC analysis support. The logging model is to log
|
||||||
##! logs a very limited subset of the IRC protocol by default. The points
|
##! IRC commands along with the associated response and some additional
|
||||||
##! that it logs at are NICK commands, USER commands, and JOIN commands. It
|
##! metadata about the connection if it's available.
|
||||||
##! log various bits of meta data as indicated in the :bro:type:`Info` record
|
|
||||||
##! along with the command at the command arguments.
|
|
||||||
|
|
||||||
module IRC;
|
module IRC;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
type Tag: enum {
|
|
||||||
EMPTY
|
|
||||||
};
|
|
||||||
|
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## Timestamp when the command was seen.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
|
## Nick name given for the connection.
|
||||||
nick: string &log &optional;
|
nick: string &log &optional;
|
||||||
|
## User name given for the connection.
|
||||||
user: string &log &optional;
|
user: string &log &optional;
|
||||||
channels: set[string] &log &optional;
|
|
||||||
|
|
||||||
|
## Command given by the client.
|
||||||
command: string &log &optional;
|
command: string &log &optional;
|
||||||
|
## Value for the command given by the client.
|
||||||
value: string &log &optional;
|
value: string &log &optional;
|
||||||
|
## Any additional data for the command.
|
||||||
addl: string &log &optional;
|
addl: string &log &optional;
|
||||||
tags: set[Tag] &log;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
|
## Event that can be handled to access the IRC record as it is sent on
|
||||||
|
## to the logging framework.
|
||||||
global irc_log: event(rec: Info);
|
global irc_log: event(rec: Info);
|
||||||
}
|
}
|
||||||
|
|
||||||
redef record connection += {
|
redef record connection += {
|
||||||
|
## IRC session information.
|
||||||
irc: Info &optional;
|
irc: Info &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -14,15 +14,17 @@
|
||||||
module SSH;
|
module SSH;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The SSH protocol logging stream identifier.
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
## This indicates that a heuristically detected "successful" SSH
|
## Indicates that a heuristically detected "successful" SSH
|
||||||
## authentication occurred.
|
## authentication occurred.
|
||||||
Login
|
Login
|
||||||
};
|
};
|
||||||
|
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## Time when the SSH connection began.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
|
@ -34,11 +36,11 @@ export {
|
||||||
## would be set for the opposite situation.
|
## would be set for the opposite situation.
|
||||||
# TODO: handle local-local and remote-remote better.
|
# TODO: handle local-local and remote-remote better.
|
||||||
direction: Direction &log &optional;
|
direction: Direction &log &optional;
|
||||||
## The software string given by the client.
|
## Software string given by the client.
|
||||||
client: string &log &optional;
|
client: string &log &optional;
|
||||||
## The software string given by the server.
|
## Software string given by the server.
|
||||||
server: string &log &optional;
|
server: string &log &optional;
|
||||||
## The amount of data returned from the server. This is currently
|
## Amount of data returned from the server. This is currently
|
||||||
## the only measure of the success heuristic and it is logged to
|
## the only measure of the success heuristic and it is logged to
|
||||||
## assist analysts looking at the logs to make their own determination
|
## assist analysts looking at the logs to make their own determination
|
||||||
## about the success on a case-by-case basis.
|
## about the success on a case-by-case basis.
|
||||||
|
@ -48,8 +50,8 @@ export {
|
||||||
done: bool &default=F;
|
done: bool &default=F;
|
||||||
};
|
};
|
||||||
|
|
||||||
## The size in bytes at which the SSH connection is presumed to be
|
## The size in bytes of data sent by the server at which the SSH
|
||||||
## successful.
|
## connection is presumed to be successful.
|
||||||
const authentication_data_size = 5500 &redef;
|
const authentication_data_size = 5500 &redef;
|
||||||
|
|
||||||
## If true, we tell the event engine to not look at further data
|
## If true, we tell the event engine to not look at further data
|
||||||
|
@ -58,14 +60,16 @@ export {
|
||||||
## kinds of analyses (e.g., tracking connection size).
|
## kinds of analyses (e.g., tracking connection size).
|
||||||
const skip_processing_after_detection = F &redef;
|
const skip_processing_after_detection = F &redef;
|
||||||
|
|
||||||
## This event is generated when the heuristic thinks that a login
|
## Event that is generated when the heuristic thinks that a login
|
||||||
## was successful.
|
## was successful.
|
||||||
global heuristic_successful_login: event(c: connection);
|
global heuristic_successful_login: event(c: connection);
|
||||||
|
|
||||||
## This event is generated when the heuristic thinks that a login
|
## Event that is generated when the heuristic thinks that a login
|
||||||
## failed.
|
## failed.
|
||||||
global heuristic_failed_login: event(c: connection);
|
global heuristic_failed_login: event(c: connection);
|
||||||
|
|
||||||
|
## Event that can be handled to access the :bro:type:`SSH::Info`
|
||||||
|
## record as it is sent on to the logging framework.
|
||||||
global log_ssh: event(rec: Info);
|
global log_ssh: event(rec: Info);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,18 +1,65 @@
|
||||||
module SSL;
|
module SSL;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
|
||||||
const SSLv2 = 0x0002;
|
const SSLv2 = 0x0002;
|
||||||
const SSLv3 = 0x0300;
|
const SSLv3 = 0x0300;
|
||||||
const TLSv10 = 0x0301;
|
const TLSv10 = 0x0301;
|
||||||
const TLSv11 = 0x0302;
|
const TLSv11 = 0x0302;
|
||||||
|
const TLSv12 = 0x0303;
|
||||||
|
## Mapping between the constants and string values for SSL/TLS versions.
|
||||||
const version_strings: table[count] of string = {
|
const version_strings: table[count] of string = {
|
||||||
[SSLv2] = "SSLv2",
|
[SSLv2] = "SSLv2",
|
||||||
[SSLv3] = "SSLv3",
|
[SSLv3] = "SSLv3",
|
||||||
[TLSv10] = "TLSv10",
|
[TLSv10] = "TLSv10",
|
||||||
[TLSv11] = "TLSv11",
|
[TLSv11] = "TLSv11",
|
||||||
|
[TLSv12] = "TLSv12",
|
||||||
} &default="UNKNOWN";
|
} &default="UNKNOWN";
|
||||||
|
|
||||||
|
## Mapping between numeric codes and human readable strings for alert
|
||||||
|
## levels.
|
||||||
|
const alert_levels: table[count] of string = {
|
||||||
|
[1] = "warning",
|
||||||
|
[2] = "fatal",
|
||||||
|
} &default=function(i: count):string { return fmt("unknown-%d", i); };
|
||||||
|
|
||||||
|
## Mapping between numeric codes and human readable strings for alert
|
||||||
|
## descriptions..
|
||||||
|
const alert_descriptions: table[count] of string = {
|
||||||
|
[0] = "close_notify",
|
||||||
|
[10] = "unexpected_message",
|
||||||
|
[20] = "bad_record_mac",
|
||||||
|
[21] = "decryption_failed",
|
||||||
|
[22] = "record_overflow",
|
||||||
|
[30] = "decompression_failure",
|
||||||
|
[40] = "handshake_failure",
|
||||||
|
[41] = "no_certificate",
|
||||||
|
[42] = "bad_certificate",
|
||||||
|
[43] = "unsupported_certificate",
|
||||||
|
[44] = "certificate_revoked",
|
||||||
|
[45] = "certificate_expired",
|
||||||
|
[46] = "certificate_unknown",
|
||||||
|
[47] = "illegal_parameter",
|
||||||
|
[48] = "unknown_ca",
|
||||||
|
[49] = "access_denied",
|
||||||
|
[50] = "decode_error",
|
||||||
|
[51] = "decrypt_error",
|
||||||
|
[60] = "export_restriction",
|
||||||
|
[70] = "protocol_version",
|
||||||
|
[71] = "insufficient_security",
|
||||||
|
[80] = "internal_error",
|
||||||
|
[90] = "user_canceled",
|
||||||
|
[100] = "no_renegotiation",
|
||||||
|
[110] = "unsupported_extension",
|
||||||
|
[111] = "certificate_unobtainable",
|
||||||
|
[112] = "unrecognized_name",
|
||||||
|
[113] = "bad_certificate_status_response",
|
||||||
|
[114] = "bad_certificate_hash_value",
|
||||||
|
[115] = "unknown_psk_identity",
|
||||||
|
} &default=function(i: count):string { return fmt("unknown-%d", i); };
|
||||||
|
|
||||||
|
## Mapping between numeric codes and human readable strings for SSL/TLS
|
||||||
|
## extensions.
|
||||||
|
# More information can be found here:
|
||||||
# http://www.iana.org/assignments/tls-extensiontype-values/tls-extensiontype-values.xml
|
# http://www.iana.org/assignments/tls-extensiontype-values/tls-extensiontype-values.xml
|
||||||
const extensions: table[count] of string = {
|
const extensions: table[count] of string = {
|
||||||
[0] = "server_name",
|
[0] = "server_name",
|
||||||
|
@ -31,10 +78,11 @@ export {
|
||||||
[13] = "signature_algorithms",
|
[13] = "signature_algorithms",
|
||||||
[14] = "use_srtp",
|
[14] = "use_srtp",
|
||||||
[35] = "SessionTicket TLS",
|
[35] = "SessionTicket TLS",
|
||||||
|
[13172] = "next_protocol_negotiation",
|
||||||
[65281] = "renegotiation_info"
|
[65281] = "renegotiation_info"
|
||||||
} &default=function(i: count):string { return fmt("unknown-%d", i); };
|
} &default=function(i: count):string { return fmt("unknown-%d", i); };
|
||||||
|
|
||||||
## SSLv2
|
# SSLv2
|
||||||
const SSLv20_CK_RC4_128_WITH_MD5 = 0x010080;
|
const SSLv20_CK_RC4_128_WITH_MD5 = 0x010080;
|
||||||
const SSLv20_CK_RC4_128_EXPORT40_WITH_MD5 = 0x020080;
|
const SSLv20_CK_RC4_128_EXPORT40_WITH_MD5 = 0x020080;
|
||||||
const SSLv20_CK_RC2_128_CBC_WITH_MD5 = 0x030080;
|
const SSLv20_CK_RC2_128_CBC_WITH_MD5 = 0x030080;
|
||||||
|
@ -43,7 +91,7 @@ export {
|
||||||
const SSLv20_CK_DES_64_CBC_WITH_MD5 = 0x060040;
|
const SSLv20_CK_DES_64_CBC_WITH_MD5 = 0x060040;
|
||||||
const SSLv20_CK_DES_192_EDE3_CBC_WITH_MD5 = 0x0700C0;
|
const SSLv20_CK_DES_192_EDE3_CBC_WITH_MD5 = 0x0700C0;
|
||||||
|
|
||||||
## TLS
|
# TLS
|
||||||
const TLS_NULL_WITH_NULL_NULL = 0x0000;
|
const TLS_NULL_WITH_NULL_NULL = 0x0000;
|
||||||
const TLS_RSA_WITH_NULL_MD5 = 0x0001;
|
const TLS_RSA_WITH_NULL_MD5 = 0x0001;
|
||||||
const TLS_RSA_WITH_NULL_SHA = 0x0002;
|
const TLS_RSA_WITH_NULL_SHA = 0x0002;
|
||||||
|
@ -261,12 +309,10 @@ export {
|
||||||
const SSL_RSA_WITH_3DES_EDE_CBC_MD5 = 0xFF83;
|
const SSL_RSA_WITH_3DES_EDE_CBC_MD5 = 0xFF83;
|
||||||
const TLS_EMPTY_RENEGOTIATION_INFO_SCSV = 0x00FF;
|
const TLS_EMPTY_RENEGOTIATION_INFO_SCSV = 0x00FF;
|
||||||
|
|
||||||
# --- This is a table of all known cipher specs.
|
## This is a table of all known cipher specs. It can be used for
|
||||||
# --- It can be used for detecting unknown ciphers and for
|
## detecting unknown ciphers and for converting the cipher spec constants
|
||||||
# --- converting the cipher spec constants into a human readable format.
|
## into a human readable format.
|
||||||
|
|
||||||
const cipher_desc: table[count] of string = {
|
const cipher_desc: table[count] of string = {
|
||||||
# --- sslv20 ---
|
|
||||||
[SSLv20_CK_RC4_128_EXPORT40_WITH_MD5] =
|
[SSLv20_CK_RC4_128_EXPORT40_WITH_MD5] =
|
||||||
"SSLv20_CK_RC4_128_EXPORT40_WITH_MD5",
|
"SSLv20_CK_RC4_128_EXPORT40_WITH_MD5",
|
||||||
[SSLv20_CK_RC4_128_WITH_MD5] = "SSLv20_CK_RC4_128_WITH_MD5",
|
[SSLv20_CK_RC4_128_WITH_MD5] = "SSLv20_CK_RC4_128_WITH_MD5",
|
||||||
|
@ -278,7 +324,6 @@ export {
|
||||||
"SSLv20_CK_DES_192_EDE3_CBC_WITH_MD5",
|
"SSLv20_CK_DES_192_EDE3_CBC_WITH_MD5",
|
||||||
[SSLv20_CK_DES_64_CBC_WITH_MD5] = "SSLv20_CK_DES_64_CBC_WITH_MD5",
|
[SSLv20_CK_DES_64_CBC_WITH_MD5] = "SSLv20_CK_DES_64_CBC_WITH_MD5",
|
||||||
|
|
||||||
# --- TLS ---
|
|
||||||
[TLS_NULL_WITH_NULL_NULL] = "TLS_NULL_WITH_NULL_NULL",
|
[TLS_NULL_WITH_NULL_NULL] = "TLS_NULL_WITH_NULL_NULL",
|
||||||
[TLS_RSA_WITH_NULL_MD5] = "TLS_RSA_WITH_NULL_MD5",
|
[TLS_RSA_WITH_NULL_MD5] = "TLS_RSA_WITH_NULL_MD5",
|
||||||
[TLS_RSA_WITH_NULL_SHA] = "TLS_RSA_WITH_NULL_SHA",
|
[TLS_RSA_WITH_NULL_SHA] = "TLS_RSA_WITH_NULL_SHA",
|
||||||
|
@ -492,6 +537,7 @@ export {
|
||||||
[SSL_RSA_FIPS_WITH_3DES_EDE_CBC_SHA_2] = "SSL_RSA_FIPS_WITH_3DES_EDE_CBC_SHA_2",
|
[SSL_RSA_FIPS_WITH_3DES_EDE_CBC_SHA_2] = "SSL_RSA_FIPS_WITH_3DES_EDE_CBC_SHA_2",
|
||||||
} &default="UNKNOWN";
|
} &default="UNKNOWN";
|
||||||
|
|
||||||
|
## Mapping between the constants and string values for SSL/TLS errors.
|
||||||
const x509_errors: table[count] of string = {
|
const x509_errors: table[count] of string = {
|
||||||
[0] = "ok",
|
[0] = "ok",
|
||||||
[1] = "unable to get issuer cert",
|
[1] = "unable to get issuer cert",
|
||||||
|
@ -526,8 +572,7 @@ export {
|
||||||
[30] = "akid issuer serial mismatch",
|
[30] = "akid issuer serial mismatch",
|
||||||
[31] = "keyusage no certsign",
|
[31] = "keyusage no certsign",
|
||||||
[32] = "unable to get crl issuer",
|
[32] = "unable to get crl issuer",
|
||||||
[33] = "unhandled critical extension"
|
[33] = "unhandled critical extension",
|
||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,3 +1,6 @@
|
||||||
|
##! Base SSL analysis script. This script logs information about the SSL/TLS
|
||||||
|
##! handshaking and encryption establishment process.
|
||||||
|
|
||||||
@load ./consts
|
@load ./consts
|
||||||
|
|
||||||
module SSL;
|
module SSL;
|
||||||
|
@ -6,27 +9,41 @@ export {
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## Time when the SSL connection began.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
|
## SSL/TLS version the server offered.
|
||||||
version: string &log &optional;
|
version: string &log &optional;
|
||||||
|
## SSL/TLS cipher suite the server chose.
|
||||||
cipher: string &log &optional;
|
cipher: string &log &optional;
|
||||||
|
## Value of the Server Name Indicator SSL/TLS extension. It
|
||||||
|
## indicates the server name that the client was requesting.
|
||||||
server_name: string &log &optional;
|
server_name: string &log &optional;
|
||||||
|
## Session ID offered by the client for session resumption.
|
||||||
session_id: string &log &optional;
|
session_id: string &log &optional;
|
||||||
|
## Subject of the X.509 certificate offered by the server.
|
||||||
subject: string &log &optional;
|
subject: string &log &optional;
|
||||||
|
## NotValidBefore field value from the server certificate.
|
||||||
not_valid_before: time &log &optional;
|
not_valid_before: time &log &optional;
|
||||||
|
## NotValidAfter field value from the serve certificate.
|
||||||
not_valid_after: time &log &optional;
|
not_valid_after: time &log &optional;
|
||||||
|
## Last alert that was seen during the connection.
|
||||||
|
last_alert: string &log &optional;
|
||||||
|
|
||||||
|
## Full binary server certificate stored in DER format.
|
||||||
cert: string &optional;
|
cert: string &optional;
|
||||||
|
## Chain of certificates offered by the server to validate its
|
||||||
|
## complete signing chain.
|
||||||
cert_chain: vector of string &optional;
|
cert_chain: vector of string &optional;
|
||||||
|
|
||||||
## This stores the analyzer id used for the analyzer instance attached
|
## The analyzer ID used for the analyzer instance attached
|
||||||
## to each connection. It is not used for logging since it's a
|
## to each connection. It is not used for logging since it's a
|
||||||
## meaningless arbitrary number.
|
## meaningless arbitrary number.
|
||||||
analyzer_id: count &optional;
|
analyzer_id: count &optional;
|
||||||
};
|
};
|
||||||
|
|
||||||
## This is where the default root CA bundle is defined. By loading the
|
## The default root CA bundle. By loading the
|
||||||
## mozilla-ca-list.bro script it will be set to Mozilla's root CA list.
|
## mozilla-ca-list.bro script it will be set to Mozilla's root CA list.
|
||||||
const root_certs: table[string] of string = {} &redef;
|
const root_certs: table[string] of string = {} &redef;
|
||||||
|
|
||||||
|
@ -40,12 +57,9 @@ export {
|
||||||
## utility.
|
## utility.
|
||||||
const openssl_util = "openssl" &redef;
|
const openssl_util = "openssl" &redef;
|
||||||
|
|
||||||
|
## Event that can be handled to access the SSL
|
||||||
|
## record as it is sent on to the logging framework.
|
||||||
global log_ssl: event(rec: Info);
|
global log_ssl: event(rec: Info);
|
||||||
|
|
||||||
const ports = {
|
|
||||||
443/tcp, 563/tcp, 585/tcp, 614/tcp, 636/tcp,
|
|
||||||
989/tcp, 990/tcp, 992/tcp, 993/tcp, 995/tcp, 5223/tcp
|
|
||||||
} &redef;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
redef record connection += {
|
redef record connection += {
|
||||||
|
@ -72,6 +86,11 @@ redef capture_filters += {
|
||||||
["xmpps"] = "tcp port 5223",
|
["xmpps"] = "tcp port 5223",
|
||||||
};
|
};
|
||||||
|
|
||||||
|
const ports = {
|
||||||
|
443/tcp, 563/tcp, 585/tcp, 614/tcp, 636/tcp,
|
||||||
|
989/tcp, 990/tcp, 992/tcp, 993/tcp, 995/tcp, 5223/tcp
|
||||||
|
};
|
||||||
|
|
||||||
redef dpd_config += {
|
redef dpd_config += {
|
||||||
[[ANALYZER_SSL]] = [$ports = ports]
|
[[ANALYZER_SSL]] = [$ports = ports]
|
||||||
};
|
};
|
||||||
|
@ -112,10 +131,14 @@ event ssl_server_hello(c: connection, version: count, possible_ts: time, session
|
||||||
c$ssl$cipher = cipher_desc[cipher];
|
c$ssl$cipher = cipher_desc[cipher];
|
||||||
}
|
}
|
||||||
|
|
||||||
event x509_certificate(c: connection, cert: X509, is_server: bool, chain_idx: count, chain_len: count, der_cert: string) &priority=5
|
event x509_certificate(c: connection, is_orig: bool, cert: X509, chain_idx: count, chain_len: count, der_cert: string) &priority=5
|
||||||
{
|
{
|
||||||
set_session(c);
|
set_session(c);
|
||||||
|
|
||||||
|
# We aren't doing anything with client certificates yet.
|
||||||
|
if ( is_orig )
|
||||||
|
return;
|
||||||
|
|
||||||
if ( chain_idx == 0 )
|
if ( chain_idx == 0 )
|
||||||
{
|
{
|
||||||
# Save the primary cert.
|
# Save the primary cert.
|
||||||
|
@ -133,14 +156,21 @@ event x509_certificate(c: connection, cert: X509, is_server: bool, chain_idx: co
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
event ssl_extension(c: connection, code: count, val: string) &priority=5
|
event ssl_extension(c: connection, is_orig: bool, code: count, val: string) &priority=5
|
||||||
{
|
{
|
||||||
set_session(c);
|
set_session(c);
|
||||||
|
|
||||||
if ( extensions[code] == "server_name" )
|
if ( is_orig && extensions[code] == "server_name" )
|
||||||
c$ssl$server_name = sub_bytes(val, 6, |val|);
|
c$ssl$server_name = sub_bytes(val, 6, |val|);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
event ssl_alert(c: connection, is_orig: bool, level: count, desc: count) &priority=5
|
||||||
|
{
|
||||||
|
set_session(c);
|
||||||
|
|
||||||
|
c$ssl$last_alert = alert_descriptions[desc];
|
||||||
|
}
|
||||||
|
|
||||||
event ssl_established(c: connection) &priority=5
|
event ssl_established(c: connection) &priority=5
|
||||||
{
|
{
|
||||||
set_session(c);
|
set_session(c);
|
||||||
|
|
|
@ -1,6 +1,9 @@
|
||||||
|
##! Constants definitions for syslog.
|
||||||
|
|
||||||
module Syslog;
|
module Syslog;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## Mapping between the constants and string values for syslog facilities.
|
||||||
const facility_codes: table[count] of string = {
|
const facility_codes: table[count] of string = {
|
||||||
[0] = "KERN",
|
[0] = "KERN",
|
||||||
[1] = "USER",
|
[1] = "USER",
|
||||||
|
@ -28,6 +31,7 @@ export {
|
||||||
[23] = "LOCAL7",
|
[23] = "LOCAL7",
|
||||||
} &default=function(c: count): string { return fmt("?-%d", c); };
|
} &default=function(c: count): string { return fmt("?-%d", c); };
|
||||||
|
|
||||||
|
## Mapping between the constants and string values for syslog severities.
|
||||||
const severity_codes: table[count] of string = {
|
const severity_codes: table[count] of string = {
|
||||||
[0] = "EMERG",
|
[0] = "EMERG",
|
||||||
[1] = "ALERT",
|
[1] = "ALERT",
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
##! Core script support for logging syslog messages.
|
##! Core script support for logging syslog messages. This script represents
|
||||||
|
##! one syslog message as one logged record.
|
||||||
|
|
||||||
@load ./consts
|
@load ./consts
|
||||||
|
|
||||||
|
@ -8,19 +9,23 @@ export {
|
||||||
redef enum Log::ID += { LOG };
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
type Info: record {
|
type Info: record {
|
||||||
|
## Timestamp of when the syslog message was seen.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
uid: string &log;
|
uid: string &log;
|
||||||
id: conn_id &log;
|
id: conn_id &log;
|
||||||
|
## Protocol over which the message was seen.
|
||||||
proto: transport_proto &log;
|
proto: transport_proto &log;
|
||||||
|
## Syslog facility for the message.
|
||||||
facility: string &log;
|
facility: string &log;
|
||||||
|
## Syslog severity for the message.
|
||||||
severity: string &log;
|
severity: string &log;
|
||||||
|
## The plain text message.
|
||||||
message: string &log;
|
message: string &log;
|
||||||
};
|
};
|
||||||
|
|
||||||
const ports = { 514/udp } &redef;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
redef capture_filters += { ["syslog"] = "port 514" };
|
redef capture_filters += { ["syslog"] = "port 514" };
|
||||||
|
const ports = { 514/udp } &redef;
|
||||||
redef dpd_config += { [ANALYZER_SYSLOG_BINPAC] = [$ports = ports] };
|
redef dpd_config += { [ANALYZER_SYSLOG_BINPAC] = [$ports = ports] };
|
||||||
|
|
||||||
redef likely_server_ports += { 514/udp };
|
redef likely_server_ports += { 514/udp };
|
||||||
|
|
|
@ -18,7 +18,7 @@ export {
|
||||||
const local_nets: set[subnet] &redef;
|
const local_nets: set[subnet] &redef;
|
||||||
|
|
||||||
## This is used for retrieving the subnet when you multiple
|
## This is used for retrieving the subnet when you multiple
|
||||||
## :bro:id:`local_nets`. A membership query can be done with an
|
## :bro:id:`Site::local_nets`. A membership query can be done with an
|
||||||
## :bro:type:`addr` and the table will yield the subnet it was found
|
## :bro:type:`addr` and the table will yield the subnet it was found
|
||||||
## within.
|
## within.
|
||||||
global local_nets_table: table[subnet] of subnet = {};
|
global local_nets_table: table[subnet] of subnet = {};
|
||||||
|
|
|
@ -1,3 +1,12 @@
|
||||||
|
##! The controllee portion of the control framework. Load this script if remote
|
||||||
|
##! runtime control of the Bro process is desired.
|
||||||
|
##!
|
||||||
|
##! A controllee only needs to load the controllee script in addition
|
||||||
|
##! to the specific analysis scripts desired. It may also need a node
|
||||||
|
##! configured as a controller node in the communications nodes configuration::
|
||||||
|
##!
|
||||||
|
##! bro <scripts> frameworks/control/controllee
|
||||||
|
|
||||||
@load base/frameworks/control
|
@load base/frameworks/control
|
||||||
# If an instance is a controllee, it implicitly needs to listen for remote
|
# If an instance is a controllee, it implicitly needs to listen for remote
|
||||||
# connections.
|
# connections.
|
||||||
|
|
|
@ -1,3 +1,10 @@
|
||||||
|
##! This is a utility script that implements the controller interface for the
|
||||||
|
##! control framework. It's intended to be run to control a remote Bro
|
||||||
|
##! and then shutdown.
|
||||||
|
##!
|
||||||
|
##! It's intended to be used from the command line like this::
|
||||||
|
##! bro <scripts> frameworks/control/controller Control::host=<host_addr> Control::port=<host_port> Control::cmd=<command> [Control::arg=<arg>]
|
||||||
|
|
||||||
@load base/frameworks/control
|
@load base/frameworks/control
|
||||||
@load base/frameworks/communication
|
@load base/frameworks/communication
|
||||||
|
|
||||||
|
|
|
@ -8,7 +8,6 @@ module ProtocolDetector;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
Off_Port_Protocol_Found, # raised for each connection found
|
|
||||||
Protocol_Found,
|
Protocol_Found,
|
||||||
Server_Found,
|
Server_Found,
|
||||||
};
|
};
|
||||||
|
@ -155,13 +154,10 @@ function report_protocols(c: connection)
|
||||||
{
|
{
|
||||||
if ( [a, c$id$resp_h, c$id$resp_p] in valids )
|
if ( [a, c$id$resp_h, c$id$resp_p] in valids )
|
||||||
do_notice(c, a, valids[a, c$id$resp_h, c$id$resp_p]);
|
do_notice(c, a, valids[a, c$id$resp_h, c$id$resp_p]);
|
||||||
|
|
||||||
else if ( [a, 0.0.0.0, c$id$resp_p] in valids )
|
else if ( [a, 0.0.0.0, c$id$resp_p] in valids )
|
||||||
do_notice(c, a, valids[a, 0.0.0.0, c$id$resp_p]);
|
do_notice(c, a, valids[a, 0.0.0.0, c$id$resp_p]);
|
||||||
else
|
else
|
||||||
do_notice(c, a, NONE);
|
do_notice(c, a, NONE);
|
||||||
|
|
||||||
append_addl(c, analyzer_name(a));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
delete conns[c$id];
|
delete conns[c$id];
|
||||||
|
@ -218,20 +214,6 @@ event protocol_confirmation(c: connection, atype: count, aid: count)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
# event connection_analyzer_disabled(c: connection, analyzer: count)
|
|
||||||
# {
|
|
||||||
# if ( c$id !in conns )
|
|
||||||
# return;
|
|
||||||
#
|
|
||||||
# delete conns[c$id][analyzer];
|
|
||||||
# }
|
|
||||||
|
|
||||||
function append_proto_addl(c: connection)
|
|
||||||
{
|
|
||||||
for ( a in conns[c$id] )
|
|
||||||
append_addl(c, fmt_protocol(get_protocol(c, a)));
|
|
||||||
}
|
|
||||||
|
|
||||||
function found_protocol(c: connection, analyzer: count, protocol: string)
|
function found_protocol(c: connection, analyzer: count, protocol: string)
|
||||||
{
|
{
|
||||||
# Don't report anything running on a well-known port.
|
# Don't report anything running on a well-known port.
|
||||||
|
|
|
@ -1,3 +1,6 @@
|
||||||
|
##! An example of using the metrics framework to collect connection metrics
|
||||||
|
##! aggregated into /24 CIDR ranges.
|
||||||
|
|
||||||
@load base/frameworks/metrics
|
@load base/frameworks/metrics
|
||||||
@load base/utils/site
|
@load base/utils/site
|
||||||
|
|
||||||
|
|
|
@ -1,9 +1,17 @@
|
||||||
|
##! Provides an example of aggregating and limiting collection down to
|
||||||
|
##! only local networks. Additionally, the status code for the response from
|
||||||
|
##! the request is added into the metric.
|
||||||
|
|
||||||
@load base/frameworks/metrics
|
@load base/frameworks/metrics
|
||||||
@load base/protocols/http
|
@load base/protocols/http
|
||||||
@load base/utils/site
|
@load base/utils/site
|
||||||
|
|
||||||
redef enum Metrics::ID += {
|
redef enum Metrics::ID += {
|
||||||
|
## Measures HTTP requests indexed on both the request host and the response
|
||||||
|
## code from the server.
|
||||||
HTTP_REQUESTS_BY_STATUS_CODE,
|
HTTP_REQUESTS_BY_STATUS_CODE,
|
||||||
|
|
||||||
|
## Currently unfinished and not working.
|
||||||
HTTP_REQUESTS_BY_HOST_HEADER,
|
HTTP_REQUESTS_BY_HOST_HEADER,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -11,13 +19,13 @@ event bro_init()
|
||||||
{
|
{
|
||||||
# TODO: these are waiting on a fix with table vals + records before they will work.
|
# TODO: these are waiting on a fix with table vals + records before they will work.
|
||||||
#Metrics::add_filter(HTTP_REQUESTS_BY_HOST_HEADER,
|
#Metrics::add_filter(HTTP_REQUESTS_BY_HOST_HEADER,
|
||||||
# [$pred(index: Index) = { return Site:is_local_addr(index$host) },
|
# [$pred(index: Metrics::Index) = { return Site::is_local_addr(index$host); },
|
||||||
# $aggregation_mask=24,
|
# $aggregation_mask=24,
|
||||||
# $break_interval=5mins]);
|
# $break_interval=1min]);
|
||||||
#
|
|
||||||
## Site::local_nets must be defined in order for this to actually do anything.
|
# Site::local_nets must be defined in order for this to actually do anything.
|
||||||
#Metrics::add_filter(HTTP_REQUESTS_BY_STATUS_CODE, [$aggregation_table=Site::local_nets_table,
|
Metrics::add_filter(HTTP_REQUESTS_BY_STATUS_CODE, [$aggregation_table=Site::local_nets_table,
|
||||||
# $break_interval=5mins]);
|
$break_interval=1min]);
|
||||||
}
|
}
|
||||||
|
|
||||||
event HTTP::log_http(rec: HTTP::Info)
|
event HTTP::log_http(rec: HTTP::Info)
|
||||||
|
|
|
@ -1,3 +1,8 @@
|
||||||
|
##! Provides an example of using the metrics framework to collect the number
|
||||||
|
##! of times a specific server name indicator value is seen in SSL session
|
||||||
|
##! establishments. Names ending in google.com are being filtered out as an
|
||||||
|
##! example of the predicate based filtering in metrics filters.
|
||||||
|
|
||||||
@load base/frameworks/metrics
|
@load base/frameworks/metrics
|
||||||
@load base/protocols/ssl
|
@load base/protocols/ssl
|
||||||
|
|
||||||
|
|
|
@ -1,3 +1,7 @@
|
||||||
|
##! Provides the possibly to define software names that are interesting to
|
||||||
|
##! watch for changes. A notice is generated if software versions change on a
|
||||||
|
##! host.
|
||||||
|
|
||||||
@load base/frameworks/notice
|
@load base/frameworks/notice
|
||||||
@load base/frameworks/software
|
@load base/frameworks/software
|
||||||
|
|
||||||
|
@ -5,24 +9,17 @@ module Software;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
## For certain softwares, a version changing may matter. In that case,
|
## For certain software, a version changing may matter. In that case,
|
||||||
## this notice will be generated. Software that matters if the version
|
## this notice will be generated. Software that matters if the version
|
||||||
## changes can be configured with the
|
## changes can be configured with the
|
||||||
## :bro:id:`Software::interesting_version_changes` variable.
|
## :bro:id:`Software::interesting_version_changes` variable.
|
||||||
Software_Version_Change,
|
Software_Version_Change,
|
||||||
};
|
};
|
||||||
|
|
||||||
## Some software is more interesting when the version changes and this
|
## Some software is more interesting when the version changes and this is
|
||||||
## a set of all software that should raise a notice when a different
|
## a set of all software that should raise a notice when a different
|
||||||
## version is seen on a host.
|
## version is seen on a host.
|
||||||
const interesting_version_changes: set[string] = {
|
const interesting_version_changes: set[string] = { } &redef;
|
||||||
"SSH"
|
|
||||||
} &redef;
|
|
||||||
|
|
||||||
## Some software is more interesting when the version changes and this
|
|
||||||
## a set of all software that should raise a notice when a different
|
|
||||||
## version is seen on a host.
|
|
||||||
const interesting_type_changes: set[string] = {};
|
|
||||||
}
|
}
|
||||||
|
|
||||||
event log_software(rec: Info)
|
event log_software(rec: Info)
|
||||||
|
|
|
@ -1,3 +1,7 @@
|
||||||
|
##! Provides a variable to define vulnerable versions of software and if a
|
||||||
|
##! a version of that software as old or older than the defined version a
|
||||||
|
##! notice will be generated.
|
||||||
|
|
||||||
@load base/frameworks/notice
|
@load base/frameworks/notice
|
||||||
@load base/frameworks/software
|
@load base/frameworks/software
|
||||||
|
|
||||||
|
@ -5,6 +9,7 @@ module Software;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
|
## Indicates that a vulnerable version of software was detected.
|
||||||
Vulnerable_Version,
|
Vulnerable_Version,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -18,6 +23,7 @@ event log_software(rec: Info)
|
||||||
if ( rec$name in vulnerable_versions &&
|
if ( rec$name in vulnerable_versions &&
|
||||||
cmp_versions(rec$version, vulnerable_versions[rec$name]) <= 0 )
|
cmp_versions(rec$version, vulnerable_versions[rec$name]) <= 0 )
|
||||||
{
|
{
|
||||||
NOTICE([$note=Vulnerable_Version, $src=rec$host, $msg=software_fmt(rec)]);
|
NOTICE([$note=Vulnerable_Version, $src=rec$host,
|
||||||
|
$msg=fmt("A vulnerable version of software was detected: %s", software_fmt(rec))]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -15,7 +15,7 @@ export {
|
||||||
alert: AlertData &log;
|
alert: AlertData &log;
|
||||||
};
|
};
|
||||||
|
|
||||||
## This can convert a Barnyard :bro:type:`PacketID` value to a
|
## This can convert a Barnyard :bro:type:`Barnyard2::PacketID` value to a
|
||||||
## :bro:type:`conn_id` value in the case that you might need to index
|
## :bro:type:`conn_id` value in the case that you might need to index
|
||||||
## into an existing data structure elsewhere within Bro.
|
## into an existing data structure elsewhere within Bro.
|
||||||
global pid2cid: function(p: PacketID): conn_id;
|
global pid2cid: function(p: PacketID): conn_id;
|
||||||
|
|
|
@ -17,7 +17,7 @@ export {
|
||||||
|
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
## Report if the detected capture loss exceeds the percentage
|
## Report if the detected capture loss exceeds the percentage
|
||||||
## threshold
|
## threshold.
|
||||||
Too_Much_Loss
|
Too_Much_Loss
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -42,9 +42,9 @@ export {
|
||||||
const watch_interval = 15mins &redef;
|
const watch_interval = 15mins &redef;
|
||||||
|
|
||||||
## The percentage of missed data that is considered "too much"
|
## The percentage of missed data that is considered "too much"
|
||||||
## when the :bro:enum:`Too_Much_Loss` notice should be generated.
|
## when the :bro:enum:`CaptureLoss::Too_Much_Loss` notice should be
|
||||||
## The value is expressed as a double between 0 and 1 with 1 being
|
## generated. The value is expressed as a double between 0 and 1 with 1
|
||||||
## 100%
|
## being 100%
|
||||||
const too_much_loss: double = 0.1 &redef;
|
const too_much_loss: double = 0.1 &redef;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
##!
|
##! Log the loaded scripts.
|
||||||
|
|
||||||
module LoadedScripts;
|
module LoadedScripts;
|
||||||
|
|
||||||
|
|
|
@ -2,14 +2,13 @@
|
||||||
|
|
||||||
module Profiling;
|
module Profiling;
|
||||||
|
|
||||||
|
## Set the profiling output file.
|
||||||
redef profiling_file = open_log_file("prof");
|
redef profiling_file = open_log_file("prof");
|
||||||
|
|
||||||
export {
|
## Set the cheap profiling interval.
|
||||||
## Cheap profiling every 15 seconds.
|
redef profiling_interval = 15 secs;
|
||||||
redef profiling_interval = 15 secs &redef;
|
|
||||||
}
|
|
||||||
|
|
||||||
# Expensive profiling every 5 minutes.
|
## Set the expensive profiling interval.
|
||||||
redef expensive_profiling_multiple = 20;
|
redef expensive_profiling_multiple = 20;
|
||||||
|
|
||||||
event bro_init()
|
event bro_init()
|
||||||
|
|
83
scripts/policy/misc/stats.bro
Normal file
83
scripts/policy/misc/stats.bro
Normal file
|
@ -0,0 +1,83 @@
|
||||||
|
##! Log memory/packet/lag statistics. Differs from profiling.bro in that this
|
||||||
|
##! is lighter-weight (much less info, and less load to generate).
|
||||||
|
|
||||||
|
@load base/frameworks/notice
|
||||||
|
|
||||||
|
module Stats;
|
||||||
|
|
||||||
|
export {
|
||||||
|
redef enum Log::ID += { LOG };
|
||||||
|
|
||||||
|
## How often stats are reported.
|
||||||
|
const stats_report_interval = 1min &redef;
|
||||||
|
|
||||||
|
type Info: record {
|
||||||
|
## Timestamp for the measurement.
|
||||||
|
ts: time &log;
|
||||||
|
## Peer that generated this log. Mostly for clusters.
|
||||||
|
peer: string &log;
|
||||||
|
## Amount of memory currently in use in MB.
|
||||||
|
mem: count &log;
|
||||||
|
## Number of packets processed since the last stats interval.
|
||||||
|
pkts_proc: count &log;
|
||||||
|
## Number of events that been processed since the last stats interval.
|
||||||
|
events_proc: count &log;
|
||||||
|
## Number of events that have been queued since the last stats interval.
|
||||||
|
events_queued: count &log;
|
||||||
|
|
||||||
|
## Lag between the wall clock and packet timestamps if reading live traffic.
|
||||||
|
lag: interval &log &optional;
|
||||||
|
## Number of packets received since the last stats interval if reading
|
||||||
|
## live traffic.
|
||||||
|
pkts_recv: count &log &optional;
|
||||||
|
## Number of packets dropped since the last stats interval if reading
|
||||||
|
## live traffic.
|
||||||
|
pkts_dropped: count &log &optional;
|
||||||
|
## Number of packets seen on the link since the last stats interval
|
||||||
|
## if reading live traffic.
|
||||||
|
pkts_link: count &log &optional;
|
||||||
|
};
|
||||||
|
|
||||||
|
## Event to catch stats as they are written to the logging stream.
|
||||||
|
global log_stats: event(rec: Info);
|
||||||
|
}
|
||||||
|
|
||||||
|
event bro_init() &priority=5
|
||||||
|
{
|
||||||
|
Log::create_stream(Stats::LOG, [$columns=Info, $ev=log_stats]);
|
||||||
|
}
|
||||||
|
|
||||||
|
event check_stats(last_ts: time, last_ns: NetStats, last_res: bro_resources)
|
||||||
|
{
|
||||||
|
local now = current_time();
|
||||||
|
local ns = net_stats();
|
||||||
|
local res = resource_usage();
|
||||||
|
|
||||||
|
if ( bro_is_terminating() )
|
||||||
|
# No more stats will be written or scheduled when Bro is
|
||||||
|
# shutting down.
|
||||||
|
return;
|
||||||
|
|
||||||
|
local info: Info = [$ts=now, $peer=peer_description, $mem=res$mem/1000000,
|
||||||
|
$pkts_proc=res$num_packets - last_res$num_packets,
|
||||||
|
$events_proc=res$num_events_dispatched - last_res$num_events_dispatched,
|
||||||
|
$events_queued=res$num_events_queued - last_res$num_events_queued];
|
||||||
|
|
||||||
|
if ( reading_live_traffic() )
|
||||||
|
{
|
||||||
|
info$lag = now - network_time();
|
||||||
|
# Someone's going to have to explain what this is and add a field to the Info record.
|
||||||
|
# info$util = 100.0*((res$user_time + res$system_time) - (last_res$user_time + last_res$system_time))/(now-last_ts);
|
||||||
|
info$pkts_recv = ns$pkts_recvd - last_ns$pkts_recvd;
|
||||||
|
info$pkts_dropped = ns$pkts_dropped - last_ns$pkts_dropped;
|
||||||
|
info$pkts_link = ns$pkts_link - last_ns$pkts_link;
|
||||||
|
}
|
||||||
|
|
||||||
|
Log::write(Stats::LOG, info);
|
||||||
|
schedule stats_report_interval { check_stats(now, ns, res) };
|
||||||
|
}
|
||||||
|
|
||||||
|
event bro_init()
|
||||||
|
{
|
||||||
|
schedule stats_report_interval { check_stats(current_time(), net_stats(), resource_usage()) };
|
||||||
|
}
|
|
@ -10,7 +10,8 @@ export {
|
||||||
## This event can be generated externally to this script if on-demand
|
## This event can be generated externally to this script if on-demand
|
||||||
## tracefile rotation is required with the caveat that the script doesn't
|
## tracefile rotation is required with the caveat that the script doesn't
|
||||||
## currently attempt to get back on schedule automatically and the next
|
## currently attempt to get back on schedule automatically and the next
|
||||||
## trim will likely won't happen on the :bro:id:`trim_interval`.
|
## trim will likely won't happen on the
|
||||||
|
## :bro:id:`TrimTraceFile::trim_interval`.
|
||||||
global go: event(first_trim: bool);
|
global go: event(first_trim: bool);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -8,8 +8,10 @@
|
||||||
module Known;
|
module Known;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The known-hosts logging stream identifier.
|
||||||
redef enum Log::ID += { HOSTS_LOG };
|
redef enum Log::ID += { HOSTS_LOG };
|
||||||
|
|
||||||
|
## The record type which contains the column fields of the known-hosts log.
|
||||||
type HostsInfo: record {
|
type HostsInfo: record {
|
||||||
## The timestamp at which the host was detected.
|
## The timestamp at which the host was detected.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
|
@ -19,7 +21,7 @@ export {
|
||||||
};
|
};
|
||||||
|
|
||||||
## The hosts whose existence should be logged and tracked.
|
## The hosts whose existence should be logged and tracked.
|
||||||
## Choices are: LOCAL_HOSTS, REMOTE_HOSTS, ALL_HOSTS, NO_HOSTS
|
## See :bro:type:`Host` for possible choices.
|
||||||
const host_tracking = LOCAL_HOSTS &redef;
|
const host_tracking = LOCAL_HOSTS &redef;
|
||||||
|
|
||||||
## The set of all known addresses to store for preventing duplicate
|
## The set of all known addresses to store for preventing duplicate
|
||||||
|
@ -29,6 +31,8 @@ export {
|
||||||
## of each individual address is logged each day.
|
## of each individual address is logged each day.
|
||||||
global known_hosts: set[addr] &create_expire=1day &synchronized &redef;
|
global known_hosts: set[addr] &create_expire=1day &synchronized &redef;
|
||||||
|
|
||||||
|
## An event that can be handled to access the :bro:type:`Known::HostsInfo`
|
||||||
|
## record as it is sent on to the logging framework.
|
||||||
global log_known_hosts: event(rec: HostsInfo);
|
global log_known_hosts: event(rec: HostsInfo);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -8,29 +8,41 @@
|
||||||
module Known;
|
module Known;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
|
## The known-services logging stream identifier.
|
||||||
redef enum Log::ID += { SERVICES_LOG };
|
redef enum Log::ID += { SERVICES_LOG };
|
||||||
|
|
||||||
|
## The record type which contains the column fields of the known-services
|
||||||
|
## log.
|
||||||
type ServicesInfo: record {
|
type ServicesInfo: record {
|
||||||
|
## The time at which the service was detected.
|
||||||
ts: time &log;
|
ts: time &log;
|
||||||
|
## The host address on which the service is running.
|
||||||
host: addr &log;
|
host: addr &log;
|
||||||
|
## The port number on which the service is running.
|
||||||
port_num: port &log;
|
port_num: port &log;
|
||||||
|
## The transport-layer protocol which the service uses.
|
||||||
port_proto: transport_proto &log;
|
port_proto: transport_proto &log;
|
||||||
|
## A set of protocols that match the service's connection payloads.
|
||||||
service: set[string] &log;
|
service: set[string] &log;
|
||||||
|
|
||||||
done: bool &default=F;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
## The hosts whose services should be tracked and logged.
|
## The hosts whose services should be tracked and logged.
|
||||||
|
## See :bro:type:`Host` for possible choices.
|
||||||
const service_tracking = LOCAL_HOSTS &redef;
|
const service_tracking = LOCAL_HOSTS &redef;
|
||||||
|
|
||||||
|
## Tracks the set of daily-detected services for preventing the logging
|
||||||
|
## of duplicates, but can also be inspected by other scripts for
|
||||||
|
## different purposes.
|
||||||
global known_services: set[addr, port] &create_expire=1day &synchronized;
|
global known_services: set[addr, port] &create_expire=1day &synchronized;
|
||||||
|
|
||||||
|
## Event that can be handled to access the :bro:type:`Known::ServicesInfo`
|
||||||
|
## record as it is sent on to the logging framework.
|
||||||
global log_known_services: event(rec: ServicesInfo);
|
global log_known_services: event(rec: ServicesInfo);
|
||||||
}
|
}
|
||||||
|
|
||||||
redef record connection += {
|
redef record connection += {
|
||||||
## This field is to indicate whether or not the processing for detecting
|
# This field is to indicate whether or not the processing for detecting
|
||||||
## and logging the service for this connection is complete.
|
# and logging the service for this connection is complete.
|
||||||
known_services_done: bool &default=F;
|
known_services_done: bool &default=F;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -7,7 +7,7 @@ module FTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
## This indicates that a successful response to a "SITE EXEC"
|
## Indicates that a successful response to a "SITE EXEC"
|
||||||
## command/arg pair was seen.
|
## command/arg pair was seen.
|
||||||
Site_Exec_Success,
|
Site_Exec_Success,
|
||||||
};
|
};
|
||||||
|
|
|
@ -12,8 +12,10 @@ module FTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Software::Type += {
|
redef enum Software::Type += {
|
||||||
FTP_CLIENT,
|
## Identifier for FTP clients in the software framework.
|
||||||
FTP_SERVER,
|
CLIENT,
|
||||||
|
## Not currently implemented.
|
||||||
|
SERVER,
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -21,7 +23,7 @@ event ftp_request(c: connection, command: string, arg: string) &priority=4
|
||||||
{
|
{
|
||||||
if ( command == "CLNT" )
|
if ( command == "CLNT" )
|
||||||
{
|
{
|
||||||
local si = Software::parse(arg, c$id$orig_h, FTP_CLIENT);
|
local si = Software::parse(arg, c$id$orig_h, CLIENT);
|
||||||
Software::found(c$id, si);
|
Software::found(c$id, si);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,15 +1,18 @@
|
||||||
##! This script takes MD5 sums of files transferred over HTTP and checks them with
|
##! Detect file downloads over HTTP that have MD5 sums matching files in Team
|
||||||
##! Team Cymru's Malware Hash Registry (http://www.team-cymru.org/Services/MHR/).
|
##! Cymru's Malware Hash Registry (http://www.team-cymru.org/Services/MHR/).
|
||||||
##! By default, not all file transfers will have MD5 sums calculated. Read the
|
##! By default, not all file transfers will have MD5 sums calculated. Read the
|
||||||
##! documentation for the :doc:base/protocols/http/file-hash.bro script to see how to
|
##! documentation for the :doc:base/protocols/http/file-hash.bro script to see
|
||||||
##! configure which transfers will have hashes calculated.
|
##! how to configure which transfers will have hashes calculated.
|
||||||
|
|
||||||
@load base/frameworks/notice
|
@load base/frameworks/notice
|
||||||
@load base/protocols/http
|
@load base/protocols/http
|
||||||
|
|
||||||
|
module HTTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Notice::Type += {
|
redef enum Notice::Type += {
|
||||||
## If the MD5 sum of a file transferred over HTTP
|
## The MD5 sum of a file transferred over HTTP matched in the
|
||||||
|
## malware hash registry.
|
||||||
Malware_Hash_Registry_Match
|
Malware_Hash_Registry_Match
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
##! Intelligence based HTTP detections.
|
##! Intelligence based HTTP detections. Not yet working!
|
||||||
|
|
||||||
@load base/protocols/http/main
|
@load base/protocols/http/main
|
||||||
@load base/protocols/http/utils
|
@load base/protocols/http/utils
|
||||||
|
|
|
@ -12,12 +12,14 @@ export {
|
||||||
SQL_Injection_Attacker,
|
SQL_Injection_Attacker,
|
||||||
## Indicates that a host was seen to have SQL injection attacks against
|
## Indicates that a host was seen to have SQL injection attacks against
|
||||||
## it. This is tracked by IP address as opposed to hostname.
|
## it. This is tracked by IP address as opposed to hostname.
|
||||||
SQL_Injection_Attack_Against,
|
SQL_Injection_Victim,
|
||||||
};
|
};
|
||||||
|
|
||||||
redef enum Metrics::ID += {
|
redef enum Metrics::ID += {
|
||||||
SQL_ATTACKER,
|
## Metric to track SQL injection attackers.
|
||||||
SQL_ATTACKS_AGAINST,
|
SQLI_ATTACKER,
|
||||||
|
## Metrics to track SQL injection victims.
|
||||||
|
SQLI_VICTIM,
|
||||||
};
|
};
|
||||||
|
|
||||||
redef enum Tags += {
|
redef enum Tags += {
|
||||||
|
@ -30,17 +32,17 @@ export {
|
||||||
COOKIE_SQLI,
|
COOKIE_SQLI,
|
||||||
};
|
};
|
||||||
|
|
||||||
## This defines the threshold that determines if an SQL injection attack
|
## Defines the threshold that determines if an SQL injection attack
|
||||||
## is ongoing based on the number of requests that appear to be SQL
|
## is ongoing based on the number of requests that appear to be SQL
|
||||||
## injection attacks.
|
## injection attacks.
|
||||||
const sqli_requests_threshold = 50 &redef;
|
const sqli_requests_threshold = 50 &redef;
|
||||||
|
|
||||||
## Interval at which to watch for the :bro:id:`sqli_requests_threshold`
|
## Interval at which to watch for the
|
||||||
## variable to be crossed. At the end of each interval the counter is
|
## :bro:id:`HTTP::sqli_requests_threshold` variable to be crossed.
|
||||||
## reset.
|
## At the end of each interval the counter is reset.
|
||||||
const sqli_requests_interval = 5min &redef;
|
const sqli_requests_interval = 5min &redef;
|
||||||
|
|
||||||
## This regular expression is used to match URI based SQL injections
|
## Regular expression is used to match URI based SQL injections.
|
||||||
const match_sql_injection_uri =
|
const match_sql_injection_uri =
|
||||||
/[\?&][^[:blank:]\x00-\x37\|]+?=[\-[:alnum:]%]+([[:blank:]\x00-\x37]|\/\*.*?\*\/)*['"]?([[:blank:]\x00-\x37]|\/\*.*?\*\/|\)?;)+.*?([hH][aA][vV][iI][nN][gG]|[uU][nN][iI][oO][nN]|[eE][xX][eE][cC]|[sS][eE][lL][eE][cC][tT]|[dD][eE][lL][eE][tT][eE]|[dD][rR][oO][pP]|[dD][eE][cC][lL][aA][rR][eE]|[cC][rR][eE][aA][tT][eE]|[iI][nN][sS][eE][rR][tT])([[:blank:]\x00-\x37]|\/\*.*?\*\/)+/
|
/[\?&][^[:blank:]\x00-\x37\|]+?=[\-[:alnum:]%]+([[:blank:]\x00-\x37]|\/\*.*?\*\/)*['"]?([[:blank:]\x00-\x37]|\/\*.*?\*\/|\)?;)+.*?([hH][aA][vV][iI][nN][gG]|[uU][nN][iI][oO][nN]|[eE][xX][eE][cC]|[sS][eE][lL][eE][cC][tT]|[dD][eE][lL][eE][tT][eE]|[dD][rR][oO][pP]|[dD][eE][cC][lL][aA][rR][eE]|[cC][rR][eE][aA][tT][eE]|[iI][nN][sS][eE][rR][tT])([[:blank:]\x00-\x37]|\/\*.*?\*\/)+/
|
||||||
| /[\?&][^[:blank:]\x00-\x37\|]+?=[\-0-9%]+([[:blank:]\x00-\x37]|\/\*.*?\*\/)*['"]?([[:blank:]\x00-\x37]|\/\*.*?\*\/|\)?;)+([xX]?[oO][rR]|[nN]?[aA][nN][dD])([[:blank:]\x00-\x37]|\/\*.*?\*\/)+['"]?(([^a-zA-Z&]+)?=|[eE][xX][iI][sS][tT][sS])/
|
| /[\?&][^[:blank:]\x00-\x37\|]+?=[\-0-9%]+([[:blank:]\x00-\x37]|\/\*.*?\*\/)*['"]?([[:blank:]\x00-\x37]|\/\*.*?\*\/|\)?;)+([xX]?[oO][rR]|[nN]?[aA][nN][dD])([[:blank:]\x00-\x37]|\/\*.*?\*\/)+['"]?(([^a-zA-Z&]+)?=|[eE][xX][iI][sS][tT][sS])/
|
||||||
|
@ -56,14 +58,14 @@ event bro_init() &priority=3
|
||||||
# determine when it looks like an actual attack and how to respond when
|
# determine when it looks like an actual attack and how to respond when
|
||||||
# thresholds are crossed.
|
# thresholds are crossed.
|
||||||
|
|
||||||
Metrics::add_filter(SQL_ATTACKER, [$log=F,
|
Metrics::add_filter(SQLI_ATTACKER, [$log=F,
|
||||||
$notice_threshold=sqli_requests_threshold,
|
$notice_threshold=sqli_requests_threshold,
|
||||||
$break_interval=sqli_requests_interval,
|
$break_interval=sqli_requests_interval,
|
||||||
$note=SQL_Injection_Attacker]);
|
$note=SQL_Injection_Attacker]);
|
||||||
Metrics::add_filter(SQL_ATTACKS_AGAINST, [$log=F,
|
Metrics::add_filter(SQLI_VICTIM, [$log=F,
|
||||||
$notice_threshold=sqli_requests_threshold,
|
$notice_threshold=sqli_requests_threshold,
|
||||||
$break_interval=sqli_requests_interval,
|
$break_interval=sqli_requests_interval,
|
||||||
$note=SQL_Injection_Attack_Against]);
|
$note=SQL_Injection_Victim]);
|
||||||
}
|
}
|
||||||
|
|
||||||
event http_request(c: connection, method: string, original_URI: string,
|
event http_request(c: connection, method: string, original_URI: string,
|
||||||
|
@ -73,7 +75,7 @@ event http_request(c: connection, method: string, original_URI: string,
|
||||||
{
|
{
|
||||||
add c$http$tags[URI_SQLI];
|
add c$http$tags[URI_SQLI];
|
||||||
|
|
||||||
Metrics::add_data(SQL_ATTACKER, [$host=c$id$orig_h], 1);
|
Metrics::add_data(SQLI_ATTACKER, [$host=c$id$orig_h], 1);
|
||||||
Metrics::add_data(SQL_ATTACKS_AGAINST, [$host=c$id$resp_h], 1);
|
Metrics::add_data(SQLI_VICTIM, [$host=c$id$resp_h], 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,3 +1,5 @@
|
||||||
|
##! Detect and log web applications through the software framework.
|
||||||
|
|
||||||
@load base/frameworks/signatures
|
@load base/frameworks/signatures
|
||||||
@load base/frameworks/software
|
@load base/frameworks/software
|
||||||
@load base/protocols/http
|
@load base/protocols/http
|
||||||
|
@ -10,10 +12,12 @@ redef Signatures::ignored_ids += /^webapp-/;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Software::Type += {
|
redef enum Software::Type += {
|
||||||
|
## Identifier for web applications in the software framework.
|
||||||
WEB_APPLICATION,
|
WEB_APPLICATION,
|
||||||
};
|
};
|
||||||
|
|
||||||
redef record Software::Info += {
|
redef record Software::Info += {
|
||||||
|
## Most root URL where the software was discovered.
|
||||||
url: string &optional &log;
|
url: string &optional &log;
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
##! This script take advantage of a few ways that installed plugin information
|
##! Detect browser plugins as they leak through requests to Omniture
|
||||||
##! leaks from web browsers.
|
##! advertising servers.
|
||||||
|
|
||||||
@load base/protocols/http
|
@load base/protocols/http
|
||||||
@load base/frameworks/software
|
@load base/frameworks/software
|
||||||
|
@ -13,6 +13,7 @@ export {
|
||||||
};
|
};
|
||||||
|
|
||||||
redef enum Software::Type += {
|
redef enum Software::Type += {
|
||||||
|
## Identifier for browser plugins in the software framework.
|
||||||
BROWSER_PLUGIN
|
BROWSER_PLUGIN
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
|
@ -6,8 +6,11 @@ module HTTP;
|
||||||
|
|
||||||
export {
|
export {
|
||||||
redef enum Software::Type += {
|
redef enum Software::Type += {
|
||||||
|
## Identifier for web servers in the software framework.
|
||||||
SERVER,
|
SERVER,
|
||||||
|
## Identifier for app servers in the software framework.
|
||||||
APPSERVER,
|
APPSERVER,
|
||||||
|
## Identifier for web browsers in the software framework.
|
||||||
BROWSER,
|
BROWSER,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
Some files were not shown because too many files have changed in this diff Show more
Loading…
Add table
Add a link
Reference in a new issue