parser job from logsearch/204.0.0
Github source:
6b05207d or
master branch
Properties¶
logstash¶
metadata_level¶Whether to include additional metadata throughout the event lifecycle. NONE = disabled, DEBUG = fully enabled
- Default
NONE
output¶
elasticsearch¶
data_hosts¶The list of elasticsearch data node IPs
- Default
- 127.0.0.1
flush_size¶Controls how many logs will be buffered and sent to Elasticsearch for bulk indexing
- Default
500
logstash_parser¶
debug¶Debug level logging
- Default
false
deployment_dictionary¶A file that contains a hash of job name-deployment name keypairs for @source.deployment lookup
- Default
/var/vcap/packages/logsearch-config/deployment_lookup.yml
elasticsearch_document_id¶Use a specific, dynamic ID rather than an auto-generated identifier.
elasticsearch_index¶The specific, dynamic index name to write events to.
- Default
logstash-%{+YYYY.MM.dd}
elasticsearch_index_type¶The specific, dynamic index type name to write events to.
- Default
'%{@type}'
elasticsearch_routing¶The routing to be used when indexing a document.
enable_json_filter¶Toggles the if_it_looks_like_json.conf filter rule
- Default
false
filters¶The configuration to embed into the logstash filters section. Can either be a set of parsing rules as a string or a list of hashes in the form of [{name: path_to_parsing_rules.conf}]
- Default
""
idle_flush_time¶How frequently to flush events if the output queue is not full.
inputs¶A list of input plugins, with a hash of options for each of them. Please refer to example below.
- Default
- options: {} plugin: redis- Example
inputs: - options: host: 192.168.1.1 password: c1oudbunny user: logsearch plugin: rabbitmq
message_max_size¶Maximum log message length. Anything larger is truncated (TODO: move this to ingestor?)
- Default
1.048576e+06
outputs¶The configuration to embed into the logstash outputs section
timecop¶
reject_greater_than_hours¶Logs with timestamps greater than this many hours in the future won’t be parsed and will get tagged with fail/timecop
- Default
1
reject_less_than_hours¶Logs with timestamps less than this many hours in the past won’t be parsed and will get tagged with fail/timecop
- Default
24
wait_for_templates¶A list of index templates that need to be present in ElasticSearch before the process starts
- Default
- index_template
workers¶The number of worker threads that logstash should use (default: auto = one per CPU)
- Default
auto
redis¶
host¶Redis host of queue
key¶Name of queue to pull messages from
- Default
logstash
port¶Redis port of queue
- Default
6379
Templates¶
Templates are rendered and placed onto corresponding
instances during the deployment process. This job's templates
will be placed into /var/vcap/jobs/parser/ directory
(learn more).
bin/monit_debugger(frombin/monit_debugger)bin/parser_ctl(frombin/parser_ctl)config/filters_override.conf(fromconfig/filters_override.conf.erb)config/filters_post.conf(fromconfig/filters_post.conf.erb)config/filters_pre.conf(fromconfig/filters_pre.conf.erb)config/input_and_output.conf(fromconfig/input_and_output.conf.erb)data/properties.sh(fromdata/properties.sh.erb)helpers/ctl_setup.sh(fromhelpers/ctl_setup.sh)helpers/ctl_utils.sh(fromhelpers/ctl_utils.sh)logsearch/logs.yml(fromlogsearch/logs.yml)
Packages¶
Packages are compiled and placed onto corresponding
instances during the deployment process. Packages will be
placed into /var/vcap/packages/ directory.