Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Per prospector configurable pipeline #3433

Merged
merged 2 commits into from
Jan 24, 2017
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions CHANGELOG.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -93,6 +93,7 @@ https://github.com/elastic/beats/compare/v5.1.1...master[Check the HEAD diff]
*Filebeat*
- Add enabled config option to prospectors. {pull}3157[3157]
- Add target option for decoded_json_field. {pull}3169[3169]
- Add the `pipeline` config option at the prospector level, for configuring the Ingest Node pipeline ID. {pull}3433[3433]

*Winlogbeat*

Expand Down
4 changes: 4 additions & 0 deletions filebeat/_meta/beat.full.yml
Original file line number Diff line number Diff line change
Expand Up @@ -143,6 +143,10 @@ filebeat.prospectors:
# this can mean that the first entries of a new file are skipped.
#tail_files: false

# The Ingest Node pipeline ID associated with this prospector. If this is set, it
# overwrites the pipeline option from the Elasticsearch output.
#pipeline:

# Experimental: If symlinks is enabled, symlinks are opened and harvested. The harvester is openening the
# original for harvesting but will report the symlink name as source.
#symlinks: false
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -405,6 +405,14 @@ This option applies to files that Filebeat has not already processed. If you ran

NOTE: You can use this setting to avoid indexing old log lines when you run Filebeat on a set of log files for the first time. After the first run, we recommend disabling this option, or you risk losing lines during file rotation.

===== pipeline

The Ingest Node pipeline ID to set for the events generated by this prospector.

NOTE: The pipeline ID can also be configured in the Elasticsearch output, but this
option usually results in simpler configuration files. If the pipeline is configured both
in the prospector and in the output, the option from the prospector is the one used.

===== symlinks

experimental[]
Expand Down
4 changes: 4 additions & 0 deletions filebeat/filebeat.full.yml
Original file line number Diff line number Diff line change
Expand Up @@ -143,6 +143,10 @@ filebeat.prospectors:
# this can mean that the first entries of a new file are skipped.
#tail_files: false

# The Ingest Node pipeline ID associated with this prospector. If this is set, it
# overwrites the pipeline option from the Elasticsearch output.
#pipeline:

# Experimental: If symlinks is enabled, symlinks are opened and harvested. The harvester is openening the
# original for harvesting but will report the symlink name as source.
#symlinks: false
Expand Down
1 change: 1 addition & 0 deletions filebeat/harvester/config.go
Original file line number Diff line number Diff line change
Expand Up @@ -52,6 +52,7 @@ type harvesterConfig struct {
MaxBytes int `config:"max_bytes" validate:"min=0,nonzero"`
Multiline *reader.MultilineConfig `config:"multiline"`
JSON *reader.JSONConfig `config:"json"`
Pipeline string `config:"pipeline"`
}

func (config *harvesterConfig) Validate() error {
Expand Down
1 change: 1 addition & 0 deletions filebeat/harvester/log.go
Original file line number Diff line number Diff line change
Expand Up @@ -137,6 +137,7 @@ func (h *Harvester) Harvest(r reader.Reader) {
event.InputType = h.config.InputType
event.DocumentType = h.config.DocumentType
event.JSONConfig = h.config.JSON
event.Pipeline = h.config.Pipeline
}

// Always send event to update state, also if lines was skipped
Expand Down
12 changes: 12 additions & 0 deletions filebeat/input/event.go
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@ type Event struct {
JSONConfig *reader.JSONConfig
State file.State
Data common.MapStr // Use in readers to add data to the event
Pipeline string
}

func NewEvent(state file.State) *Event {
Expand Down Expand Up @@ -58,6 +59,17 @@ func (e *Event) ToMapStr() common.MapStr {
return event
}

// Metadata creates a common.MapStr containing the metadata to
// be associated with the event.
func (e *Event) Metadata() common.MapStr {
if e.Pipeline != "" {
return common.MapStr{
"pipeline": e.Pipeline,
}
}
return nil
}
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

output metadata should also work if meta == nil =>

func (e *Event) Metadata() common.MapStr {
  if e.Pipeline != "" {
    return common.MapStr{
      "pipeline": e.Pipeline,
    }
  }
  return nil

}

This way you don't need to create a map if no pipeline is configured.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice, I didn't realize that would work.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I pushed a version with that.


// HasData returns true if the event itself contains data
// Events without data are only state updates
func (e *Event) HasData() bool {
Expand Down
6 changes: 4 additions & 2 deletions filebeat/publisher/async.go
Original file line number Diff line number Diff line change
Expand Up @@ -84,10 +84,12 @@ func (p *asyncLogPublisher) Start() {
flag: 0,
events: events,
}
dataEvents, meta := getDataEvents(events)
p.client.PublishEvents(
getDataEvents(events),
dataEvents,
publisher.Signal(batch),
publisher.Guaranteed)
publisher.Guaranteed,
publisher.MetadataBatch(meta))

p.active.append(batch)
case <-ticker.C:
Expand Down
9 changes: 6 additions & 3 deletions filebeat/publisher/publisher.go
Original file line number Diff line number Diff line change
Expand Up @@ -45,12 +45,15 @@ var (
)

// getDataEvents returns all events which contain data (not only state updates)
func getDataEvents(events []*input.Event) []common.MapStr {
dataEvents := make([]common.MapStr, 0, len(events))
// together with their associated metadata
func getDataEvents(events []*input.Event) (dataEvents []common.MapStr, meta []common.MapStr) {
dataEvents = make([]common.MapStr, 0, len(events))
meta = make([]common.MapStr, 0, len(events))
for _, event := range events {
if event.HasData() {
dataEvents = append(dataEvents, event.ToMapStr())
meta = append(meta, event.Metadata())
}
}
return dataEvents
return dataEvents, meta
}
4 changes: 3 additions & 1 deletion filebeat/publisher/sync.go
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,9 @@ func (p *syncLogPublisher) Publish() error {
case events = <-p.in:
}

ok := p.client.PublishEvents(getDataEvents(events), publisher.Sync, publisher.Guaranteed)
dataEvents, meta := getDataEvents(events)
ok := p.client.PublishEvents(dataEvents, publisher.Sync, publisher.Guaranteed,
publisher.MetadataBatch(meta))
if !ok {
// PublishEvents will only returns false, if p.client has been closed.
return sigPublisherStop
Expand Down
12 changes: 10 additions & 2 deletions filebeat/tests/system/config/filebeat.yml.j2
Original file line number Diff line number Diff line change
Expand Up @@ -28,6 +28,7 @@ filebeat.prospectors:
clean_removed: {{clean_removed}}
harvester_limit: {{harvester_limit | default(0) }}
symlinks: {{symlinks}}
pipeline: {{pipeline}}

{% if fields %}
fields:
Expand Down Expand Up @@ -132,11 +133,18 @@ processors:
# Configure what outputs to use when sending the data collected by the beat.
# Multiple outputs may be used.

#------------------------------- File output ----------------------------------
{%- if logstash %}
{%- if elasticsearch %}
#------------------------------- Elasticsearch output ----------------------------
output.elasticsearch:
hosts: ["{{ elasticsearch.host }}"]
pipeline: {{elasticsearch.pipeline}}
index: {{elasticsearch.index}}
{%- elif logstash %}
#------------------------------- Logstash output ---------------------------------
output.logstash:
hosts: ["{{ logstash.host }}"]
{%- else %}
#------------------------------- File output ----------------------------------
output.file:
path: {{ output_file_path|default(beat.working_dir + "/output") }}
filename: "{{ output_file_filename|default("filebeat") }}"
Expand Down
62 changes: 62 additions & 0 deletions filebeat/tests/system/test_modules.py
Original file line number Diff line number Diff line change
Expand Up @@ -110,3 +110,65 @@ def run_on_file(self, module, fileset, test_file, cfgfile):
if not found:
raise Exception("The following expected object was" +
" not found: {}".format(obj))

@unittest.skipIf(not INTEGRATION_TESTS or
os.getenv("TESTING_ENVIRONMENT") == "2x",
"integration test not available on 2.x")
def test_prospector_pipeline_config(self):
"""
Tests that the pipeline configured in the prospector overwrites
the one from the output.
"""
self.init()
index_name = "filebeat-test-prospector"
try:
self.es.indices.delete(index=index_name)
except:
pass
self.wait_until(lambda: not self.es.indices.exists(index_name))

self.render_config_template(
path=os.path.abspath(self.working_dir) + "/log/*",
elasticsearch=dict(
host=self.elasticsearch_url,
pipeline="estest",
index=index_name),
pipeline="test",
)

os.mkdir(self.working_dir + "/log/")
testfile = self.working_dir + "/log/test.log"
with open(testfile, 'a') as file:
file.write("Hello World1\n")

# put pipeline
self.es.transport.perform_request("PUT", "/_ingest/pipeline/test",
body={
"processors": [{
"set": {
"field": "x-pipeline",
"value": "test-pipeline",
}
}]})

filebeat = self.start_beat()

# Wait until the event is in ES
self.wait_until(lambda: self.es.indices.exists(index_name))

def search_objects():
try:
self.es.indices.refresh(index=index_name)
res = self.es.search(index=index_name,
body={"query": {"match_all": {}}})
return [o["_source"] for o in res["hits"]["hits"]]
except:
return []

self.wait_until(lambda: len(search_objects()) > 0, max_timeout=20)
filebeat.check_kill_and_wait()

objects = search_objects()
assert len(objects) == 1
o = objects[0]
assert o["x-pipeline"] == "test-pipeline"
2 changes: 1 addition & 1 deletion libbeat/publisher/client.go
Original file line number Diff line number Diff line change
Expand Up @@ -238,5 +238,5 @@ func MakeContext(opts []ClientOption) ([]common.MapStr, Context) {
}
}
}
return nil, ctx
return meta, ctx
}