-
Notifications
You must be signed in to change notification settings - Fork 39
/
Copy pathtelegraf.conf
100 lines (90 loc) · 3.37 KB
/
telegraf.conf
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
[agent]
interval = "1s"
round_interval = true
metric_batch_size = 1000
metric_buffer_limit = 10000
collection_jitter = "0s"
flush_interval = "1s"
flush_jitter = "0s"
precision = ""
debug = false
quiet = false
logfile = ""
hostname = ""
omit_hostname = false
###############################################################################
# OUTPUT PLUGINS #
###############################################################################
[[outputs.remp_elasticsearch]]
urls = ["http://elasticsearch:9200"] # required
timeout = "5s"
index_name = "pageviews"
namepass = ["pageviews"]
manage_template = false
id_field = "remp_pageview_id"
tagexclude = ["host"]
[[outputs.remp_elasticsearch]]
urls = ["http://elasticsearch:9200"] # required
timeout = "5s"
index_name = "pageviews_time_spent"
namepass = ["pageviews_time_spent"]
manage_template = false
id_field = "remp_pageview_id"
updated_fields = ["timespent"]
tagexclude = ["host"]
field_whitelist = ["token", "subscriber", "signed_in", "locked", "timespent", "unload"]
[[outputs.remp_elasticsearch]]
urls = ["http://elasticsearch:9200"] # required
timeout = "5s"
index_name = "pageviews_progress"
namepass = ["pageviews_progress"]
manage_template = false
id_field = "remp_pageview_id"
updated_fields = ["page_progress", "article_progress"]
field_whitelist = ["token", "subscriber", "signed_in", "locked", "page_progress", "article_progress", "unload"]
[[outputs.remp_elasticsearch]]
urls = ["http://elasticsearch:9200"] # required
timeout = "5s"
index_name = "events"
namepass = ["events_v2"]
manage_template = false
id_field = "remp_event_id"
tagexclude = ["host"]
[[outputs.remp_elasticsearch]]
urls = ["http://elasticsearch:9200"] # required
timeout = "5s"
index_name = "commerce"
namepass = ["commerce"]
manage_template = false
id_field = "remp_commerce_id"
tagexclude = ["host"]
[[outputs.remp_elasticsearch]]
urls = ["http://elasticsearch:9200"] # required
timeout = "5s"
index_name = "concurrents_by_browser"
# if you to track timespent, keep current setting as is so the concurrents can be based on more recent data
namepass = ["pageviews_time_spent"]
# if you don't plan to track timespent, you can make concurrents_by_browser based on regular pageviews; the data will be a bit less accurate
# namepass = ["pageviews"]
manage_template = false
id_field = "browser_id"
# TODO replace URL with canonical URL, without URL fragments
updated_fields = ["time", "article_id", "token", "derived_referer_host_with_path", "derived_referer_medium", "url"]
field_whitelist = ["time", "article_id", "token", "derived_referer_host_with_path", "derived_referer_medium", "url"]
[[outputs.remp_elasticsearch]]
urls = ["http://elasticsearch:9200"] # required
timeout = "5s"
index_name = "entities"
namepass = ["entities"]
manage_template = false
id_field = "remp_entity_id"
tagexclude = ["host"]
###############################################################################
# SERVICE INPUT PLUGINS #
###############################################################################
[[inputs.kafka_consumer]]
topics = ["beam_events"]
brokers = ["kafka:9092"]
consumer_group = "beam_consumers"
offset = "oldest"
data_format = "influx"