The Installed Collector can gather data from several different types of Sources. You should manually migrate your Sources to an OpenTelemetry Configuration.
- General Configuration Concepts
- Collector
- Cloud Based Management
- Local Configuration File
- Collector
- Common Parameters
- Local File Source (LocalFile)
- Remote File Source (RemoteFileV2)
- Syslog Source (Syslog)
- Docker Logs Source (DockerLog)
- Docker Stats Source (DockerStats)
- Script Source (Script)
- Streaming Metrics Source (StreamingMetrics)
- Host Metrics Source (SystemStats)
- Local Windows Event Log Source (LocalWindowsEventLog)
- Remote Windows Event Log Source (RemoteWindowsEventLog)
- Local Windows Performance Source (LocalWindowsPerfMon)
- Remote Windows Performance Source (RemoteWindowsPerfMon)
- Windows Active Directory Source (ActiveDirectory)
Let's consider the following example:
extensions:
sumologic:
install_token: <install_token>
receivers:
filelog:
include:
- /var/log/syslog
tcplog:
listen_address: 0.0.0.0:514
processors:
memory_limiter:
check_interval: 1s
limit_mib: 4000
spike_limit_mib: 800
resource:
attributes:
- key: author
value: me
action: insert
exporters:
sumologic:
service:
extensions:
- sumologic
pipelines:
logs/example pipeline:
receivers:
- filelog
- tcplog
processors:
- memory_limiter
- resource
exporters:
- sumologic
We can differentiate four types of modules:
- extensions - unrelated to data processing, but responsible for additional actions, like collector registration (eg. sumologic extension)
- receivers - responsible for receiving data and pushing it to processors
- processors - responsible for data modification, like adding fields, limiting memory and so on
- exporters - responsible for sending data, received by receivers and processed by processors
To use those configured modules, they need to be mentioned in the service
section.
service
consists of extensions
(they are global across collector) and pipelines
.
Pipelines
can be logs
, metrics
, and traces
and each of them can have
receivers
, processors
and exporters
. Multiple pipelines of one type can be configured using aliases,
such as example pipeline
for logs
in the example above.
Collector registration and configuration is handled by the sumologicextension.
Collector name can be specified by setting the collector_name
option:
extensions:
sumologic:
install_token: <install_token>
collector_name: my_collector
To set a description, use the collector_description
option:
extensions:
sumologic:
install_token: <install_token>
collector_name: my_collector
collector_description: This is my and only my collector
Host name can be set in the sumologic exporter configuration. The exporter will set the host name for every record sent to Sumo Logic:
extensions:
sumologic:
install_token: <install_token>
collector_name: my_collector
collector_description: This is my and only my collector
exporters:
sumologic:
source_host: My hostname
To set a Collector category, use the collector_category
option:
extensions:
sumologic:
install_token: <install_token>
collector_name: my_collector
collector_description: This is my and only my collector
collector_category: example
exporters:
sumologic:
source_host: My hostname
Fields in the Opentelemetry Collector can be added with the resourceprocessor.
For example, to add a field with the key author
with the value me
to every record,
you could use the following configuration:
extensions:
sumologic:
install_token: <install_token>
collector_name: my_collector
collector_description: This is my and only my collector
collector_category: example
processors:
resource:
attributes:
- key: author
value: me
action: insert
exporters:
sumologic:
source_host: My hostname
Assignment to an Ingest Budget is not supported by Opentelemetry Collector.
To set the Collector time zone, use the time_zone
option.
For example, the following examples sets the time zone to America/Tijuana
:
extensions:
sumologic:
install_token: <install_token>
collector_name: my_collector
collector_description: This is my and only my collector
collector_category: example
time_zone: America/Tijuana
processors:
resource:
attributes:
- key: author
value: me
action: insert
exporters:
sumologic:
source_host: My hostname
CPU Target is not supported by the Opentelemetry Collector.
Currently, the Opentelemetry Collector can only be managed with Local Configuration File Management. Depending on your setup, follow the steps in Cloud Based Management or Local Configuration File for migration details.
This section describes migration steps for Sources managed from the Cloud.
Below is an example of an OpenTelemetry configuration for a Local File Source.
extensions:
sumologic:
install_token: <install_token>
## Time Zone is a substitute of Installed Collector `Time Zone`
## with `Use time zone from log file. If none is detected use:` option.
## This is used only if `clear_logs_timestamp` is set to `true` in sumologic exporter.
## Full list of time zones is available on wikipedia:
## https://en.wikipedia.org/wiki/List_of_tz_database_time_zones#List
time_zone: America/Tijuana
receivers:
## There is no substitute for `Description` in current project phase.
## It is recommended to use comments for that purpose, like this one.
## filelog/<source group name>:
## <source group name> can be substitute of Installed Collector `Name`.
filelog/log source:
## List of local files which should be read.
## Installed Collector `File path` substitute.
include:
- /var/log/*.log
- /opt/app/logs/*.log
## List of local files which shouldn't be read.
## Installed Collector `Denylist` substitute.
exclude:
- /var/log/auth.log
- /opt/app/logs/security_*.log
## There is no substitute of Installed Collector `Collection should begin`.
## This is nearest config and can take one of two values: `beginning` or `end`.
start_at: beginning
## encoding is substitute for Installed Collector `Encoding`.
## List of supported encodings:
## https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/v0.33.0/receiver/filelogreceiver
encoding: utf-8
## multiline is Opentelemetry Collector substitute for `Enable Multiline Processing`.
## As multiline detection behaves slightly different than in Installed Collector
## the following section in filelog documentation is recommended to read:
## https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/v0.33.0/receiver/filelogreceiver#multiline-configuration
multiline:
## line_start_pattern is substitute of `Boundary Regex`.
line_start_pattern: ^\d{4}
processors:
## The following configuration will add two fields to every record
resource/log source:
attributes:
- key: cloud.availability_zone
value: zone-1
action: insert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic:
## Set _sourceName
source_name: my example name
## Installed Collector substitute for `Source Category`.
source_category: example category
## Installed Collector substitute for `Source Host`.
source_host: example host
## clear_logs_timestamp is by default set to True.
## If it's set to true, it works like `Enable Timestamp Parsing`,
## and `Time Zone` is going to be taken from `extensions` section.
## There is no possibility to configure several time zones in one exporter.
## clear_logs_timestamp sets to true also behaves like
## `Timestamp Format` would be set to `Automatically detect the format`
## in terms of Installed Collector configuration.
clear_logs_timestamp: true
service:
extensions:
- sumologic
pipelines:
logs/log source:
receivers:
- filelog/log source
processors:
- resource/log source
exporters:
- sumologic
Define the name after the slash /
in the receiver name.
To set _sourceName
, use resourceprocessor
or set it in sumologicexporter.
For example, the following snippet configures the name as my example name
:
receivers:
filelog/my example name:
# ...
exporters:
sumologic:
source_name: my example name
A description can be added as a comment just above the receiver name.
For example, the following snippet configures the description as All my example logs
:
receivers:
## All my example logs
filelog/my example name:
# ...
exporters:
sumologic:
source_name: my example name
Like the Installed Collector, the OpenTelemetry Collector supports regular expression for paths.
In addition, you can specify multiple different path expressions.
Add them as elements of the include
configuration option.
For example, the following snippet configures the path to all .log
files from /var/log/
and /opt/my_app/
:
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
# ...
exporters:
sumologic:
source_name: my example name
The OpenTelemetry Collector doesn't have a substitute for this Installed Collector option.
It supports two options, starting at the beginning or end of a file.
Starting at the beginning
will read the entire file every time it's started.
Starting at the end
will read only logs appended to file after it's started.
This is configurable with the start_at
option.
For example, the following snippet configures the Collector to only read appended logs:
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
start_at: end
# ...
exporters:
sumologic:
source_name: my example name
The Source Host is set in the exporter configuration with the source_host
option.
For example, the following snippet configures the Source Host as My Host
:
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
start_at: end
# ...
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
The Source Category is set in the exporter configuration with the source_category
option.
For example, the following snippet configures the Source Category as My Category
:
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
start_at: end
# ...
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
Use the resourceprocessor to set custom fields.
For example, the following snippet configures two fields, cloud.availability_zone
and k8s.cluster.name
:
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
start_at: end
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
Use the exclude
option in the filelog receiver to specify files you don't want collected.
For example, the following snippet excludes /var/log/sensitive.log
from collection:
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
exclude:
- /var/log/sensitive.log
start_at: end
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
The Installed Collector option to Extract timestamp information from log file entries
in an
OpenTelemtry configuration is clear_logs_timestamp
. This is set to true
by default.
This works like Extract timestamp information from log file entries
combined with
Ignore time zone from log file and instead use:
set to Use Collector Default
.
For example, the following configuration sets the time_zone for a Collector with extensions.sumologic.time_zone
:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
exclude:
- /var/log/sensitive.log
start_at: end
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
If clear_logs_timestamp
is set to false
, timestamp parsing should be configured
manually, like in the following snippet:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
exclude:
- /var/log/sensitive.log
start_at: end
operators:
## Extract timestamp into timestamp field using regex
## rel: https://github.com/sumo-drosiek/opentelemetry-log-collection/blob/b506aadf913d6c1691cef10a534d495338c87dee/docs/operators/regex_parser.md
- type: regex_parser
regex: (?P<timestamp>^\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2},\d{3} (\+|\-)\d{4})
## Keep original record in log field
preserve_to: body.log
## Parse timestamp from timestamp field
## rel: https://github.com/sumo-drosiek/opentelemetry-log-collection/blob/b506aadf913d6c1691cef10a534d495338c87dee/docs/operators/time_parser.md
timestamp:
parse_from: body.timestamp
## Layout are substitute for Timestamp Format configuration
layout_type: gotime
layout: '2006-01-02 15:04:05,000 -0700'
## Restore record from log field
## rel: https://github.com/sumo-drosiek/opentelemetry-log-collection/blob/b506aadf913d6c1691cef10a534d495338c87dee/docs/operators/move.md
- type: move
from: body.log
to: body
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
## Keep manually parsed timestamps
clear_logs_timestamp: true
The following example snippet skips timestamp parsing so the Collector uses Receipt Time:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
exclude:
- /var/log/sensitive.log
start_at: end
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
## Keep manually parsed timestamps (use Receipt Time by default)
clear_logs_timestamp: true
Use encoding
to set the encoding of your data. Full list of supporter encodings can be obtained from filelogreceiver documentation.
The following snippet sets the encoding to UTF-8:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
exclude:
- /var/log/sensitive.log
start_at: end
encoding: utf-8
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
Multiline processing in the Opentelemetry Collector is set manually. There is no automatic boundary detection.
The following snippet sets the boundary regex as ^\d{4}-\d{2}-\d{2}
to match, for example, 2021-06-06
):
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
## All my example logs
filelog/my example name:
include:
- /var/log/*.log
- /opt/my_app/*.log
exclude:
- /var/log/sensitive.log
start_at: end
encoding: utf-8
multiline:
line_start_pattern: ^\d{4}-\d{2}-\d{2}
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: My Host
source_category: My Category
If your multiline logs have a known end pattern use the line_end_pattern
option.
More information is available in filelogreceiver documentation.
Remote File Source is not supported by the Opentelemetry Collector.
The equivalent of the Syslog Source is a combination of the tcplog or the udplog receivers and the sumologicsyslog processor.
Note: The OpenTelemetry Collector provides also Syslog Receiver. See this document for details.
Below is an example of an OpenTelemetry configuration for a Syslog Source.
extensions:
sumologic:
install_token: <install_token>
## Time Zone is a substitute of Installed Collector `Time Zone`
## with `Use time zone from log file. If none is detected use:` option.
## This is used only if `clear_logs_timestamp` is set to `true` in sumologic exporter.
## Full list of time zones is available on wikipedia:
## https://en.wikipedia.org/wiki/List_of_tz_database_time_zones#List
time_zone: America/Tijuana
receivers:
## Use tcpreceiver for TCP protocol
tcpreceiver/first receiver:
## listen address in format host:port
## host 0.0.0.0 mean all network interfaces
listen_address: 0.0.0.0:514
## Add network attributes
## `net.peer.name` is going to be used as exporters.sumologic.source_host
## rel: https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/v0.33.0/receiver/tcplogreceiver#configuration
add_attributes: true
## Use udpreceiver for UDP protocol
udpreceiver/first receiver:
## listen address in format host:port
## host 0.0.0.0 mean all network interfaces
listen_address: 0.0.0.0:514
## Add network attributes
## `net.peer.name` is going to be used as exporters.sumologic.source_host
## rel: https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/v0.33.0/receiver/udplogreceiver#configuration
add_attributes: true
processors:
## There is no substitute for `Description` in current project phase.
## It is recommended to use comments for that purpose, like this one.
## sumologic_syslog/<source group name>:
## <source group name> can be substitute of Installed Collector `Name`.
sumologic_syslog/syslog source:
## The following configuration will add two fields to every record
resource/syslog source:
attributes:
- key: cloud.availability_zone
value: zone-1
action: insert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/syslog:
## Installed Collector substitute for `Source Category`.
source_category: example category
## clear_logs_timestamp is by default set to True.
## If it's set to true, it works like `Enable Timestamp Parsing`,
## and `Time Zone` is going to be taken from `extensions` section.
## There is no possibility to configure several time zones in one exporter.
## clear_logs_timestamp sets to true also behaves like
## `Timestamp Format` would be set to `Automatically detect the format`
## in terms of Installed Collector configuration.
clear_logs_timestamp: true
## Set Source Name to be facility name
source_name: "%{facility}"
## Set Source Host to `net.peer.name`
source_host: "%{net.peer.name}
service:
extensions:
- sumologic
pipelines:
logs/syslog source:
receivers:
- filelog/syslog source
processors:
- resource/syslog source
exporters:
- sumologic/syslog
Define the name after the slash /
in the processor name.
For example, the following snippet configures the name as my example name
:
processor:
sumologic_syslog/my example name:
# ...
A description can be added as a comment just above the processor name.
For example, the following snippet configures the description as All my example logs
:
processor:
## All my example logs
sumologic_syslog/my example name:
# ...
Protocol is defined by receiver type. For UDP use udplogreceiver and for TCP use tcplogreceiver. Port can be set by listen_address
, for example to listen on port 6776
on all interfaces, use listen_address: 0.0.0.0:6776
.
You can use multiple receivers with different names nad ports like in the following example:
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
tcpreceiver/second receiver:
listen_address: 127.0.0.1:5140
udpreceiver/first receiver:
listen_address: 0.0.0.0:514
udpreceiver/second receiver:
listen_address: 127.0.0.1:5150
processor:
## All my example logs
sumologic_syslog/my example name:
# ...
A Source Category can be set in the exporter configuration with the source_category
option.
For example, the following snippet configures the Source Category as My Category
:
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
tcpreceiver/second receiver:
listen_address: 127.0.0.1:5140
udpreceiver/first receiver:
listen_address: 0.0.0.0:514
udpreceiver/second receiver:
listen_address: 127.0.0.1:5150
processor:
## All my example logs
sumologic_syslog/my example name:
# ...
exporters:
sumologic/some name:
source_category: My Category
Use the resourceprocessor to set custom fields.
For example, the following snippet configures two fields, cloud.availability_zone
and k8s.cluster.name
:
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
tcpreceiver/second receiver:
listen_address: 127.0.0.1:5140
udpreceiver/first receiver:
listen_address: 0.0.0.0:514
udpreceiver/second receiver:
listen_address: 127.0.0.1:5150
processors:
## All my example logs
sumologic_syslog/my example name:
# ...
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_category: My Category
The Installed Collector option to Extract timestamp information from log file entries
in an
OpenTelemtry configuration is clear_logs_timestamp
. This is set to true
by default.
This works like Extract timestamp information from log file entries
combined with
Ignore time zone from log file and instead use:
set to Use Collector Default
.
For example, the following configuration sets the time_zone for a Collector with extensions.sumologic.time_zone
:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
tcpreceiver/second receiver:
listen_address: 127.0.0.1:5140
udpreceiver/first receiver:
listen_address: 0.0.0.0:514
udpreceiver/second receiver:
listen_address: 127.0.0.1:5150
processors:
## All my example logs
sumologic_syslog/my example name:
# ...
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_category: My Category
If clear_logs_timestamp
is set to false
, timestamp parsing should be configured
manually, like in the following snippet:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
operators:
## Extract timestamp into timestamp field using regex
## rel: https://github.com/sumo-drosiek/opentelemetry-log-collection/blob/b506aadf913d6c1691cef10a534d495338c87dee/docs/operators/regex_parser.md
- type: regex_parser
regex: (?P<timestamp>^\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2},\d{3} (\+|\-)\d{4})
## Keep original record in log field
preserve_to: body.log
## Parse timestamp from timestamp field
## rel: https://github.com/sumo-drosiek/opentelemetry-log-collection/blob/b506aadf913d6c1691cef10a534d495338c87dee/docs/operators/time_parser.md
timestamp:
parse_from: body.timestamp
## Layout are substitute for Timestamp Format configuration
layout_type: gotime
layout: '2006-01-02 15:04:05,000 -0700'
## Restore record from log field
## rel: https://github.com/sumo-drosiek/opentelemetry-log-collection/blob/b506aadf913d6c1691cef10a534d495338c87dee/docs/operators/move.md
- type: move
from: body.log
to: body
processors:
## All my example logs
sumologic_syslog/my example name:
# ...
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_category: My Category
## Keep manually parsed timestamps
clear_logs_timestamp: true
The following example snippet skips timestamp parsing so the Collector uses Receipt Time:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
processors:
## All my example logs
sumologic_syslog/my example name:
# ...
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_category: My Category
## Keep manually parsed timestamps
clear_logs_timestamp: true
The OpenTelemetry Collector requires the Source Name to be set manually.
In the exporter configuration, use the Sumologicsyslogprocessor
to set the facility
attribute.
For example:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
processors:
## All my example logs
sumologic_syslog/my example name:
# ...
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_category: My Category
## Keep manually parsed timestamps
clear_logs_timestamp: true
## Set Source Name to facility, which is set by sumologicsyslogprocessor
source_name: "%{facility}
The OpenTelemetry Collector requires the Source Host to be set manually.
Set add_attributes
to true
for tcplogreceiver/udplogreceiver.
This adds connection related attributes,
especially net.peer.name
which should be set as the Source Host.
For example:
extensions:
sumologic:
time_zone: America/Tijuana
receivers:
tcpreceiver/first receiver:
listen_address: 0.0.0.0:514
add_attributes: true
processors:
## All my example logs
sumologic_syslog/my example name:
# ...
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_category: My Category
## Keep manually parsed timestamps
clear_logs_timestamp: true
## Set Source Name to facility, which is set by sumologicsyslogprocessor
source_name: "%{facility}
source_host: "%{net.peer.name}
Docker Logs Source is not supported by the OpenTelemetry Collector.
Docker Stats Source is not supported by the OpenTelemetry Collector.
Script Source is not supported by the OpenTelemetry Collector.
For the Streaming Metrics Source we are using the Telegraf receiver with socket_listener plugin.
Below is an example of an OpenTelemetry configuration for a Streaming Metrics Source.
extensions:
sumologic:
install_token: <install_token>
## Time Zone is a substitute of Installed Collector `Time Zone`
## Full list of time zones is available on wikipedia:
## https://en.wikipedia.org/wiki/List_of_tz_database_time_zones#List
time_zone: America/Tijuana
receivers:
## There is no substitute for `Description` in current project phase.
## It is recommended to use comments for that purpose, like this one.
## telegraf/<source group name>:
## <source group name> can be substitute of Installed Collector `Name`.
telegraf/metrics source:
## Do not add metric field separately as data point label.
separate_field: false
## Telegraf configuration
agent_config: |
[agent]
## Get metrics every 15 seconds
interval = "15s"
## Flush metrics every 15 seconds
flush_interval = "15s"
## socket_listener listen on given protocol://hostname:port for metrics
[[inputs.socket_listener]]
## listen for metrics on UDP port 2006 on localhost
service_address = "udp://localhost:2006"
## Get metrics in carbon2 format
data_format = "carbon2"
## Add additional metadata
[inputs.socket_listener.tags]
_contentType = "Carbon2"
_primaryMetricType = "carbon"
processors:
## The following configuration will add two metadata properties to every record
resource/metric source:
attributes:
- key: cloud.availability_zone
value: zone-1
action: insert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic:
## Set _sourceName
source_name: my example name
## Installed Collector substitute for `Source Category`.
source_category: example category
## Installed Collector substitute for `Source Host`.
source_host: example host
service:
extensions:
- sumologic
pipelines:
metrics/metric source:
receivers:
- telegraf/metrics source
processors:
- resource/metric source
exporters:
- sumologic
Define the name after the slash /
in the receiver name.
To set _sourceName
, use resourceprocessor
or set it in sumologicexporter.
For example, the following snippet configures the name as my example name
:
receivers:
telegraf/my example name:
# ...
exporters:
sumologic:
source_name: my example name
A description can be added as a comment just above the receiver name.
For example, the following snippet configures the description as All my example logs
:
receivers:
## All my example metrics
telegraf/my example name:
# ...
exporters:
sumologic:
source_name: my example name
Protocol and Port can be configured using service_address
in Telegraf socket_listener
plugin configuration.
For example:
receivers:
## All my example metrics
telegraf/my example name:
## Telegraf configuration
agent_config: |
## socket_listener listen on given protocol://hostname:port for metrics
[[inputs.socket_listener]]
## listen for metrics on UDP port 2006 on localhost
service_address = "udp://localhost:2006"
## Get metrics in carbon2 format
data_format = "carbon2"
## Add additional metadata
[inputs.socket_listener.tags]
_contentType = "Carbon2"
_primaryMetricType = "carbon"
# ...
exporters:
sumologic:
source_name: my example name
Content Type can be configured using data_format
in the Telegraf socket_listener
plugin configuration.
Any of the available formats can be used, especially graphite
and carbon2
.
For example:
receivers:
## All my example metrics
telegraf/my example name:
## Telegraf configuration
agent_config: |
## socket_listener listen on given protocol://hostname:port for metrics
[[inputs.socket_listener]]
## listen for metrics on UDP port 2006 on localhost
service_address = "udp://localhost:2006"
## Get metrics in carbon2 format
data_format = "carbon2"
# ...
exporters:
sumologic:
source_name: my example name
A Source Category can be set in the exporter configuration with the source_category
option.
For example, the following snippet configures the Source Category as My Category
:
receivers:
## All my example metrics
telegraf/my example name:
## Telegraf configuration
agent_config: |
## socket_listener listen on given protocol://hostname:port for metrics
[[inputs.socket_listener]]
## listen for metrics on UDP port 2006 on localhost
service_address = "udp://localhost:2006"
## Get metrics in carbon2 format
data_format = "carbon2"
exporters:
sumologic/some name:
source_name: my example name
source_category: My Category
Use the resourceprocessor to set custom metadata.
For example, the following snippet configures two additional metadata properties,
cloud.availability_zone
and k8s.cluster.name
:
receivers:
## All my example metrics
telegraf/my example name:
## Telegraf configuration
agent_config: |
## socket_listener listen on given protocol://hostname:port for metrics
[[inputs.socket_listener]]
## listen for metrics on UDP port 2006 on localhost
service_address = "udp://localhost:2006"
## Get metrics in carbon2 format
data_format = "carbon2"
processors:
# ...
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_category: My Category
The equivalent of the Host Metrics Source is the telegraf receiver with appropiate plugins.
Note: The are differences between the Installed Collector and the Openelemetry Collector host metrics. See this document to learn more.
Below is an example of an OpenTelemetry configuration for a Host Metrics Source.
extensions:
sumologic:
install_token: <install_token>
## Time Zone is a substitute of Installed Collector `Time Zone`
## Full list of time zones is available on wikipedia:
## https://en.wikipedia.org/wiki/List_of_tz_database_time_zones#List
time_zone: America/Tijuana
receivers:
## There is no substitute for `Description` in current project phase.
## It is recommended to use comments for that purpose, like this one.
## telegraf/<source group name>:
## <source group name> can be substitute of Installed Collector `Name`.
telegraf/metrics source:
## Do not add metric field separately as data point label.
separate_field: false
## Telegraf configuration
agent_config: |
[agent]
## Get metrics every 15 seconds
interval = "15s"
## Flush metrics every 15 seconds
flush_interval = "15s"
## CPU metrics
[[inputs.cpu]]
percpu = false
totalcpu = true
collect_cpu_time = false
report_active = true
namepass = [ "cpu" ]
fieldpass = [ "usage_active", "usage_steal", "usage_iowait", "usage_irq", "usage_user", "usage_idle", "usage_nice", "usage_system", "usage_softirq" ]
## CPU metrics
[[inputs.system]]
namepass = [ "system" ]
fieldpass = [ "load1", "load5", "load15" ]
## Memory metrics
[[inputs.mem]]
fieldpass = [ "total", "free", "used", "used_percent", "available", "available_percent" ]
## TCP metrics
[[inputs.netstat]]
fieldpass = [ "tcp_close", "tcp_close_wait", "tcp_closing", "tcp_established", "tcp_listen", "tcp_time_wait" ]
## Network metrics
[[inputs.net]]
interfaces = ["eth*", "en*", "lo*"]
ignore_protocol_stats = true
fieldpass = [ "bytes_sent", "bytes_recv", "packets_sent", "packets_recv" ]
## Disk metrics
[[inputs.disk]]
namepass = [ "disk" ]
fieldpass = [ "used", "used_percent", "inodes_free" ]
ignore_fs = ["tmpfs", "devtmpfs", "devfs", "iso9660", "overlay", "aufs", "squashfs"]
## Disk metrics
[[inputs.diskio]]
fieldpass = [ "reads", "read_bytes", "writes", "write_bytes" ]
processors:
## The following configuration will add two metadata properties to every record
resource/metric source:
attributes:
- key: cloud.availability_zone
value: zone-1
action: insert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic:
## Set _sourceName
source_name: my example name
## Installed Collector substitute for `Source Category`.
source_category: example category
## Installed Collector substitute for `Source Host`.
source_host: example host
## Ensure compability with Installed Colllector metric name
translate_telegraf_attributes: true
service:
extensions:
- sumologic
pipelines:
metrics/metric source:
receivers:
- telegraf/metrics source
processors:
- resource/metric source
exporters:
- sumologic
Define the name after the slash /
in the receiver name.
To set _sourceName
, use resourceprocessor
or set it in sumologicexporter.
For example, the following snippet configures the name as my example name
:
receivers:
telegraf/my example name:
# ...
exporters:
sumologic:
source_name: my example name
A description can be added as a comment just above the receiver name.
For example, the following snippet configures the description as All my example logs
:
receivers:
## All my example metrics
telegraf/my example name:
# ...
exporters:
sumologic:
source_name: my example name
A Source Host can be set in the exporter configuration with the source_host
option.
For example, the following snippet configures the Source Host as my_host
:
receivers:
## All my example metrics
telegraf/my example name:
# ...
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
A Source Category can be set in the exporter configuration with the source_category
option.
For example, the following snippet configures the Source Category as My Category
:
receivers:
## All my example metrics
telegraf/my example name:
# ...
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
Use the resourceprocessor to set custom metadata.
For example, the following snippet configures two additional metadata properties,
cloud.availability_zone
and k8s.cluster.name
:
receivers:
## All my example metrics
telegraf/my example name:
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
To set Scan Interval use interval
in Telegraf's agent configuration.
The following example shows how to set it for 1 minute:
receivers:
## All my example metrics
telegraf/my example name:
agent_config: |
[agent]
interval = "1m"
flush_interval = "1m"
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
Telegraf offers a set of various plugins you can use to get metrics. In this section, we are describing only plugins that are required for seamless migration from the Installed Collector. If you are interested in other metrics, see list of Telegraf input plugins.
Each of the subtopics contain a table that describes how Installed Collector metrics translate to Telegraf metrics.
To ensure all dashboards are working as before,
Telegraf metric names are translated to the Installed Collector by sumologicexporter.
You can disable this by setting translate_telegraf_attributes
to false
,
but in this case you need to update your dashboards.
To get CPU metrics we are using the inputs.cpu and the inputs.system Telegraf plugins.
Metric Name | Telegraf plugin | Telegraf metric name |
---|---|---|
CPU_User | inputs.cpu | cpu_usage_user |
CPU_Sys | inputs.cpu | cpu_usage_System |
CPU_Nice | inputs.cpu | cpu_usage_nice |
CPU_Idle | inputs.cpu | cpu_usage_idle |
CPU_IOWait | inputs.cpu | cpu_usage_iowait |
CPU_Irq | inputs.cpu | cpu_usage_irq |
CPU_SoftIrq | inputs.cpu | cpu_usage_softirq |
CPU_Stolen | inputs.cpu | cpu_usage_steal |
CPU_LoadAvg_1min | inputs.system | system_load1 |
CPU_LoadAvg_5min | inputs.system | system_load5 |
CPU_LoadAvg_15min | inputs.system | system_load15 |
CPU_Total | inputs.cpu | cpu_usage_active |
The following example shows the desired configuration:
receivers:
## All my example metrics
telegraf/my example name:
agent_config: |
[agent]
interval = "1m"
flush_interval = "1m"
## CPU metrics
[[inputs.cpu]]
percpu = false
totalcpu = true
collect_cpu_time = false
report_active = true
namepass = [ "cpu" ]
fieldpass = [ "usage_active", "usage_steal", "usage_iowait", "usage_irq", "usage_user", "usage_idle", "usage_nice", "usage_system", "usage_softirq" ]
## CPU metrics
[[inputs.system]]
namepass = [ "system" ]
fieldpass = [ "load1", "load5", "load15" ]
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
To get CPU metrics we are using the inputs.mem Telegraf plugin.
Metric Name | Telegraf plugin | Telegraf metric name |
---|---|---|
Mem_Total | inputs.mem | mem_total |
Mem_Used | N/A | N/A |
Mem_Free | inputs.mem | mem_free |
Mem_ActualFree | inputs.mem | mem_available |
Mem_ActualUsed | inputs.mem | mem_used |
Mem_UsedPercent | inputs.mem | mem_used_percent |
Mem_FreePercent | inputs.mem | mem_available_percent |
Mem_PhysicalRam | N/A | N/A |
The following example shows the desired configuration:
receivers:
## All my example metrics
telegraf/my example name:
agent_config: |
[agent]
interval = "1m"
flush_interval = "1m"
## CPU metrics
[[inputs.cpu]]
percpu = false
totalcpu = true
collect_cpu_time = false
report_active = true
namepass = [ "cpu" ]
fieldpass = [ "usage_active", "usage_steal", "usage_iowait", "usage_irq", "usage_user", "usage_idle", "usage_nice", "usage_system", "usage_softirq" ]
## CPU metrics
[[inputs.system]]
namepass = [ "system" ]
fieldpass = [ "load1", "load5", "load15" ]
## Memory metrics
[[inputs.mem]]
fieldpass = [ "total", "free", "used", "used_percent", "available", "available_percent" ]
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
To get TCP metrics we are using the inputs.netstat Telegraf plugin.
Metric Name | Telegraf plugin | Telegraf metric name |
---|---|---|
TCP_InboundTotal | N/A | N/A |
TCP_OutboundTotal | N/A | N/A |
TCP_Established | inputs.netstat | netstat_tcp_established |
TCP_Listen | inputs.netstat | netstat_tcp_listen |
TCP_Idle | N/A | N/A |
TCP_Closing | inputs.netstat | netstat_tcp_closing |
TCP_CloseWait | inputs.netstat | netstat_tcp_close_wait |
TCP_Close | inputs.netstat | netstat_tcp_close |
TCP_TimeWait | inputs.netstat | netstat_tcp_time_wait |
The following example shows the desired configuration:
receivers:
## All my example metrics
telegraf/my example name:
agent_config: |
[agent]
interval = "1m"
flush_interval = "1m"
## CPU metrics
[[inputs.cpu]]
percpu = false
totalcpu = true
collect_cpu_time = false
report_active = true
namepass = [ "cpu" ]
fieldpass = [ "usage_active", "usage_steal", "usage_iowait", "usage_irq", "usage_user", "usage_idle", "usage_nice", "usage_system", "usage_softirq" ]
## CPU metrics
[[inputs.system]]
namepass = [ "system" ]
fieldpass = [ "load1", "load5", "load15" ]
## Memory metrics
[[inputs.mem]]
fieldpass = [ "total", "free", "used", "used_percent", "available", "available_percent" ]
## TCP metrics
[[inputs.netstat]]
fieldpass = [ "tcp_close", "tcp_close_wait", "tcp_closing", "tcp_established", "tcp_listen", "tcp_time_wait" ]
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
To get network metrics we are using the inputs.net Telegraf plugin.
Metric Name | Telegraf plugin | Telegraf metric name |
---|---|---|
Net_InPackets | inputs.net | net_packets_recv |
Net_OutPackets | inputs.net | net_packets_sent |
Net_InBytes | inputs.net | net_bytes_recv |
Net_OutBytes | inputs.net | net_bytes_sent |
The following example shows the desired configuration:
receivers:
## All my example metrics
telegraf/my example name:
agent_config: |
[agent]
interval = "1m"
flush_interval = "1m"
## CPU metrics
[[inputs.cpu]]
percpu = false
totalcpu = true
collect_cpu_time = false
report_active = true
namepass = [ "cpu" ]
fieldpass = [ "usage_active", "usage_steal", "usage_iowait", "usage_irq", "usage_user", "usage_idle", "usage_nice", "usage_system", "usage_softirq" ]
## CPU metrics
[[inputs.system]]
namepass = [ "system" ]
fieldpass = [ "load1", "load5", "load15" ]
## Memory metrics
[[inputs.mem]]
fieldpass = [ "total", "free", "used", "used_percent", "available", "available_percent" ]
## TCP metrics
[[inputs.netstat]]
fieldpass = [ "tcp_close", "tcp_close_wait", "tcp_closing", "tcp_established", "tcp_listen", "tcp_time_wait" ]
## Network metrics
[[inputs.net]]
interfaces = ["eth*", "en*", "lo*"]
ignore_protocol_stats = true
fieldpass = [ "bytes_sent", "bytes_recv", "packets_sent", "packets_recv" ]
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
To get disk metrics we are using the inputs.diskio and the inputs.disk Telegraf plugins.
Metric Name | Telegraf plugin | Telegraf metric name |
---|---|---|
Disk_Reads | inputs.diskio | diskio_reads |
Disk_ReadBytes | inputs.diskio | diskio_read_bytes |
Disk_Writes | inputs.diskio | diskio_writes |
Disk_WriteBytes | inputs.diskio | diskio_write_bytes |
Disk_Queue | N/A | N/A |
Disk_InodesAvailable | inputs.disk | disk_inodes_free |
Disk_Used | inputs.disk | disk_used |
Disk_UsedPercent | inputs.disk | disk_used_percent |
Disk_Available | N/A | N/A |
The following example shows the desired configuration:
receivers:
## All my example metrics
telegraf/my example name:
agent_config: |
[agent]
interval = "1m"
flush_interval = "1m"
## CPU metrics
[[inputs.cpu]]
percpu = false
totalcpu = true
collect_cpu_time = false
report_active = true
namepass = [ "cpu" ]
fieldpass = [ "usage_active", "usage_steal", "usage_iowait", "usage_irq", "usage_user", "usage_idle", "usage_nice", "usage_system", "usage_softirq" ]
## CPU metrics
[[inputs.system]]
namepass = [ "system" ]
fieldpass = [ "load1", "load5", "load15" ]
## Memory metrics
[[inputs.mem]]
fieldpass = [ "total", "free", "used", "used_percent", "available", "available_percent" ]
## TCP metrics
[[inputs.netstat]]
fieldpass = [ "tcp_close", "tcp_close_wait", "tcp_closing", "tcp_established", "tcp_listen", "tcp_time_wait" ]
## Network metrics
[[inputs.net]]
interfaces = ["eth*", "en*", "lo*"]
ignore_protocol_stats = true
fieldpass = [ "bytes_sent", "bytes_recv", "packets_sent", "packets_recv" ]
## Disk metrics
[[inputs.disk]]
namepass = [ "disk" ]
fieldpass = [ "used", "used_percent", "inodes_free" ]
ignore_fs = ["tmpfs", "devtmpfs", "devfs", "iso9660", "overlay", "aufs", "squashfs"]
## Disk metrics
[[inputs.diskio]]
fieldpass = [ "reads", "read_bytes", "writes", "write_bytes" ]
# ...
processors:
resource/my example name fields:
attributes:
- key: cloud.availability_zone
value: zone-1
action: upsert
- key: k8s.cluster.name
value: my-cluster
action: insert
exporters:
sumologic/some name:
source_name: my example name
source_host: my_host
source_category: My Category
Local Windows Event Log Source is not supported by the OpenTelemetry Collector.
Local Windows Performance Monitor Log Source is not supported by the OpenTelemetry Collector.
Windows Active Directory Source is not supported by the OpenTelemetry Collector.
Script Action is not supported by the OpenTelemetry Collector.
This section describes migration steps for an Installed Collector managed with a Local Configuration File.
The following table shows the equivalent user.properties for OpenTelemetry.
user.properties key | The OpenTelemetry Collector Key |
---|---|
wrapper.java.command=JRE Bin Location |
N/A |
accessid=accessId |
N/A, use extensions.sumologic.install_token |
accesskey=accessKey |
N/A, use extensions.sumologic.install_token |
category=category |
extensions.sumologic.collector_category |
clobber=true/false |
extensions.sumologic.clobber |
description=description |
extensions.sumologic.collector_description |
disableActionSource=true/false |
N/A |
disableScriptSource=true/false |
N/A |
disableUpgrade=true/false |
N/A |
enableActionSource=true/false |
N/A |
enableScriptSource=true/false |
N/A |
ephemeral=true/false |
N/A |
fields=[list of fields] |
processors.resource |
fipsJce=true/false |
N/A |
hostName=hostname |
exporters.sumologic.source_host |
name=name |
extensions.sumologic.collector_name |
proxyHost=host |
plese see OTC documentation |
proxyNtlmDomain=NTLM domain |
plese see OTC documentation |
proxyPassword=password |
plese see OTC documentation |
proxyPort=port |
plese see OTC documentation |
proxyUser=username |
plese see OTC documentation |
skipAccessKeyRemoval=true/false |
N/A |
sources=absolute filepath or folderpath |
N/A |
syncSources=absolute filepath or folderpath |
N/A |
targetCPU=target |
N/A |
timeZone=timezone |
extensions.sumologic.time_zone |
token=token |
N/A |
url=collection endpoint |
extensions.sumologic.api.base.url |
wrapper.java.command=JRE Bin Location |
N/A |
wrapper.java.command=JRE Bin Location |
N/A |
wrapper.java.maxmemory=size |
N/A |
This section describes migration steps for common parameters.
sourceType
migration:
- LocalFile
- RemoteFileV2
- Syslog
- DockerLog
- DockerStats
- Script
- StreamingMetrics
- SystemStats
- LocalWindowsEventLog
- RemoteWindowsEventLog
- LocalWindowsPerfMon
- RemoteWindowsPerfMon
- ActiveDirectory
The Installed Collector Parameter | The OpenTelemetry Collector Key |
---|---|
name |
exporters.sumologic.source_name |
description |
A description can be added as a comment just above the receiver name. See the linked example. |
fields |
Use the resourceprocessor to set custom fields. See the linked example. |
hostName |
exporters.sumologic.source_host; See the linked example. |
category |
exporters.sumologic.source_category |
automaticDateParsing |
See Timestamp Parsing explanation |
timeZone |
See Timestamp Parsing explanation |
forceTimeZone |
See Timestamp Parsing explanation |
defaultDateFormat |
See Timestamp Parsing explanation |
defaultDateFormats |
See Timestamp Parsing explanation |
multilineProcessingEnabled |
See Multiline Processing explanation |
useAutolineMatching |
See Multiline Processing explanation |
manualPrefixRegexp |
See Multiline Processing explanation |
filters |
N/A |
cutoffTimestamp |
N/A |
cutoffRelativeTime |
N/A |
The equivalent of the Local File Source is the filelog receiver. More useful information can be found in Local File Source for Cloud Based Management.
The Installed Collector Parameter | The OpenTelemetry Collector Key |
---|---|
pathExpression |
element of receivers.filelog.include list |
denylist |
elemets of receivers.filelog.exclude list |
encoding |
receivers.filelog.encoding |
Remote File Source is not supported by the OpenTelemetry Collector.
The equivalent of the Syslog Source is a combination of the tcplog or the udplog receivers and the sumologicsyslog processor. More useful information can be found in Syslog Source for Cloud Based Management.
Note: The OpenTelemetry Collector provides also Syslog Receiver. See this document for details.
The Installed Collector Parameter | The OpenTelemetry Collector Key |
---|---|
protocol |
using tcplog or udplog receiver. See syslog explanation |
port |
receivers.tcplog.listen_address or receivers.udplog.listen_address . See syslog explanation |
Docker Logs Source is not supported by the OpenTelemetry Collector.
Docker Stats Source is not supported by the OpenTelemetry Collector.
Script Source is not supported by the OpenTelemetry Collector.
The equivalent of the Streaming Metrics Source is the telegraf receiver with appropiate plugins. More useful information can be found in Streaming Metrics Source for Cloud Based Management.
The Installed Collector Parameter | The OpenTelemetry Collector Key |
---|---|
name |
exporters.sumologic.source_name |
description |
A description can be added as a comment just above the receiver name. See the linked example. |
category |
exporters.sumologic.source_category |
contentType |
receivers.telegraf.agent_config('inputs.socket_listener'.data_format) |
protocol |
receivers.telegraf.agent_config('inputs.socket_listener'.service_address) |
port |
receivers.telegraf.agent_config('inputs.socket_listener'.service_address) |
The equivalent of the Host Metrics Source is the telegraf receiver with appropiate plugins. More useful information can be found in Host Metrics Source for Cloud Based Management.
Note: The are differences between the Installed Collector and the Openelemetry Collector host metrics. See this document to learn more.
The Installed Collector Parameter | The OpenTelemetry Collector Key |
---|---|
name |
exporters.sumologic.source_name |
description |
A description can be added as a comment just above the receiver name. See the linked example. |
category |
exporters.sumologic.source_category |
metrics |
Appropiate plugins have to be configured. By default no metrics are being processed. |
interval (ms) |
receivers.telegraf.agent_config('agent'.interval) |
hostName |
exporters.sumologic.source_host |
Local Windows Event Log Source is not supported by the OpenTelemetry Collector.
Remote Windows Event Log Source is not supported by the OpenTelemetry Collector.
Local Windows Performance Source is not supported by the OpenTelemetry Collector.
Remote Windows Performance Source is not supported by the OpenTelemetry Collector.
Windows Active Directory Source is not supported by the OpenTelemetry Collector.