目录
强大的性能,无限的扩展能力
收集、组织和处理海量高速数据。当您将任何数据视为时间序列数据时,它都会更有价值。借助 InfluxDB,这款排名第一的时间序列平台旨在与 Telegraf 一起扩展。
查看入门方法
输入和输出集成概述
此插件从重要的 OpenStack 服务收集指标,从而促进云基础设施的监控和管理。
OpenSearch 输出插件允许用户使用 HTTP 将指标直接发送到 OpenSearch 实例,从而促进 OpenSearch 生态系统内有效的数据管理和分析。
集成详情
OpenStack
OpenStack 插件允许用户从各种 OpenStack 服务(如 CINDER、GLANCE、HEAT、KEYSTONE、NEUTRON 和 NOVA)收集性能指标。它支持多个 OpenStack API 来获取与这些服务相关的关键指标,从而实现对云资源的全面监控和管理。随着组织越来越多地采用 OpenStack 作为其云基础设施,此插件在提供对整个云环境中的资源使用情况、可用性和性能的洞察方面发挥着至关重要的作用。配置选项允许自定义轮询间隔和过滤不需要的标签,以优化性能和基数。
OpenSearch
OpenSearch Telegraf 插件通过 HTTP 与 OpenSearch 数据库集成,从而实现指标的简化收集和存储。作为一个专为 2.x 及更高版本的 OpenSearch 版本设计的强大工具,该插件提供了强大的功能,同时通过原始 Elasticsearch 插件提供与 1.x 的兼容性。此插件有助于在 OpenSearch 中创建和管理索引,自动管理模板并确保数据结构化以进行有效分析。该插件支持各种配置选项,例如索引名称、身份验证、运行状况检查和值处理,使其可以根据不同的操作要求进行定制。其功能使其对于希望利用 OpenSearch 的强大功能进行指标存储和查询的组织至关重要。
配置
OpenStack
[[inputs.openstack]]
## The recommended interval to poll is '30m'
## The identity endpoint to authenticate against and get the service catalog from.
authentication_endpoint = "https://my.openstack.cloud:5000"
## The domain to authenticate against when using a V3 identity endpoint.
# domain = "default"
## The project to authenticate as.
# project = "admin"
## User authentication credentials. Must have admin rights.
username = "admin"
password = "password"
## Available services are:
## "agents", "aggregates", "cinder_services", "flavors", "hypervisors",
## "networks", "nova_services", "ports", "projects", "servers",
## "serverdiagnostics", "services", "stacks", "storage_pools", "subnets",
## "volumes"
# enabled_services = ["services", "projects", "hypervisors", "flavors", "networks", "volumes"]
## Query all instances of all tenants for the volumes and server services
## NOTE: Usually this is only permitted for administrators!
# query_all_tenants = true
## output secrets (such as adminPass(for server) and UserID(for volume)).
# output_secrets = false
## Amount of time allowed to complete the HTTP(s) request.
# timeout = "5s"
## HTTP Proxy support
# http_proxy_url = ""
## Optional TLS Config
# tls_ca = /path/to/cafile
# tls_cert = /path/to/certfile
# tls_key = /path/to/keyfile
## Use TLS but skip chain & host verification
# insecure_skip_verify = false
## Options for tags received from Openstack
# tag_prefix = "openstack_tag_"
# tag_value = "true"
## Timestamp format for timestamp data received from Openstack.
## If false format is unix nanoseconds.
# human_readable_timestamps = false
## Measure Openstack call duration
# measure_openstack_requests = false
OpenSearch
[[outputs.opensearch]]
## URLs
## The full HTTP endpoint URL for your OpenSearch instance. Multiple URLs can
## be specified as part of the same cluster, but only one URLs is used to
## write during each interval.
urls = ["http://node1.os.example.com:9200"]
## Index Name
## Target index name for metrics (OpenSearch will create if it not exists).
## This is a Golang template (see https://pkg.go.dev/text/template)
## You can also specify
## metric name (`{{.Name}}`), tag value (`{{.Tag "tag_name"}}`), field value (`{{.Field "field_name"}}`)
## If the tag does not exist, the default tag value will be empty string "".
## the timestamp (`{{.Time.Format "xxxxxxxxx"}}`).
## For example: "telegraf-{{.Time.Format \"2006-01-02\"}}-{{.Tag \"host\"}}" would set it to telegraf-2023-07-27-HostName
index_name = ""
## Timeout
## OpenSearch client timeout
# timeout = "5s"
## Sniffer
## Set to true to ask OpenSearch a list of all cluster nodes,
## thus it is not necessary to list all nodes in the urls config option
# enable_sniffer = false
## GZIP Compression
## Set to true to enable gzip compression
# enable_gzip = false
## Health Check Interval
## Set the interval to check if the OpenSearch nodes are available
## Setting to "0s" will disable the health check (not recommended in production)
# health_check_interval = "10s"
## Set the timeout for periodic health checks.
# health_check_timeout = "1s"
## HTTP basic authentication details.
# username = ""
# password = ""
## HTTP bearer token authentication details
# auth_bearer_token = ""
## Optional TLS Config
## Set to true/false to enforce TLS being enabled/disabled. If not set,
## enable TLS only if any of the other options are specified.
# tls_enable =
## Trusted root certificates for server
# tls_ca = "/path/to/cafile"
## Used for TLS client certificate authentication
# tls_cert = "/path/to/certfile"
## Used for TLS client certificate authentication
# tls_key = "/path/to/keyfile"
## Send the specified TLS server name via SNI
# tls_server_name = "kubernetes.example.com"
## Use TLS but skip chain & host verification
# insecure_skip_verify = false
## Template Config
## Manage templates
## Set to true if you want telegraf to manage its index template.
## If enabled it will create a recommended index template for telegraf indexes
# manage_template = true
## Template Name
## The template name used for telegraf indexes
# template_name = "telegraf"
## Overwrite Templates
## Set to true if you want telegraf to overwrite an existing template
# overwrite_template = false
## Document ID
## If set to true a unique ID hash will be sent as
## sha256(concat(timestamp,measurement,series-hash)) string. It will enable
## data resend and update metric points avoiding duplicated metrics with
## different id's
# force_document_id = false
## Value Handling
## Specifies the handling of NaN and Inf values.
## This option can have the following values:
## none -- do not modify field-values (default); will produce an error
## if NaNs or infs are encountered
## drop -- drop fields containing NaNs or infs
## replace -- replace with the value in "float_replacement_value" (default: 0.0)
## NaNs and inf will be replaced with the given number, -inf with the negative of that number
# float_handling = "none"
# float_replacement_value = 0.0
## Pipeline Config
## To use a ingest pipeline, set this to the name of the pipeline you want to use.
# use_pipeline = "my_pipeline"
## Pipeline Name
## Additionally, you can specify a tag name using the notation (`{{.Tag "tag_name"}}`)
## which will be used as the pipeline name (e.g. "{{.Tag \"os_pipeline\"}}").
## If the tag does not exist, the default pipeline will be used as the pipeline.
## If no default pipeline is set, no pipeline is used for the metric.
# default_pipeline = ""
输入和输出集成示例
OpenStack
-
跨云管理:利用 OpenStack 插件从单个 Telegraf 实例监控和管理多个 OpenStack 云。通过聚合不同云的指标,组织可以深入了解资源利用率,并优化其云架构以实现成本和性能的优化。
-
基于指标的自动扩展:将从 OpenStack 收集的指标集成到自动扩展解决方案中。例如,如果插件检测到特定服务的性能下降,它可以触发自动扩展规则以启动额外的实例,从而确保系统性能在不同的工作负载下保持最佳状态。
-
性能监控仪表板:使用 OpenStack Telegraf 插件收集的数据来支持实时监控仪表板。此设置提供了来自 OpenStack 服务的关键指标的可视化,使利益干系人能够快速识别趋势、查明问题,并在管理其云基础设施时做出数据驱动的决策。
-
服务可用性报告和分析:通过利用从各种 OpenStack 服务收集的指标,团队可以生成关于服务可用性和随时间推移的性能的详细报告。此信息可以帮助识别重复出现的问题、改进服务交付,并就基础设施或服务配置的更改做出明智的决策。
OpenSearch
-
时间序列数据的动态索引:利用 OpenSearch Telegraf 插件为时间序列指标动态创建索引,确保数据以有组织的方式存储,从而有利于基于时间的查询。通过使用 Go 模板定义索引模式,用户可以利用该插件创建每日或每月索引,这可以大大简化数据管理和随时间推移的检索,从而提高分析性能。
-
多租户应用程序的集中式日志记录:在多租户应用程序中实施 OpenSearch 插件,其中每个租户的日志都发送到单独的索引。这使得可以对每个租户进行有针对性的分析和监控,同时保持数据隔离。通过利用索引名称模板功能,用户可以自动创建特定于租户的索引,这不仅简化了流程,而且还增强了租户数据的安全性和可访问性。
-
与机器学习集成以进行异常检测:将 OpenSearch 插件与机器学习工具结合使用,以自动检测指标数据中的异常。通过配置插件以将实时指标发送到 OpenSearch,用户可以将机器学习模型应用于传入的数据流,以识别异常值或异常模式,从而促进主动监控和快速补救措施。
-
使用 OpenSearch 增强监控仪表板:使用从 OpenSearch 收集的指标创建实时仪表板,以深入了解系统性能。通过将指标馈送到 OpenSearch,组织可以利用 OpenSearch Dashboards 可视化关键绩效指标,使运营团队能够快速评估运行状况和性能,并做出数据驱动的决策。
反馈
感谢您成为我们社区的一份子!如果您有任何一般性反馈或在这些页面上发现任何错误,我们欢迎并鼓励您提出意见。请在 InfluxDB 社区 Slack 中提交您的反馈。
强大的性能,无限的扩展能力
收集、组织和处理海量高速数据。当您将任何数据视为时间序列数据时,它都会更有价值。借助 InfluxDB,这款排名第一的时间序列平台旨在与 Telegraf 一起扩展。
查看入门方法