Reference architectures
Recommended patterns for reliably sending high-volume telemetry to Axiom using industry-standard collectors like OpenTelemetry and Vector.
Successfully adopting a new platform at scale requires a robust and reliable data collection strategy. In complex distributed systems, how you collect and forward telemetry is as important as the platform that analyzes it.
While Axiom is flexible and supports dozens of methods, there is a confident, battle-tested recommendation for large-scale environments. This guide outlines Axiom’s recommended architectural patterns using two popular tools: the OpenTelemetry Collector and Vector.
These patterns are designed to provide:
- Resilience: Ensure no data is lost during network partitions or downstream issues.
- Enrichment: Add valuable, consistent metadata to all your telemetry at the source.
- Performance: Offload processing and batching from your applications.
- Governance: Centrally manage data routing, filtering, and sampling to control costs.
While these patterns are recommended for large-scale environments, Axiom has many other simpler methods available to send data.
The core patterns
There are two primary deployment patterns for a collector: the Agent model and the Aggregator model. Most large organizations use a hybrid of both.
Most organizations find it easy to send data to Axiom. If you are having trouble, please contact the Axiom team for help. We can provide ad-hoc guidance or design tailored implementation services for larger projects.
Agent pattern
In this model, a lightweight collector runs as an agent on every host or as a sidecar in every pod. It’s responsible for collecting telemetry from applications on that specific node and forwarding it directly to Axiom.
Best for: Capturing rich, host-specific metadata (e.g., k8s.pod.name
, host.id
) and providing a resilient, decentralized collection point with local buffering.
Aggregator pattern
In this model, a separate, horizontally-scalable pool of collectors acts as a centralized aggregation layer. Applications and agents send their data to this layer, which then handles final processing, enrichment, and forwarding to Axiom.
Best for: Centralized control over data routing and filtering, managing data from sources that cannot run an agent (e.g., third-party APIs, cloud provider log streams), and simplifying management by maintaining a small fleet of aggregators instead of thousands of agents.
Tool recommendations
Both the OpenTelemetry Collector and Vector are excellent choices that can be deployed in either an Agent or Aggregator pattern. The best choice depends on your team’s existing ecosystem and primary use case.
OpenTelemetry Collector
The OpenTelemetry Collector is the CNCF-backed, vendor-neutral standard for collecting and processing telemetry. It’s the ideal choice when your organization is standardizing on the OpenTelemetry framework for all signals (logs, metrics, and traces).
Use the OTel Collector when:
- You are instrumenting your applications with OpenTelemetry SDKs.
- You need to process traces, metrics, and logs in a single, unified pipeline.
- You want to align with the fastest-growing standard in the observability community.
Axiom natively supports the OpenTelemetry Line Protocol (OTLP). Configuring the collector to send data to Axiom is simple.
Example Collector Configuration (otel-collector-config.yaml
):
Replace AXIOM_DOMAIN
with api.axiom.co
if your organization uses the US region, and with api.eu.axiom.co
if your organization uses the EU region. For more information, see Regions.
Replace API_TOKEN
with the Axiom API token you have generated. For added security, store the API token in an environment variable.
Replace DATASET_NAME
with the name of the Axiom dataset where you send your data.
Vector
Vector is a high-performance, Rust-based observability data pipeline. It excels at log collection and transformation, offering a powerful domain-specific language (VRL) for complex data manipulation.
Use Vector when:
- Your primary focus is on logs from a wide variety of sources (files, syslog, APIs).
- You need to perform complex parsing, filtering, enrichment, or redaction on your event data.
- You require an extremely lightweight and memory-efficient agent for edge deployments.
Vector has a native sink for Axiom, making configuration straightforward.
Example Vector Configuration (vector.toml
):
Replace VECTOR_SOURCE_ID
with the Vector source ID.
Replace PATH_TO_LOGS
with the path to the log files. For example, /var/log/**/*.log
.
Replace SINK_ID
with the sink ID.
Replace API_TOKEN
with the Axiom API token you have generated. For added security, store the API token in an environment variable.
Replace DATASET_NAME
with the name of the Axiom dataset where you send your data.
Summary
Aspect | OpenTelemetry Collector | Vector |
---|---|---|
Primary Use Case | Unified pipeline for logs, metrics, and traces | High-performance log collection and transformation |
Ecosystem | CNCF and OpenTelemetry standard | Standalone, broad source/sink compatibility |
Transformation | Processors (limited for complex logic) | Vector Remap Language (VRL) for advanced logic |
Performance | Excellent | Excellent, often with lower resource footprint for logs |
Both are first-class choices for sending data to Axiom. Your decision should be based on whether you need a unified OTel pipeline or a specialized, high-performance log processing tool.
What’s next?
- Explore the complete range of options for sending data in the Methods page.
- For direct ingestion, see the Axiom REST API.