Skip to content

Commit f343aa0

Browse files
authored
Merge branch 'main' into value-map-collect
2 parents edbbeaf + 500d913 commit f343aa0

File tree

10 files changed

+85
-327
lines changed

10 files changed

+85
-327
lines changed

examples/self-diagnostics/Cargo.toml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,7 @@ publish = false
99
opentelemetry = { path = "../../opentelemetry" }
1010
opentelemetry_sdk = { path = "../../opentelemetry-sdk", features = ["rt-tokio"]}
1111
opentelemetry-stdout = { path = "../../opentelemetry-stdout"}
12-
opentelemetry-appender-tracing = { path = "../../opentelemetry-appender-tracing"}
1312
tokio = { workspace = true, features = ["full"] }
1413
tracing = { workspace = true, features = ["std"]}
1514
tracing-core = { workspace = true }
1615
tracing-subscriber = { version = "0.3.18", features = ["env-filter","registry", "std"]}
17-
opentelemetry-otlp = { path = "../../opentelemetry-otlp", features = ["http-proto", "reqwest-client", "logs"] }
18-
once_cell ={ version = "1.19.0"}
19-
ctrlc = "3.4"

examples/self-diagnostics/Dockerfile

Lines changed: 0 additions & 6 deletions
This file was deleted.
Lines changed: 23 additions & 88 deletions
Original file line numberDiff line numberDiff line change
@@ -1,93 +1,28 @@
11
# Basic OpenTelemetry metrics example with custom error handler:
22

3-
This example shows how to setup the custom error handler for self-diagnostics.
4-
5-
## Custom Error Handling:
6-
7-
A custom error handler is set up to capture and record errors using the `tracing` crate's `error!` macro. These errors are then exported to a collector using the `opentelemetry-appender-tracing` crate, which utilizes the OTLP log exporter over `HTTP/protobuf`. As a result, any errors generated by the configured OTLP metrics pipeline are funneled through this custom error handler for proper recording and export.
3+
This example shows how to self-diagnose OpenTelemetry by enabling its internal
4+
logs. OpenTelemetry crates publish internal logs when "internal-logs" feature is
5+
enabled. This feature is enabled by default. Internal logs are published using
6+
`tracing` events, and hence, a `tracing` subscriber must be configured without
7+
which the logs are simply discarded.
88

99
## Filtering logs from external dependencies of OTLP Exporter:
1010

11-
The example configures a tracing `filter` to restrict logs from external crates (`hyper`, `tonic`, and `reqwest`) used by the OTLP Exporter to the `error` level. This helps prevent an infinite loop of log generation when these crates emit logs that are picked up by the tracing subscriber.
12-
13-
## Ensure that the internally generated errors are logged only once:
14-
15-
By using a hashset to track seen errors, the custom error handler ensures that the same error is not logged multiple times. This is particularly useful for handling scenarios where continuous error logging might occur, such as when the OpenTelemetry collector is not running.
16-
17-
18-
## Usage
19-
20-
### `docker-compose`
21-
22-
By default runs against the `otel/opentelemetry-collector:latest` image, and uses `reqwest-client`
23-
as the http client, using http as the transport.
24-
25-
```shell
26-
docker-compose up
27-
```
28-
29-
In another terminal run the application `cargo run`
30-
31-
The docker-compose terminal will display logs, traces, metrics.
32-
33-
Press Ctrl+C to stop the collector, and then tear it down:
34-
35-
```shell
36-
docker-compose down
37-
```
38-
39-
### Manual
40-
41-
If you don't want to use `docker-compose`, you can manually run the `otel/opentelemetry-collector` container
42-
and inspect the logs to see traces being transferred.
43-
44-
On Unix based systems use:
45-
46-
```shell
47-
# From the current directory, run `opentelemetry-collector`
48-
docker run --rm -it -p 4318:4318 -v $(pwd):/cfg otel/opentelemetry-collector:latest --config=/cfg/otel-collector-config.yaml
49-
```
50-
51-
On Windows use:
52-
53-
```shell
54-
# From the current directory, run `opentelemetry-collector`
55-
docker run --rm -it -p 4318:4318 -v "%cd%":/cfg otel/opentelemetry-collector:latest --config=/cfg/otel-collector-config.yaml
56-
```
57-
58-
Run the app which exports logs, metrics and traces via OTLP to the collector
59-
60-
```shell
61-
cargo run
62-
```
63-
64-
### Output:
65-
66-
- If the docker instance for collector is running, below error should be logged into the container. There won't be any logs from the `hyper`, `reqwest` and `tonic` crates.
67-
```
68-
otel-collector-1 | 2024-06-05T17:09:46.926Z info LogExporter {"kind": "exporter", "data_type": "logs", "name": "logging", "resource logs": 1, "log records": 1}
69-
otel-collector-1 | 2024-06-05T17:09:46.926Z info ResourceLog #0
70-
otel-collector-1 | Resource SchemaURL:
71-
otel-collector-1 | Resource attributes:
72-
otel-collector-1 | -> telemetry.sdk.name: Str(opentelemetry)
73-
otel-collector-1 | -> telemetry.sdk.version: Str(0.23.0)
74-
otel-collector-1 | -> telemetry.sdk.language: Str(rust)
75-
otel-collector-1 | -> service.name: Str(unknown_service)
76-
otel-collector-1 | ScopeLogs #0
77-
otel-collector-1 | ScopeLogs SchemaURL:
78-
otel-collector-1 | InstrumentationScope opentelemetry-appender-tracing 0.4.0
79-
otel-collector-1 | LogRecord #0
80-
otel-collector-1 | ObservedTimestamp: 2024-06-05 17:09:45.931951161 +0000 UTC
81-
otel-collector-1 | Timestamp: 1970-01-01 00:00:00 +0000 UTC
82-
otel-collector-1 | SeverityText: ERROR
83-
otel-collector-1 | SeverityNumber: Error(17)
84-
otel-collector-1 | Body: Str(OpenTelemetry metrics error occurred: Metrics error: Warning: Maximum data points for metric stream exceeded. Entry added to overflow. Subsequent overflows to same metric until next collect will not be logged.)
85-
otel-collector-1 | Attributes:
86-
otel-collector-1 | -> name: Str(event examples/self-diagnostics/src/main.rs:42)
87-
otel-collector-1 | Trace ID:
88-
otel-collector-1 | Span ID:
89-
otel-collector-1 | Flags: 0
90-
otel-collector-1 | {"kind": "exporter", "data_type": "logs", "name": "logging"}
91-
```
92-
93-
- The SDK will keep trying to upload metrics at regular intervals if the collector's Docker instance is down. To avoid a logging loop, internal errors like 'Connection refused' will be attempted to be logged only once.
11+
The example configures a tracing `filter` to restrict logs from external crates
12+
(`hyper`, `tonic`, and `reqwest` etc.) used by the OTLP Exporter to the `error`
13+
level. This helps prevent an infinite loop of log generation when these crates
14+
emit logs that are picked up by the tracing subscriber. This is only a
15+
workaround until [the root
16+
issue](https://github.com/open-telemetry/opentelemetry-rust/issues/761) is
17+
resolved.
18+
19+
## Filtering logs to be send to OpenTelemetry itself
20+
21+
If you use [OpenTelemetry Tracing
22+
Appender](../../opentelemetry-appender-tracing/README.md) to send `tracing` logs
23+
to OpenTelemetry, then enabling OpenTelemetry internal logs can also cause
24+
infinite, recursive logging. You can filter out all OpenTelemetry internal logs
25+
from being sent to [OpenTelemetry Tracing
26+
Appender](../../opentelemetry-appender-tracing/README.md) using a filter, like
27+
"add_directive("opentelemetry=off".parse().unwrap())" being done for tracing's
28+
`FmtSubscriber`.

examples/self-diagnostics/docker-compose.yaml

Lines changed: 0 additions & 11 deletions
This file was deleted.

examples/self-diagnostics/otel-collector-config.yaml

Lines changed: 0 additions & 29 deletions
This file was deleted.
Lines changed: 36 additions & 146 deletions
Original file line numberDiff line numberDiff line change
@@ -1,123 +1,16 @@
1-
use opentelemetry::global::{self, set_error_handler, Error as OtelError};
1+
use opentelemetry::global;
22
use opentelemetry::KeyValue;
3-
use opentelemetry_appender_tracing::layer;
4-
use opentelemetry_otlp::{LogExporter, MetricExporter, WithExportConfig};
53
use opentelemetry_sdk::metrics::PeriodicReader;
6-
use tracing_subscriber::filter::{EnvFilter, LevelFilter};
4+
use std::error::Error;
5+
use tracing::info;
76
use tracing_subscriber::fmt;
87
use tracing_subscriber::prelude::*;
9-
10-
use std::error::Error;
11-
use tracing::error;
12-
13-
use once_cell::sync::Lazy;
14-
use std::collections::HashSet;
15-
use std::sync::{Arc, Mutex};
16-
17-
use std::sync::mpsc::channel;
18-
19-
struct ErrorState {
20-
seen_errors: Mutex<HashSet<String>>,
21-
}
22-
23-
impl ErrorState {
24-
fn new() -> Self {
25-
ErrorState {
26-
seen_errors: Mutex::new(HashSet::new()),
27-
}
28-
}
29-
30-
fn mark_as_seen(&self, err: &OtelError) -> bool {
31-
let mut seen_errors = self.seen_errors.lock().unwrap();
32-
seen_errors.insert(err.to_string())
33-
}
34-
}
35-
36-
static GLOBAL_ERROR_STATE: Lazy<Arc<ErrorState>> = Lazy::new(|| Arc::new(ErrorState::new()));
37-
38-
fn custom_error_handler(err: OtelError) {
39-
if GLOBAL_ERROR_STATE.mark_as_seen(&err) {
40-
// log error not already seen
41-
match err {
42-
OtelError::Metric(err) => error!("OpenTelemetry metrics error occurred: {}", err),
43-
OtelError::Trace(err) => error!("OpenTelemetry trace error occurred: {}", err),
44-
OtelError::Log(err) => error!("OpenTelemetry log error occurred: {}", err),
45-
OtelError::Propagation(err) => {
46-
error!("OpenTelemetry propagation error occurred: {}", err)
47-
}
48-
OtelError::Other(err_msg) => error!("OpenTelemetry error occurred: {}", err_msg),
49-
_ => error!("OpenTelemetry error occurred: {:?}", err),
50-
}
51-
}
52-
}
53-
54-
fn init_logger_provider() -> opentelemetry_sdk::logs::LoggerProvider {
55-
let exporter = LogExporter::builder()
56-
.with_http()
57-
.with_endpoint("http://localhost:4318/v1/logs")
58-
.build()
59-
.unwrap();
60-
61-
let provider = opentelemetry_sdk::logs::LoggerProvider::builder()
62-
.with_batch_exporter(exporter, opentelemetry_sdk::runtime::Tokio)
63-
.build();
64-
65-
let cloned_provider = provider.clone();
66-
67-
// Add a tracing filter to filter events from crates used by opentelemetry-otlp.
68-
// The filter levels are set as follows:
69-
// - Allow `info` level and above by default.
70-
// - Restrict `hyper`, `tonic`, and `reqwest` to `error` level logs only.
71-
// This ensures events generated from these crates within the OTLP Exporter are not looped back,
72-
// thus preventing infinite event generation.
73-
// Note: This will also drop events from these crates used outside the OTLP Exporter.
74-
// For more details, see: https://github.com/open-telemetry/opentelemetry-rust/issues/761
75-
let filter = EnvFilter::new("info")
76-
.add_directive("hyper=error".parse().unwrap())
77-
.add_directive("tonic=error".parse().unwrap())
78-
.add_directive("reqwest=error".parse().unwrap());
79-
80-
// Configuring the formatting layer specifically for OpenTelemetry internal logs.
81-
// These logs starts with "opentelemetry" prefix in target. This allows specific logs
82-
// from the OpenTelemetry-related components to be filtered and handled separately
83-
// from the application logs
84-
85-
let opentelemetry_filter = tracing_subscriber::filter::filter_fn(|metadata| {
86-
metadata.target().starts_with("opentelemetry")
87-
});
88-
89-
let fmt_opentelemetry_layer = fmt::layer()
90-
.with_filter(LevelFilter::DEBUG)
91-
.with_filter(opentelemetry_filter);
92-
93-
// Configures the appender tracing layer, filtering out OpenTelemetry internal logs
94-
// to prevent infinite logging loops.
95-
96-
let non_opentelemetry_filter = tracing_subscriber::filter::filter_fn(|metadata| {
97-
!metadata.target().starts_with("opentelemetry")
98-
});
99-
100-
let otel_layer = layer::OpenTelemetryTracingBridge::new(&cloned_provider)
101-
.with_filter(non_opentelemetry_filter.clone());
102-
103-
tracing_subscriber::registry()
104-
.with(fmt_opentelemetry_layer)
105-
.with(fmt::layer().with_filter(filter))
106-
.with(otel_layer)
107-
.init();
108-
provider
109-
}
8+
use tracing_subscriber::EnvFilter;
1109

11110
fn init_meter_provider() -> opentelemetry_sdk::metrics::SdkMeterProvider {
112-
let exporter = MetricExporter::builder()
113-
.with_http()
114-
.with_endpoint("http://localhost:4318/v1/metrics")
115-
.build()
116-
.unwrap();
11+
let exporter = opentelemetry_stdout::MetricExporterBuilder::default().build();
11712

118-
let reader = PeriodicReader::builder(exporter, opentelemetry_sdk::runtime::Tokio)
119-
.with_interval(std::time::Duration::from_secs(1))
120-
.build();
13+
let reader = PeriodicReader::builder(exporter, opentelemetry_sdk::runtime::Tokio).build();
12114

12215
let provider = opentelemetry_sdk::metrics::SdkMeterProvider::builder()
12316
.with_reader(reader)
@@ -130,46 +23,43 @@ fn init_meter_provider() -> opentelemetry_sdk::metrics::SdkMeterProvider {
13023

13124
#[tokio::main]
13225
async fn main() -> Result<(), Box<dyn Error + Send + Sync + 'static>> {
133-
// Set the custom error handler
134-
if let Err(err) = set_error_handler(custom_error_handler) {
135-
eprintln!("Failed to set custom error handler: {}", err);
136-
}
26+
// OpenTelemetry uses `tracing` crate for its internal logging. Unless a
27+
// tracing subscriber is set, the logs will be discarded. In this example,
28+
// we configure a `tracing` subscriber to:
29+
// 1. Print logs of level INFO or higher to stdout.
30+
// 2. Filter logs from OpenTelemetry's dependencies (like tonic, hyper,
31+
// reqwest etc. which are commonly used by the OTLP exporter) to only print
32+
// ERROR-level logs. This filtering helps reduce repetitive log messages
33+
// that could otherwise create an infinite loop of log output. This is a
34+
// workaround until
35+
// https://github.com/open-telemetry/opentelemetry-rust/issues/761 is
36+
// resolved.
37+
38+
// Target name used by OpenTelemetry always start with "opentelemetry".
39+
// Hence, one may use "add_directive("opentelemetry=off".parse().unwrap())"
40+
// to turn off all logs from OpenTelemetry.
13741

138-
let logger_provider = init_logger_provider();
42+
let filter = EnvFilter::new("info")
43+
.add_directive("hyper=error".parse().unwrap())
44+
.add_directive("tonic=error".parse().unwrap())
45+
.add_directive("h2=error".parse().unwrap())
46+
.add_directive("tower=error".parse().unwrap())
47+
.add_directive("reqwest=error".parse().unwrap());
48+
tracing_subscriber::registry()
49+
.with(fmt::layer().with_thread_names(true).with_filter(filter))
50+
.init();
13951

14052
// Initialize the MeterProvider with the stdout Exporter.
14153
let meter_provider = init_meter_provider();
54+
info!("Starting self-diagnostics example");
14255

143-
// Create a meter from the above MeterProvider.
14456
let meter = global::meter("example");
145-
// Create a Counter Instrument.
146-
let counter = meter.u64_counter("my_counter").build();
147-
148-
// Record measurements with unique key-value pairs to exceed the cardinality limit
149-
// of 2000 and trigger error message
150-
for i in 0..3000 {
151-
counter.add(
152-
10,
153-
&[KeyValue::new(
154-
format!("mykey{}", i),
155-
format!("myvalue{}", i),
156-
)],
157-
);
158-
}
159-
160-
let (tx, rx) = channel();
161-
162-
ctrlc::set_handler(move || tx.send(()).expect("Could not send signal on channel."))
163-
.expect("Error setting Ctrl-C handler");
164-
165-
println!("Press Ctrl-C to continue...");
166-
rx.recv().expect("Could not receive from channel.");
167-
println!("Got Ctrl-C, Doing shutdown and existing.");
57+
// Create a counter using an invalid name to trigger
58+
// internal log about the same.
59+
let counter = meter.u64_counter("my_counter with_space").build();
60+
counter.add(10, &[KeyValue::new("key", "value")]);
16861

169-
// MeterProvider is configured with an OTLP Exporter to export metrics every 1 second,
170-
// however shutting down the MeterProvider here instantly flushes
171-
// the metrics, instead of waiting for the 1 sec interval.
17262
meter_provider.shutdown()?;
173-
let _ = logger_provider.shutdown();
63+
info!("Shutdown complete. Bye!");
17464
Ok(())
17565
}

0 commit comments

Comments
 (0)