diff --git a/exporter/alertmanagerexporter/alertmanager_exporter.go b/exporter/alertmanagerexporter/alertmanager_exporter.go index 73ae3cb45310c..36feed0c6ec38 100644 --- a/exporter/alertmanagerexporter/alertmanager_exporter.go +++ b/exporter/alertmanagerexporter/alertmanager_exporter.go @@ -189,7 +189,7 @@ func (s *alertmanagerExporter) pushTraces(ctx context.Context, td ptrace.Traces) } func (s *alertmanagerExporter) start(ctx context.Context, host component.Host) error { - client, err := s.config.ClientConfig.ToClient(ctx, host, s.settings) + client, err := s.config.ToClient(ctx, host, s.settings) if err != nil { return fmt.Errorf("failed to create HTTP Client: %w", err) } @@ -209,7 +209,7 @@ func newAlertManagerExporter(cfg *Config, set component.TelemetrySettings) *aler config: cfg, settings: set, tracesMarshaler: &ptrace.JSONMarshaler{}, - endpoint: fmt.Sprintf("%s/api/%s/alerts", cfg.ClientConfig.Endpoint, cfg.APIVersion), + endpoint: fmt.Sprintf("%s/api/%s/alerts", cfg.Endpoint, cfg.APIVersion), generatorURL: cfg.GeneratorURL, defaultSeverity: cfg.DefaultSeverity, severityAttribute: cfg.SeverityAttribute, diff --git a/exporter/alertmanagerexporter/config.go b/exporter/alertmanagerexporter/config.go index c45b91d263a55..ef5496f3e2fa8 100644 --- a/exporter/alertmanagerexporter/config.go +++ b/exporter/alertmanagerexporter/config.go @@ -30,7 +30,7 @@ var _ component.Config = (*Config)(nil) // Validate checks if the exporter configuration is valid func (cfg *Config) Validate() error { - if cfg.ClientConfig.Endpoint == "" { + if cfg.Endpoint == "" { return errors.New("endpoint must be non-empty") } if cfg.DefaultSeverity == "" { diff --git a/exporter/alertmanagerexporter/config_test.go b/exporter/alertmanagerexporter/config_test.go index b8d970c06733e..dbc5d59fd2dcc 100644 --- a/exporter/alertmanagerexporter/config_test.go +++ b/exporter/alertmanagerexporter/config_test.go @@ -112,7 +112,7 @@ func TestConfig_Validate(t *testing.T) { name: "NoEndpoint", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "" + cfg.Endpoint = "" return cfg }(), wantErr: "endpoint must be non-empty", diff --git a/exporter/awsemfexporter/datapoint.go b/exporter/awsemfexporter/datapoint.go index fbb48507a7c7e..ae05bb13000e1 100644 --- a/exporter/awsemfexporter/datapoint.go +++ b/exporter/awsemfexporter/datapoint.go @@ -138,7 +138,7 @@ func (split *dataPointSplit) appendMetricData(metricVal float64, count uint64) { // CalculateDeltaDatapoints retrieves the NumberDataPoint at the given index and performs rate/delta calculation if necessary. func (dps numberDataPointSlice) CalculateDeltaDatapoints(i int, instrumentationScopeName string, _ bool, calculators *emfCalculators) ([]dataPoint, bool) { - metric := dps.NumberDataPointSlice.At(i) + metric := dps.At(i) labels := createLabels(metric.Attributes(), instrumentationScopeName) timestampMs := unixNanoToMilliseconds(metric.Timestamp()) @@ -177,7 +177,7 @@ func (dps numberDataPointSlice) CalculateDeltaDatapoints(i int, instrumentationS } func (dps numberDataPointSlice) IsStaleNaNInf(i int) (bool, pcommon.Map) { - metric := dps.NumberDataPointSlice.At(i) + metric := dps.At(i) if metric.Flags().NoRecordedValue() { return true, metric.Attributes() } @@ -189,7 +189,7 @@ func (dps numberDataPointSlice) IsStaleNaNInf(i int) (bool, pcommon.Map) { // CalculateDeltaDatapoints retrieves the HistogramDataPoint at the given index. func (dps histogramDataPointSlice) CalculateDeltaDatapoints(i int, instrumentationScopeName string, _ bool, _ *emfCalculators) ([]dataPoint, bool) { - metric := dps.HistogramDataPointSlice.At(i) + metric := dps.At(i) labels := createLabels(metric.Attributes(), instrumentationScopeName) timestamp := unixNanoToMilliseconds(metric.Timestamp()) @@ -207,7 +207,7 @@ func (dps histogramDataPointSlice) CalculateDeltaDatapoints(i int, instrumentati } func (dps histogramDataPointSlice) IsStaleNaNInf(i int) (bool, pcommon.Map) { - metric := dps.HistogramDataPointSlice.At(i) + metric := dps.At(i) if metric.Flags().NoRecordedValue() { return true, metric.Attributes() } @@ -229,7 +229,7 @@ func (dps histogramDataPointSlice) IsStaleNaNInf(i int) (bool, pcommon.Map) { // - Sum is only assigned to the first split to ensure the total sum of the datapoints after aggregation is correct. // - Count is accumulated based on the bucket counts within each split. func (dps exponentialHistogramDataPointSlice) CalculateDeltaDatapoints(idx int, instrumentationScopeName string, _ bool, _ *emfCalculators) ([]dataPoint, bool) { - metric := dps.ExponentialHistogramDataPointSlice.At(idx) + metric := dps.At(idx) const splitThreshold = 100 currentBucketIndex := 0 @@ -412,7 +412,7 @@ func collectDatapointsWithNegativeBuckets(split *dataPointSplit, metric pmetric. } func (dps exponentialHistogramDataPointSlice) IsStaleNaNInf(i int) (bool, pcommon.Map) { - metric := dps.ExponentialHistogramDataPointSlice.At(i) + metric := dps.At(i) if metric.Flags().NoRecordedValue() { return true, metric.Attributes() } @@ -430,7 +430,7 @@ func (dps exponentialHistogramDataPointSlice) IsStaleNaNInf(i int) (bool, pcommo // CalculateDeltaDatapoints retrieves the SummaryDataPoint at the given index and perform calculation with sum and count while retain the quantile value. func (dps summaryDataPointSlice) CalculateDeltaDatapoints(i int, instrumentationScopeName string, detailedMetrics bool, calculators *emfCalculators) ([]dataPoint, bool) { - metric := dps.SummaryDataPointSlice.At(i) + metric := dps.At(i) labels := createLabels(metric.Attributes(), instrumentationScopeName) timestampMs := unixNanoToMilliseconds(metric.Timestamp()) @@ -485,7 +485,7 @@ func (dps summaryDataPointSlice) CalculateDeltaDatapoints(i int, instrumentation } func (dps summaryDataPointSlice) IsStaleNaNInf(i int) (bool, pcommon.Map) { - metric := dps.SummaryDataPointSlice.At(i) + metric := dps.At(i) if metric.Flags().NoRecordedValue() { return true, metric.Attributes() } diff --git a/exporter/awsemfexporter/datapoint_test.go b/exporter/awsemfexporter/datapoint_test.go index 30a10963bd6bf..5239870e6a5a5 100644 --- a/exporter/awsemfexporter/datapoint_test.go +++ b/exporter/awsemfexporter/datapoint_test.go @@ -2052,7 +2052,7 @@ func TestGetDataPoints(t *testing.T) { expectedDPS := tc.expectedDatapointSlice.(numberDataPointSlice) assert.Equal(t, expectedDPS.deltaMetricMetadata, convertedDPS.deltaMetricMetadata) assert.Equal(t, 1, convertedDPS.Len()) - dp := convertedDPS.NumberDataPointSlice.At(0) + dp := convertedDPS.At(0) switch dp.ValueType() { case pmetric.NumberDataPointValueTypeDouble: assert.Equal(t, 0.1, dp.DoubleValue()) @@ -2062,14 +2062,14 @@ func TestGetDataPoints(t *testing.T) { assert.Equal(t, tc.expectedAttributes, dp.Attributes().AsRaw()) case histogramDataPointSlice: assert.Equal(t, 1, convertedDPS.Len()) - dp := convertedDPS.HistogramDataPointSlice.At(0) + dp := convertedDPS.At(0) assert.Equal(t, 35.0, dp.Sum()) assert.Equal(t, uint64(18), dp.Count()) assert.Equal(t, []float64{0, 10}, dp.ExplicitBounds().AsRaw()) assert.Equal(t, tc.expectedAttributes, dp.Attributes().AsRaw()) case exponentialHistogramDataPointSlice: assert.Equal(t, 1, convertedDPS.Len()) - dp := convertedDPS.ExponentialHistogramDataPointSlice.At(0) + dp := convertedDPS.At(0) assert.Equal(t, float64(0), dp.Sum()) assert.Equal(t, uint64(4), dp.Count()) assert.Equal(t, []uint64{1, 0, 1}, dp.Positive().BucketCounts().AsRaw()) @@ -2080,7 +2080,7 @@ func TestGetDataPoints(t *testing.T) { expectedDPS := tc.expectedDatapointSlice.(summaryDataPointSlice) assert.Equal(t, expectedDPS.deltaMetricMetadata, convertedDPS.deltaMetricMetadata) assert.Equal(t, 1, convertedDPS.Len()) - dp := convertedDPS.SummaryDataPointSlice.At(0) + dp := convertedDPS.At(0) assert.Equal(t, 15.0, dp.Sum()) assert.Equal(t, uint64(5), dp.Count()) assert.Equal(t, 2, dp.QuantileValues().Len()) diff --git a/exporter/awsemfexporter/grouped_metric.go b/exporter/awsemfexporter/grouped_metric.go index 8d4f8a57aa056..d0456f8740053 100644 --- a/exporter/awsemfexporter/grouped_metric.go +++ b/exporter/awsemfexporter/grouped_metric.go @@ -86,7 +86,7 @@ func addToGroupedMetric( } // Extra params to use when grouping metrics - metadata.groupedMetricMetadata.batchIndex = i + metadata.batchIndex = i groupKey := aws.NewKey(metadata.groupedMetricMetadata, labels) if _, ok := groupedMetrics[groupKey]; ok { // if MetricName already exists in metrics map, print warning log diff --git a/exporter/awsemfexporter/grouped_metric_test.go b/exporter/awsemfexporter/grouped_metric_test.go index 5509d9fb28276..31c9b2cdef8dd 100644 --- a/exporter/awsemfexporter/grouped_metric_test.go +++ b/exporter/awsemfexporter/grouped_metric_test.go @@ -309,9 +309,10 @@ func TestAddToGroupedMetric(t *testing.T) { } assert.Equal(t, expectedLabels, group.labels) - if group.metadata.logGroup == "log-group-2" { + switch group.metadata.logGroup { + case "log-group-2": seenLogGroup2 = true - } else if group.metadata.logGroup == "log-group-1" { + case "log-group-1": seenLogGroup1 = true } } @@ -441,10 +442,10 @@ func TestAddToGroupedMetric(t *testing.T) { for _, v := range groupedMetrics { assert.Len(t, v.metrics, 1) assert.Len(t, v.labels, 2) - assert.Contains(t, expectedMetadata, v.metadata.groupedMetricMetadata.batchIndex) - assert.Equal(t, expectedMetadata[v.metadata.groupedMetricMetadata.batchIndex], v.metadata) + assert.Contains(t, expectedMetadata, v.metadata.batchIndex) + assert.Equal(t, expectedMetadata[v.metadata.batchIndex], v.metadata) assert.Equal(t, expectedLabels, v.labels) - delete(expectedMetadata, v.metadata.groupedMetricMetadata.batchIndex) + delete(expectedMetadata, v.metadata.batchIndex) } }) } diff --git a/exporter/awsemfexporter/metric_translator_test.go b/exporter/awsemfexporter/metric_translator_test.go index eefd0d7928592..c495ee7c1e176 100644 --- a/exporter/awsemfexporter/metric_translator_test.go +++ b/exporter/awsemfexporter/metric_translator_test.go @@ -348,16 +348,16 @@ func TestTranslateOtToGroupedMetric(t *testing.T) { for _, v := range groupedMetrics { assert.Equal(t, tc.expectedNamespace, v.metadata.namespace) - switch { - case v.metadata.metricDataType == pmetric.MetricTypeSum: + switch v.metadata.metricDataType { + case pmetric.MetricTypeSum: assert.Len(t, v.metrics, 2) assert.Equal(t, tc.counterLabels, v.labels) assert.Equal(t, counterSumMetrics, v.metrics) - case v.metadata.metricDataType == pmetric.MetricTypeGauge: + case pmetric.MetricTypeGauge: assert.Len(t, v.metrics, 2) assert.Equal(t, tc.counterLabels, v.labels) assert.Equal(t, counterGaugeMetrics, v.metrics) - case v.metadata.metricDataType == pmetric.MetricTypeHistogram: + case pmetric.MetricTypeHistogram: assert.Len(t, v.metrics, 1) assert.Equal(t, tc.timerLabels, v.labels) assert.Equal(t, timerMetrics, v.metrics) diff --git a/exporter/awskinesisexporter/exporter.go b/exporter/awskinesisexporter/exporter.go index 762afb09fa572..1ecd5c48dbdfe 100644 --- a/exporter/awskinesisexporter/exporter.go +++ b/exporter/awskinesisexporter/exporter.go @@ -90,7 +90,7 @@ func createExporter(ctx context.Context, c component.Config, log *zap.Logger, op } encoder, err := batch.NewEncoder( - conf.Encoding.Name, + conf.Name, batch.WithMaxRecordSize(conf.MaxRecordSize), batch.WithMaxRecordsPerBatch(conf.MaxRecordsPerBatch), batch.WithCompressionType(conf.Compression), @@ -99,7 +99,7 @@ func createExporter(ctx context.Context, c component.Config, log *zap.Logger, op return nil, err } - if conf.Encoding.Name == "otlp_json" { + if conf.Name == "otlp_json" { log.Info("otlp_json is considered experimental and should not be used in a production environment") } diff --git a/exporter/awss3exporter/sumo_marshaler.go b/exporter/awss3exporter/sumo_marshaler.go index 9ec2c54e64495..c3084d4fd136e 100644 --- a/exporter/awss3exporter/sumo_marshaler.go +++ b/exporter/awss3exporter/sumo_marshaler.go @@ -32,7 +32,7 @@ func newSumoICMarshaler() sumoMarshaler { } func logEntry(buf *bytes.Buffer, format string, a ...any) { - buf.WriteString(fmt.Sprintf(format, a...)) + fmt.Fprintf(buf, format, a...) buf.WriteString("\n") } diff --git a/exporter/awsxrayexporter/internal/translator/cause_test.go b/exporter/awsxrayexporter/internal/translator/cause_test.go index 8d9603ec59b6c..9843ac662cb44 100644 --- a/exporter/awsxrayexporter/internal/translator/cause_test.go +++ b/exporter/awsxrayexporter/internal/translator/cause_test.go @@ -83,8 +83,8 @@ func TestMakeCauseAwsSdkSpan(t *testing.T) { assert.False(t, isThrottle) assert.NotNil(t, cause) - assert.Len(t, cause.CauseObject.Exceptions, 1) - exception := cause.CauseObject.Exceptions[0] + assert.Len(t, cause.Exceptions, 1) + exception := cause.Exceptions[0] assert.Equal(t, AwsIndividualHTTPErrorEventType, *exception.Type) assert.True(t, *exception.Remote) diff --git a/exporter/awsxrayexporter/internal/translator/http.go b/exporter/awsxrayexporter/internal/translator/http.go index b8a868e590d01..781583afe37cf 100644 --- a/exporter/awsxrayexporter/internal/translator/http.go +++ b/exporter/awsxrayexporter/internal/translator/http.go @@ -198,7 +198,7 @@ func constructClientURL(urlParts map[string]string) string { } } url = scheme + "://" + host - if len(port) > 0 && !(scheme == "http" && port == "80") && !(scheme == "https" && port == "443") { + if len(port) > 0 && (scheme != "http" || port != "80") && (scheme != "https" || port != "443") { url += ":" + port } target, ok := urlParts[conventionsv112.AttributeHTTPTarget] @@ -246,7 +246,7 @@ func constructServerURL(urlParts map[string]string) string { } } url = scheme + "://" + host - if len(port) > 0 && !(scheme == "http" && port == "80") && !(scheme == "https" && port == "443") { + if len(port) > 0 && (scheme != "http" || port != "80") && (scheme != "https" || port != "443") { url += ":" + port } target, ok := urlParts[conventionsv112.AttributeHTTPTarget] diff --git a/exporter/awsxrayexporter/internal/translator/segment.go b/exporter/awsxrayexporter/internal/translator/segment.go index 59fff57da3706..0c610fdf6a89d 100644 --- a/exporter/awsxrayexporter/internal/translator/segment.go +++ b/exporter/awsxrayexporter/internal/translator/segment.go @@ -169,7 +169,7 @@ func MakeDependencySubsegmentForLocalRootDependencySpan(span ptrace.Span, resour func MakeServiceSegmentForLocalRootDependencySpan(span ptrace.Span, resource pcommon.Resource, indexedAttrs []string, indexAllAttrs bool, logGroupNames []string, skipTimestampValidation bool, serviceSegmentID pcommon.SpanID) (*awsxray.Segment, error) { // We always create a segment for the service - var serviceSpan ptrace.Span = ptrace.NewSpan() + serviceSpan := ptrace.NewSpan() span.CopyTo(serviceSpan) // Set the span id to the one internally generated diff --git a/exporter/azuredataexplorerexporter/config.go b/exporter/azuredataexplorerexporter/config.go index 9aa4201ed19b0..f15a776a67914 100644 --- a/exporter/azuredataexplorerexporter/config.go +++ b/exporter/azuredataexplorerexporter/config.go @@ -66,7 +66,7 @@ func (adxCfg *Config) Validate() error { return errors.New(`either ["application_id" , "application_key" , "tenant_id"] or ["managed_identity_id"] or ["use_azure_auth"] must be provided for auth`) } - if !(adxCfg.IngestionType == managedIngestType || adxCfg.IngestionType == queuedIngestTest || isEmpty(adxCfg.IngestionType)) { + if adxCfg.IngestionType != managedIngestType && adxCfg.IngestionType != queuedIngestTest && !isEmpty(adxCfg.IngestionType) { return fmt.Errorf("unsupported configuration for ingestion_type. Accepted types [%s, %s] Provided [%s]", managedIngestType, queuedIngestTest, adxCfg.IngestionType) } // Validate managed identity ID. Use system for system assigned managed identity or UserManagedIdentityID (objectID) for user assigned managed identity diff --git a/exporter/azuremonitorexporter/metricexporter_test.go b/exporter/azuremonitorexporter/metricexporter_test.go index 1b19510010516..d470cc64f80d3 100644 --- a/exporter/azuremonitorexporter/metricexporter_test.go +++ b/exporter/azuremonitorexporter/metricexporter_test.go @@ -107,7 +107,7 @@ func TestSummaryEnvelopes(t *testing.T) { } func getDataPoint(tb testing.TB, metric pmetric.Metric) *contracts.DataPoint { - var envelopes []*contracts.Envelope = getMetricPacker().MetricToEnvelopes(metric, getResource(), getScope()) + envelopes := getMetricPacker().MetricToEnvelopes(metric, getResource(), getScope()) require.Len(tb, envelopes, 1) envelope := envelopes[0] require.NotNil(tb, envelope) diff --git a/exporter/azuremonitorexporter/trace_to_envelope.go b/exporter/azuremonitorexporter/trace_to_envelope.go index cbbebe8a3e361..4332e5c8458a7 100644 --- a/exporter/azuremonitorexporter/trace_to_envelope.go +++ b/exporter/azuremonitorexporter/trace_to_envelope.go @@ -85,7 +85,8 @@ func spanToEnvelopes( envelope.Tags[contracts.UserId] = userID.Str() } - if spanKind == ptrace.SpanKindServer || spanKind == ptrace.SpanKindConsumer { + switch spanKind { + case ptrace.SpanKindServer, ptrace.SpanKindConsumer: requestData := spanToRequestData(span, incomingSpanType) dataProperties = requestData.Properties dataSanitizeFunc = requestData.Sanitize @@ -93,7 +94,7 @@ func spanToEnvelopes( envelope.Tags[contracts.OperationName] = requestData.Name data.BaseData = requestData data.BaseType = requestData.BaseType() - } else if spanKind == ptrace.SpanKindClient || spanKind == ptrace.SpanKindProducer || spanKind == ptrace.SpanKindInternal { + case ptrace.SpanKindClient, ptrace.SpanKindProducer, ptrace.SpanKindInternal: remoteDependencyData := spanToRemoteDependencyData(span, incomingSpanType) // Regardless of the detected Span type, if the SpanKind is Internal we need to set data.Type to InProc diff --git a/exporter/coralogixexporter/config.go b/exporter/coralogixexporter/config.go index e5dfc0671a3fa..8c6c585aa0dc8 100644 --- a/exporter/coralogixexporter/config.go +++ b/exporter/coralogixexporter/config.go @@ -89,11 +89,11 @@ func (c *Config) Validate() error { } // check if headers exists - if len(c.ClientConfig.Headers) == 0 { - c.ClientConfig.Headers = make(map[string]configopaque.String) + if len(c.Headers) == 0 { + c.Headers = make(map[string]configopaque.String) } - c.ClientConfig.Headers["ACCESS_TOKEN"] = c.PrivateKey - c.ClientConfig.Headers["appName"] = configopaque.String(c.AppName) + c.Headers["ACCESS_TOKEN"] = c.PrivateKey + c.Headers["appName"] = configopaque.String(c.AppName) return nil } diff --git a/exporter/coralogixexporter/logs_client.go b/exporter/coralogixexporter/logs_client.go index 9ca361ef0eaf6..d2a9b855557f5 100644 --- a/exporter/coralogixexporter/logs_client.go +++ b/exporter/coralogixexporter/logs_client.go @@ -94,7 +94,7 @@ func (e *logsExporter) pushLogs(ctx context.Context, ld plog.Logs) error { } partialSuccess := resp.PartialSuccess() - if !(partialSuccess.ErrorMessage() == "" && partialSuccess.RejectedLogRecords() == 0) { + if partialSuccess.ErrorMessage() != "" || partialSuccess.RejectedLogRecords() != 0 { e.settings.Logger.Error("Partial success response from Coralogix", zap.String("message", partialSuccess.ErrorMessage()), zap.Int64("rejected_log_records", partialSuccess.RejectedLogRecords()), diff --git a/exporter/coralogixexporter/metrics_client.go b/exporter/coralogixexporter/metrics_client.go index 9d97696b72123..de60ad0254916 100644 --- a/exporter/coralogixexporter/metrics_client.go +++ b/exporter/coralogixexporter/metrics_client.go @@ -92,7 +92,7 @@ func (e *metricsExporter) pushMetrics(ctx context.Context, md pmetric.Metrics) e } partialSuccess := resp.PartialSuccess() - if !(partialSuccess.ErrorMessage() == "" && partialSuccess.RejectedDataPoints() == 0) { + if partialSuccess.ErrorMessage() != "" || partialSuccess.RejectedDataPoints() != 0 { e.settings.Logger.Error("Partial success response from Coralogix", zap.String("message", partialSuccess.ErrorMessage()), zap.Int64("rejected_data_points", partialSuccess.RejectedDataPoints()), diff --git a/exporter/coralogixexporter/profiles_client.go b/exporter/coralogixexporter/profiles_client.go index 35171d46ec7bc..ede0479a5b87f 100644 --- a/exporter/coralogixexporter/profiles_client.go +++ b/exporter/coralogixexporter/profiles_client.go @@ -91,7 +91,7 @@ func (e *profilesExporter) pushProfiles(ctx context.Context, md pprofile.Profile } partialSuccess := resp.PartialSuccess() - if !(partialSuccess.ErrorMessage() == "" && partialSuccess.RejectedProfiles() == 0) { + if partialSuccess.ErrorMessage() != "" || partialSuccess.RejectedProfiles() != 0 { e.settings.Logger.Error("Partial success response from Coralogix", zap.String("message", partialSuccess.ErrorMessage()), zap.Int64("rejected_profiles", partialSuccess.RejectedProfiles()), diff --git a/exporter/coralogixexporter/traces_client.go b/exporter/coralogixexporter/traces_client.go index 8717e8d17cf69..4cee0f1df3dfc 100644 --- a/exporter/coralogixexporter/traces_client.go +++ b/exporter/coralogixexporter/traces_client.go @@ -89,7 +89,7 @@ func (e *tracesExporter) pushTraces(ctx context.Context, td ptrace.Traces) error } partialSuccess := resp.PartialSuccess() - if !(partialSuccess.ErrorMessage() == "" && partialSuccess.RejectedSpans() == 0) { + if partialSuccess.ErrorMessage() != "" || partialSuccess.RejectedSpans() != 0 { e.settings.Logger.Error("Partial success response from Coralogix", zap.String("message", partialSuccess.ErrorMessage()), zap.Int64("rejected_spans", partialSuccess.RejectedSpans()), diff --git a/exporter/datadogexporter/agent_components.go b/exporter/datadogexporter/agent_components.go index c9658846be1be..db9f7f18fd3ce 100644 --- a/exporter/datadogexporter/agent_components.go +++ b/exporter/datadogexporter/agent_components.go @@ -37,7 +37,7 @@ func newConfigComponent(set component.TelemetrySettings, cfg *datadogconfig.Conf pkgconfig.Set("logs_config.batch_wait", cfg.Logs.BatchWait, pkgconfigmodel.SourceFile) pkgconfig.Set("logs_config.use_compression", cfg.Logs.UseCompression, pkgconfigmodel.SourceFile) pkgconfig.Set("logs_config.compression_level", cfg.Logs.CompressionLevel, pkgconfigmodel.SourceFile) - pkgconfig.Set("logs_config.logs_dd_url", cfg.Logs.TCPAddrConfig.Endpoint, pkgconfigmodel.SourceFile) + pkgconfig.Set("logs_config.logs_dd_url", cfg.Logs.Endpoint, pkgconfigmodel.SourceFile) pkgconfig.Set("logs_config.auditor_ttl", pkgconfigsetup.DefaultAuditorTTL, pkgconfigmodel.SourceDefault) pkgconfig.Set("logs_config.batch_max_content_size", pkgconfigsetup.DefaultBatchMaxContentSize, pkgconfigmodel.SourceDefault) pkgconfig.Set("logs_config.batch_max_size", pkgconfigsetup.DefaultBatchMaxSize, pkgconfigmodel.SourceDefault) diff --git a/exporter/datadogexporter/factory.go b/exporter/datadogexporter/factory.go index 083a2116e15c4..789c8aef51b29 100644 --- a/exporter/datadogexporter/factory.go +++ b/exporter/datadogexporter/factory.go @@ -251,7 +251,7 @@ func (f *factory) createMetricsExporter( set exporter.Settings, c component.Config, ) (exporter.Metrics, error) { - cfg := checkAndCastConfig(c, set.TelemetrySettings.Logger) + cfg := checkAndCastConfig(c, set.Logger) hostProvider, err := f.SourceProvider(set.TelemetrySettings, cfg.Hostname, cfg.HostMetadata.GetSourceTimeout()) if err != nil { return nil, fmt.Errorf("failed to build hostname provider: %w", err) @@ -404,7 +404,7 @@ func (f *factory) createTracesExporter( set exporter.Settings, c component.Config, ) (exporter.Traces, error) { - cfg := checkAndCastConfig(c, set.TelemetrySettings.Logger) + cfg := checkAndCastConfig(c, set.Logger) if noAPMStatsFeatureGate.IsEnabled() { set.Logger.Info( "Trace metrics are now disabled in the Datadog Exporter by default. To continue receiving Trace Metrics, configure the Datadog Connector or disable the feature gate.", @@ -506,7 +506,7 @@ func (f *factory) createLogsExporter( set exporter.Settings, c component.Config, ) (exporter.Logs, error) { - cfg := checkAndCastConfig(c, set.TelemetrySettings.Logger) + cfg := checkAndCastConfig(c, set.Logger) if cfg.Logs.DumpPayloads && isLogsAgentExporterEnabled() { set.Logger.Warn("logs::dump_payloads is not valid when the exporter.datadogexporter.UseLogsAgentExporter feature gate is enabled") diff --git a/exporter/datadogexporter/factory_test.go b/exporter/datadogexporter/factory_test.go index 994527902b62d..54dc9c2e25d6d 100644 --- a/exporter/datadogexporter/factory_test.go +++ b/exporter/datadogexporter/factory_test.go @@ -80,7 +80,7 @@ func TestCreateAPIMetricsExporter(t *testing.T) { require.NoError(t, sub.Unmarshal(cfg)) c := cfg.(*datadogconfig.Config) - c.Metrics.TCPAddrConfig.Endpoint = server.URL + c.Metrics.Endpoint = server.URL c.HostMetadata.Enabled = false ctx := context.Background() @@ -116,7 +116,7 @@ func TestCreateAPIExporterFailOnInvalidKey_Zorkian(t *testing.T) { // Use the mock server for API key validation c := cfg.(*datadogconfig.Config) - c.Metrics.TCPAddrConfig.Endpoint = server.URL + c.Metrics.Endpoint = server.URL c.HostMetadata.Enabled = false t.Run("true", func(t *testing.T) { @@ -200,7 +200,7 @@ func TestCreateAPIExporterFailOnInvalidKey(t *testing.T) { // Use the mock server for API key validation c := cfg.(*datadogconfig.Config) - c.Metrics.TCPAddrConfig.Endpoint = server.URL + c.Metrics.Endpoint = server.URL c.HostMetadata.Enabled = false t.Run("true", func(t *testing.T) { @@ -276,7 +276,7 @@ func TestCreateAPILogsExporter(t *testing.T) { require.NoError(t, sub.Unmarshal(cfg)) c := cfg.(*datadogconfig.Config) - c.Metrics.TCPAddrConfig.Endpoint = server.URL + c.Metrics.Endpoint = server.URL c.HostMetadata.Enabled = false ctx := context.Background() @@ -358,7 +358,7 @@ func TestStopExporters(t *testing.T) { require.NoError(t, sub.Unmarshal(cfg)) c := cfg.(*datadogconfig.Config) - c.Metrics.TCPAddrConfig.Endpoint = server.URL + c.Metrics.Endpoint = server.URL c.HostMetadata.Enabled = false ctx := context.Background() diff --git a/exporter/datadogexporter/logs_exporter.go b/exporter/datadogexporter/logs_exporter.go index 9310ff57b5162..128b953ae08a9 100644 --- a/exporter/datadogexporter/logs_exporter.go +++ b/exporter/datadogexporter/logs_exporter.go @@ -74,12 +74,12 @@ func newLogsExporter( if isMetricExportV2Enabled() { apiClient := clientutil.CreateAPIClient( params.BuildInfo, - cfg.Metrics.TCPAddrConfig.Endpoint, + cfg.Metrics.Endpoint, cfg.ClientConfig) go func() { errchan <- clientutil.ValidateAPIKey(ctx, string(cfg.API.Key), params.Logger, apiClient) }() metricsAPI = datadogV2.NewMetricsApi(apiClient) } else { - client := clientutil.CreateZorkianClient(string(cfg.API.Key), cfg.Metrics.TCPAddrConfig.Endpoint) + client := clientutil.CreateZorkianClient(string(cfg.API.Key), cfg.Metrics.Endpoint) go func() { errchan <- clientutil.ValidateAPIKeyZorkian(params.Logger, client) }() } // validate the apiKey @@ -93,7 +93,7 @@ func newLogsExporter( if err != nil { return nil, fmt.Errorf("failed to create logs translator: %w", err) } - s := logs.NewSender(cfg.Logs.TCPAddrConfig.Endpoint, params.Logger, cfg.ClientConfig, cfg.Logs.DumpPayloads, string(cfg.API.Key)) + s := logs.NewSender(cfg.Logs.Endpoint, params.Logger, cfg.ClientConfig, cfg.Logs.DumpPayloads, string(cfg.API.Key)) scrubber := scrub.NewScrubber() return &logsExporter{ params: params, diff --git a/exporter/datadogexporter/metrics_exporter.go b/exporter/datadogexporter/metrics_exporter.go index ffd76b1052fe3..dbc881c67c02f 100644 --- a/exporter/datadogexporter/metrics_exporter.go +++ b/exporter/datadogexporter/metrics_exporter.go @@ -70,7 +70,7 @@ func newMetricsExporter( options = append(options, otlpmetrics.WithFallbackSourceProvider(sourceProvider)) options = append(options, otlpmetrics.WithStatsOut(statsOut)) if pkgdatadog.MetricRemappingDisabledFeatureGate.IsEnabled() { - params.TelemetrySettings.Logger.Warn("Metric remapping is disabled in the Datadog exporter. OpenTelemetry metrics must be mapped to Datadog semantics before metrics are exported to Datadog (ex: via a processor).") + params.Logger.Warn("Metric remapping is disabled in the Datadog exporter. OpenTelemetry metrics must be mapped to Datadog semantics before metrics are exported to Datadog (ex: via a processor).") } else { options = append(options, otlpmetrics.WithRemapping()) } @@ -99,12 +99,12 @@ func newMetricsExporter( if isMetricExportV2Enabled() { apiClient := clientutil.CreateAPIClient( params.BuildInfo, - cfg.Metrics.TCPAddrConfig.Endpoint, + cfg.Metrics.Endpoint, cfg.ClientConfig) go func() { errchan <- clientutil.ValidateAPIKey(ctx, string(cfg.API.Key), params.Logger, apiClient) }() exporter.metricsAPI = datadogV2.NewMetricsApi(apiClient) } else { - client := clientutil.CreateZorkianClient(string(cfg.API.Key), cfg.Metrics.TCPAddrConfig.Endpoint) + client := clientutil.CreateZorkianClient(string(cfg.API.Key), cfg.Metrics.Endpoint) client.ExtraHeader["User-Agent"] = clientutil.UserAgent(params.BuildInfo) client.HttpClient = clientutil.NewHTTPClient(cfg.ClientConfig) go func() { errchan <- clientutil.ValidateAPIKeyZorkian(params.Logger, client) }() @@ -127,7 +127,7 @@ func (exp *metricsExporter) pushSketches(ctx context.Context, sl sketches.Sketch req, err := http.NewRequestWithContext(ctx, http.MethodPost, - exp.cfg.Metrics.TCPAddrConfig.Endpoint+sketches.SketchSeriesEndpoint, + exp.cfg.Metrics.Endpoint+sketches.SketchSeriesEndpoint, bytes.NewBuffer(payload), ) if err != nil { diff --git a/exporter/datadogexporter/traces_exporter.go b/exporter/datadogexporter/traces_exporter.go index 8fe203926223a..0b6b782d7a76f 100644 --- a/exporter/datadogexporter/traces_exporter.go +++ b/exporter/datadogexporter/traces_exporter.go @@ -86,12 +86,12 @@ func newTracesExporter( if isMetricExportV2Enabled() { apiClient := clientutil.CreateAPIClient( params.BuildInfo, - cfg.Metrics.TCPAddrConfig.Endpoint, + cfg.Metrics.Endpoint, cfg.ClientConfig) go func() { errchan <- clientutil.ValidateAPIKey(ctx, string(cfg.API.Key), params.Logger, apiClient) }() exp.metricsAPI = datadogV2.NewMetricsApi(apiClient) } else { - client := clientutil.CreateZorkianClient(string(cfg.API.Key), cfg.Metrics.TCPAddrConfig.Endpoint) + client := clientutil.CreateZorkianClient(string(cfg.API.Key), cfg.Metrics.Endpoint) go func() { errchan <- clientutil.ValidateAPIKeyZorkian(params.Logger, client) }() exp.client = client } @@ -224,7 +224,7 @@ func newTraceAgentConfig(ctx context.Context, params exporter.Settings, cfg *dat acfg.Ignore["resource"] = cfg.Traces.IgnoreResources acfg.ReceiverEnabled = false // disable HTTP receiver acfg.AgentVersion = fmt.Sprintf("datadogexporter-%s-%s", params.BuildInfo.Command, params.BuildInfo.Version) - acfg.SkipSSLValidation = cfg.ClientConfig.TLSSetting.InsecureSkipVerify + acfg.SkipSSLValidation = cfg.TLSSetting.InsecureSkipVerify acfg.ComputeStatsBySpanKind = cfg.Traces.ComputeStatsBySpanKind acfg.PeerTagsAggregation = cfg.Traces.PeerTagsAggregation acfg.PeerTags = cfg.Traces.PeerTags diff --git a/exporter/datadogexporter/traces_exporter_test.go b/exporter/datadogexporter/traces_exporter_test.go index 4636e6013434a..53b1660ed93b1 100644 --- a/exporter/datadogexporter/traces_exporter_test.go +++ b/exporter/datadogexporter/traces_exporter_test.go @@ -327,7 +327,7 @@ func TestNewTracesExporter(t *testing.T) { cfg := &datadogconfig.Config{} cfg.API.Key = "ddog_32_characters_long_api_key1" - cfg.Metrics.TCPAddrConfig.Endpoint = metricsServer.URL + cfg.Metrics.Endpoint = metricsServer.URL params := exportertest.NewNopSettings(metadata.Type) // The client should have been created correctly diff --git a/exporter/datasetexporter/config.go b/exporter/datasetexporter/config.go index 673ee2a70d62d..ce004354c7a83 100644 --- a/exporter/datasetexporter/config.go +++ b/exporter/datasetexporter/config.go @@ -217,21 +217,21 @@ func (c *Config) convert() *ExporterConfig { Endpoint: c.DatasetURL, Tokens: datasetConfig.DataSetTokens{WriteLog: string(c.APIKey)}, BufferSettings: buffer_config.DataSetBufferSettings{ - MaxLifetime: c.BufferSettings.MaxLifetime, - PurgeOlderThan: c.BufferSettings.PurgeOlderThan, + MaxLifetime: c.MaxLifetime, + PurgeOlderThan: c.PurgeOlderThan, MaxSize: buffer.LimitBufferSize, - GroupBy: c.BufferSettings.GroupBy, - RetryInitialInterval: c.BufferSettings.RetryInitialInterval, - RetryMaxInterval: c.BufferSettings.RetryMaxInterval, - RetryMaxElapsedTime: c.BufferSettings.RetryMaxElapsedTime, + GroupBy: c.GroupBy, + RetryInitialInterval: c.RetryInitialInterval, + RetryMaxInterval: c.RetryMaxInterval, + RetryMaxElapsedTime: c.RetryMaxElapsedTime, RetryMultiplier: backoff.DefaultMultiplier, RetryRandomizationFactor: backoff.DefaultRandomizationFactor, - RetryShutdownTimeout: c.BufferSettings.RetryShutdownTimeout, - MaxParallelOutgoing: c.BufferSettings.MaxParallelOutgoing, + RetryShutdownTimeout: c.RetryShutdownTimeout, + MaxParallelOutgoing: c.MaxParallelOutgoing, }, ServerHostSettings: server_host_config.DataSetServerHostSettings{ - UseHostName: c.ServerHostSettings.UseHostName, - ServerHost: c.ServerHostSettings.ServerHost, + UseHostName: c.UseHostName, + ServerHost: c.ServerHost, }, Debug: c.Debug, }, diff --git a/exporter/datasetexporter/config_test.go b/exporter/datasetexporter/config_test.go index c5b407276a742..7d89bcabb571d 100644 --- a/exporter/datasetexporter/config_test.go +++ b/exporter/datasetexporter/config_test.go @@ -38,15 +38,15 @@ func TestConfigUseDefaults(t *testing.T) { assert.Equal(t, "https://example.com", config.DatasetURL) assert.Equal(t, "secret", string(config.APIKey)) assert.Equal(t, bufferMaxLifetime, config.MaxLifetime) - assert.Equal(t, logsExportResourceInfoDefault, config.LogsSettings.ExportResourceInfo) - assert.Equal(t, logsExportResourcePrefixDefault, config.LogsSettings.ExportResourcePrefix) - assert.Equal(t, logsExportScopeInfoDefault, config.LogsSettings.ExportScopeInfo) - assert.Equal(t, logsExportScopePrefixDefault, config.LogsSettings.ExportScopePrefix) - assert.Equal(t, logsDecomposeComplexMessageFieldDefault, config.LogsSettings.DecomposeComplexMessageField) - assert.Equal(t, exportSeparatorDefault, config.LogsSettings.exportSettings.ExportSeparator) - assert.Equal(t, exportDistinguishingSuffix, config.LogsSettings.exportSettings.ExportDistinguishingSuffix) - assert.Equal(t, exportSeparatorDefault, config.TracesSettings.exportSettings.ExportSeparator) - assert.Equal(t, exportDistinguishingSuffix, config.TracesSettings.exportSettings.ExportDistinguishingSuffix) + assert.Equal(t, logsExportResourceInfoDefault, config.ExportResourceInfo) + assert.Equal(t, logsExportResourcePrefixDefault, config.ExportResourcePrefix) + assert.Equal(t, logsExportScopeInfoDefault, config.ExportScopeInfo) + assert.Equal(t, logsExportScopePrefixDefault, config.ExportScopePrefix) + assert.Equal(t, logsDecomposeComplexMessageFieldDefault, config.DecomposeComplexMessageField) + assert.Equal(t, exportSeparatorDefault, config.LogsSettings.ExportSeparator) + assert.Equal(t, exportDistinguishingSuffix, config.LogsSettings.ExportDistinguishingSuffix) + assert.Equal(t, exportSeparatorDefault, config.TracesSettings.ExportSeparator) + assert.Equal(t, exportDistinguishingSuffix, config.TracesSettings.ExportDistinguishingSuffix) } func TestConfigValidate(t *testing.T) { @@ -112,7 +112,7 @@ func TestConfigUseProvidedExportResourceInfoValue(t *testing.T) { }) err := config.Unmarshal(configMap) assert.NoError(t, err) - assert.True(t, config.LogsSettings.ExportResourceInfo) + assert.True(t, config.ExportResourceInfo) } func TestConfigUseProvidedExportScopeInfoValue(t *testing.T) { @@ -127,5 +127,5 @@ func TestConfigUseProvidedExportScopeInfoValue(t *testing.T) { }) err := config.Unmarshal(configMap) assert.NoError(t, err) - assert.False(t, config.LogsSettings.ExportScopeInfo) + assert.False(t, config.ExportScopeInfo) } diff --git a/exporter/dorisexporter/config.go b/exporter/dorisexporter/config.go index 89cdfd33a8f9b..7e5bc9acc7b06 100644 --- a/exporter/dorisexporter/config.go +++ b/exporter/dorisexporter/config.go @@ -94,13 +94,13 @@ func (cfg *Config) Validate() (err error) { if !re.MatchString(cfg.Database) { err = errors.Join(err, errors.New("database name must be alphanumeric and underscore")) } - if !re.MatchString(cfg.Table.Logs) { + if !re.MatchString(cfg.Logs) { err = errors.Join(err, errors.New("logs table name must be alphanumeric and underscore")) } - if !re.MatchString(cfg.Table.Traces) { + if !re.MatchString(cfg.Traces) { err = errors.Join(err, errors.New("traces table name must be alphanumeric and underscore")) } - if !re.MatchString(cfg.Table.Metrics) { + if !re.MatchString(cfg.Metrics) { err = errors.Join(err, errors.New("metrics table name must be alphanumeric and underscore")) } diff --git a/exporter/dorisexporter/exporter_common.go b/exporter/dorisexporter/exporter_common.go index d051299c0f84d..1264bde9a390b 100644 --- a/exporter/dorisexporter/exporter_common.go +++ b/exporter/dorisexporter/exporter_common.go @@ -102,8 +102,8 @@ func streamLoadRequest(ctx context.Context, cfg *Config, table string, data []by if groupCommit == "" || groupCommit == "off_mode" { req.Header.Set("label", label) } - if cfg.ClientConfig.Timeout != 0 { - req.Header.Set("timeout", fmt.Sprintf("%d", cfg.ClientConfig.Timeout/time.Second)) + if cfg.Timeout != 0 { + req.Header.Set("timeout", fmt.Sprintf("%d", cfg.Timeout/time.Second)) } req.SetBasicAuth(cfg.Username, string(cfg.Password)) @@ -111,7 +111,7 @@ func streamLoadRequest(ctx context.Context, cfg *Config, table string, data []by } func createDorisHTTPClient(ctx context.Context, cfg *Config, host component.Host, settings component.TelemetrySettings) (*http.Client, error) { - client, err := cfg.ClientConfig.ToClient(ctx, host, settings) + client, err := cfg.ToClient(ctx, host, settings) if err != nil { return nil, err } diff --git a/exporter/dorisexporter/exporter_logs.go b/exporter/dorisexporter/exporter_logs.go index 5b185236ef436..b6294039054a5 100644 --- a/exporter/dorisexporter/exporter_logs.go +++ b/exporter/dorisexporter/exporter_logs.go @@ -69,13 +69,13 @@ func (e *logsExporter) start(ctx context.Context, host component.Host) error { return err } - ddl := fmt.Sprintf(logsDDL, e.cfg.Table.Logs, e.cfg.propertiesStr()) + ddl := fmt.Sprintf(logsDDL, e.cfg.Logs, e.cfg.propertiesStr()) _, err = conn.ExecContext(ctx, ddl) if err != nil { return err } - view := fmt.Sprintf(logsView, e.cfg.Table.Logs, e.cfg.Table.Logs) + view := fmt.Sprintf(logsView, e.cfg.Logs, e.cfg.Logs) _, err = conn.ExecContext(ctx, view) if err != nil { e.logger.Warn("failed to create materialized view", zap.Error(err)) @@ -94,7 +94,7 @@ func (e *logsExporter) shutdown(_ context.Context) error { } func (e *logsExporter) pushLogData(ctx context.Context, ld plog.Logs) error { - label := generateLabel(e.cfg, e.cfg.Table.Logs) + label := generateLabel(e.cfg, e.cfg.Logs) logs := make([]*dLog, 0, ld.LogRecordCount()) for i := 0; i < ld.ResourceLogs().Len(); i++ { @@ -147,7 +147,7 @@ func (e *logsExporter) pushLogDataInternal(ctx context.Context, logs []*dLog, la return err } - req, err := streamLoadRequest(ctx, e.cfg, e.cfg.Table.Logs, marshal, label) + req, err := streamLoadRequest(ctx, e.cfg, e.cfg.Logs, marshal, label) if err != nil { return err } diff --git a/exporter/dorisexporter/exporter_metrics.go b/exporter/dorisexporter/exporter_metrics.go index d1a9ab7251e32..5462c14563c36 100644 --- a/exporter/dorisexporter/exporter_metrics.go +++ b/exporter/dorisexporter/exporter_metrics.go @@ -59,7 +59,7 @@ func (e *metricsExporter) start(ctx context.Context, host component.Host) error } for _, ddlTemplate := range ddls { - ddl := fmt.Sprintf(ddlTemplate, e.cfg.Table.Metrics, e.cfg.propertiesStr()) + ddl := fmt.Sprintf(ddlTemplate, e.cfg.Metrics, e.cfg.propertiesStr()) _, err = conn.ExecContext(ctx, ddl) if err != nil { return err @@ -75,7 +75,7 @@ func (e *metricsExporter) start(ctx context.Context, host component.Host) error } for _, model := range models { - table := e.cfg.Table.Metrics + model.tableSuffix() + table := e.cfg.Metrics + model.tableSuffix() view := fmt.Sprintf(metricsView, table, table) _, err = conn.ExecContext(ctx, view) if err != nil { @@ -247,7 +247,7 @@ func (e *metricsExporter) pushMetricDataInternal(ctx context.Context, metrics me return err } - req, err := streamLoadRequest(ctx, e.cfg, e.cfg.Table.Metrics+metrics.tableSuffix(), marshal, metrics.label()) + req, err := streamLoadRequest(ctx, e.cfg, e.cfg.Metrics+metrics.tableSuffix(), marshal, metrics.label()) if err != nil { return err } @@ -319,5 +319,5 @@ func (e *metricsExporter) getExemplarValue(ep pmetric.Exemplar) float64 { } func (e *metricsExporter) generateMetricLabel(m metricModel) string { - return generateLabel(e.cfg, e.cfg.Table.Metrics+m.tableSuffix()) + return generateLabel(e.cfg, e.cfg.Metrics+m.tableSuffix()) } diff --git a/exporter/dorisexporter/exporter_traces.go b/exporter/dorisexporter/exporter_traces.go index 2fbb3e3364390..26a662cdb3daa 100644 --- a/exporter/dorisexporter/exporter_traces.go +++ b/exporter/dorisexporter/exporter_traces.go @@ -97,19 +97,19 @@ func (e *tracesExporter) start(ctx context.Context, host component.Host) error { return err } - ddl := fmt.Sprintf(tracesDDL, e.cfg.Table.Traces, e.cfg.propertiesStr()) + ddl := fmt.Sprintf(tracesDDL, e.cfg.Traces, e.cfg.propertiesStr()) _, err = conn.ExecContext(ctx, ddl) if err != nil { return err } - view := fmt.Sprintf(tracesView, e.cfg.Table.Traces, e.cfg.Table.Traces) + view := fmt.Sprintf(tracesView, e.cfg.Traces, e.cfg.Traces) _, err = conn.ExecContext(ctx, view) if err != nil { e.logger.Warn("failed to create materialized view", zap.Error(err)) } - ddl = fmt.Sprintf(tracesGraphDDL, e.cfg.Table.Traces, e.cfg.propertiesStr()) + ddl = fmt.Sprintf(tracesGraphDDL, e.cfg.Traces, e.cfg.propertiesStr()) _, err = conn.ExecContext(ctx, ddl) if err != nil { return err @@ -140,7 +140,7 @@ func (e *tracesExporter) shutdown(_ context.Context) error { } func (e *tracesExporter) pushTraceData(ctx context.Context, td ptrace.Traces) error { - label := generateLabel(e.cfg, e.cfg.Table.Traces) + label := generateLabel(e.cfg, e.cfg.Traces) traces := make([]*dTrace, 0, td.SpanCount()) for i := 0; i < td.ResourceSpans().Len(); i++ { @@ -229,7 +229,7 @@ func (e *tracesExporter) pushTraceDataInternal(ctx context.Context, traces []*dT return err } - req, err := streamLoadRequest(ctx, e.cfg, e.cfg.Table.Traces, marshal, label) + req, err := streamLoadRequest(ctx, e.cfg, e.cfg.Traces, marshal, label) if err != nil { return err } @@ -274,7 +274,7 @@ func (e *tracesExporter) formatDropTraceGraphJob() string { return fmt.Sprintf( "DROP JOB where jobName = '%s:%s_graph_job';", e.cfg.Database, - e.cfg.Table.Traces, + e.cfg.Traces, ) } @@ -282,9 +282,9 @@ func (e *tracesExporter) formatTraceGraphJob() string { return fmt.Sprintf( tracesGraphJob, e.cfg.Database, - e.cfg.Table.Traces, - e.cfg.Table.Traces, - e.cfg.Table.Traces, - e.cfg.Table.Traces, + e.cfg.Traces, + e.cfg.Traces, + e.cfg.Traces, + e.cfg.Traces, ) } diff --git a/exporter/elasticsearchexporter/bulkindexer.go b/exporter/elasticsearchexporter/bulkindexer.go index db93d4be66c41..8b1ba003bbe37 100644 --- a/exporter/elasticsearchexporter/bulkindexer.go +++ b/exporter/elasticsearchexporter/bulkindexer.go @@ -407,7 +407,7 @@ func (b *bulkIndexers) start( for _, mode := range allowedMappingModes { var bi bulkIndexer - bi, err = newBulkIndexer(set.TelemetrySettings.Logger, esClient, cfg, mode == MappingOTel) + bi, err = newBulkIndexer(set.Logger, esClient, cfg, mode == MappingOTel) if err != nil { return err } diff --git a/exporter/elasticsearchexporter/esclient.go b/exporter/elasticsearchexporter/esclient.go index 2701b041c865b..73e5bc4263af4 100644 --- a/exporter/elasticsearchexporter/esclient.go +++ b/exporter/elasticsearchexporter/esclient.go @@ -92,7 +92,7 @@ func newElasticsearchClient( telemetry component.TelemetrySettings, userAgent string, ) (esapi.Transport, error) { - httpClient, err := config.ClientConfig.ToClient(ctx, host, telemetry) + httpClient, err := config.ToClient(ctx, host, telemetry) if err != nil { return nil, err } diff --git a/exporter/elasticsearchexporter/integrationtest/datareceiver.go b/exporter/elasticsearchexporter/integrationtest/datareceiver.go index ab8a1d1cf2e77..b6c32a030a538 100644 --- a/exporter/elasticsearchexporter/integrationtest/datareceiver.go +++ b/exporter/elasticsearchexporter/integrationtest/datareceiver.go @@ -99,7 +99,7 @@ func (es *esDataReceiver) Start(tc consumer.Traces, mc consumer.Metrics, lc cons return fmt.Errorf("invalid ES URL specified %s: %w", es.endpoint, err) } cfg := factory.CreateDefaultConfig().(*config) - cfg.ServerConfig.Endpoint = esURL.Host + cfg.Endpoint = esURL.Host cfg.DecodeBulkRequests = es.decodeBulkRequest set := receivertest.NewNopSettings(metadata.Type) diff --git a/exporter/elasticsearchexporter/integrationtest/exporter_bench_test.go b/exporter/elasticsearchexporter/integrationtest/exporter_bench_test.go index ae65e4fbe2a94..372257696723b 100644 --- a/exporter/elasticsearchexporter/integrationtest/exporter_bench_test.go +++ b/exporter/elasticsearchexporter/integrationtest/exporter_bench_test.go @@ -58,7 +58,7 @@ func benchmarkLogs(b *testing.B, batchSize int, mappingMode string) { defer cancel() exporterSettings := exportertest.NewNopSettings(metadata.Type) - exporterSettings.TelemetrySettings.Logger = zaptest.NewLogger(b, zaptest.Level(zap.WarnLevel)) + exporterSettings.Logger = zaptest.NewLogger(b, zaptest.Level(zap.WarnLevel)) runnerCfg := prepareBenchmark(b, batchSize, mappingMode) exporter, err := runnerCfg.factory.CreateLogs( ctx, exporterSettings, runnerCfg.esCfg, @@ -88,7 +88,7 @@ func benchmarkMetrics(b *testing.B, batchSize int, mappingMode string) { defer cancel() exporterSettings := exportertest.NewNopSettings(metadata.Type) - exporterSettings.TelemetrySettings.Logger = zaptest.NewLogger(b, zaptest.Level(zap.WarnLevel)) + exporterSettings.Logger = zaptest.NewLogger(b, zaptest.Level(zap.WarnLevel)) runnerCfg := prepareBenchmark(b, batchSize, mappingMode) exporter, err := runnerCfg.factory.CreateMetrics( ctx, exporterSettings, runnerCfg.esCfg, @@ -118,7 +118,7 @@ func benchmarkTraces(b *testing.B, batchSize int, mappingMode string) { defer cancel() exporterSettings := exportertest.NewNopSettings(metadata.Type) - exporterSettings.TelemetrySettings.Logger = zaptest.NewLogger(b, zaptest.Level(zap.WarnLevel)) + exporterSettings.Logger = zaptest.NewLogger(b, zaptest.Level(zap.WarnLevel)) runnerCfg := prepareBenchmark(b, batchSize, mappingMode) exporter, err := runnerCfg.factory.CreateTraces( ctx, exporterSettings, runnerCfg.esCfg, diff --git a/exporter/elasticsearchexporter/internal/datapoints/histogram.go b/exporter/elasticsearchexporter/internal/datapoints/histogram.go index 1f681e25a7709..4be90e3b2fc3d 100644 --- a/exporter/elasticsearchexporter/internal/datapoints/histogram.go +++ b/exporter/elasticsearchexporter/internal/datapoints/histogram.go @@ -45,7 +45,7 @@ func (dp Histogram) DynamicTemplate(_ pmetric.Metric) string { } func (dp Histogram) DocCount() uint64 { - return dp.HistogramDataPoint.Count() + return dp.Count() } func (dp Histogram) Metric() pmetric.Metric { diff --git a/exporter/elasticsearchexporter/internal/datapoints/number.go b/exporter/elasticsearchexporter/internal/datapoints/number.go index 7816bb24da0c7..4acfe17b1d04f 100644 --- a/exporter/elasticsearchexporter/internal/datapoints/number.go +++ b/exporter/elasticsearchexporter/internal/datapoints/number.go @@ -46,7 +46,7 @@ func (dp Number) Value() (pcommon.Value, error) { func (dp Number) DynamicTemplate(metric pmetric.Metric) string { switch metric.Type() { case pmetric.MetricTypeSum: - switch dp.NumberDataPoint.ValueType() { + switch dp.ValueType() { case pmetric.NumberDataPointValueTypeDouble: if metric.Sum().IsMonotonic() { return "counter_double" @@ -61,7 +61,7 @@ func (dp Number) DynamicTemplate(metric pmetric.Metric) string { return "" // NumberDataPointValueTypeEmpty should already be discarded in numberToValue } case pmetric.MetricTypeGauge: - switch dp.NumberDataPoint.ValueType() { + switch dp.ValueType() { case pmetric.NumberDataPointValueTypeDouble: return "gauge_double" case pmetric.NumberDataPointValueTypeInt: diff --git a/exporter/faroexporter/config_test.go b/exporter/faroexporter/config_test.go index 101dc5f890bfb..0c22fb49c4ccf 100644 --- a/exporter/faroexporter/config_test.go +++ b/exporter/faroexporter/config_test.go @@ -30,7 +30,7 @@ func TestLoadConfig(t *testing.T) { require.NoError(t, sub.Unmarshal(cfg)) assert.NoError(t, componenttest.CheckConfigStruct(cfg)) - assert.Equal(t, "https://faro.example.com/collect", cfg.(*Config).ClientConfig.Endpoint) + assert.Equal(t, "https://faro.example.com/collect", cfg.(*Config).Endpoint) } func TestValidateConfig(t *testing.T) { diff --git a/exporter/faroexporter/exporter.go b/exporter/faroexporter/exporter.go index 73a94725c81a0..1d323e6fceb32 100644 --- a/exporter/faroexporter/exporter.go +++ b/exporter/faroexporter/exporter.go @@ -66,7 +66,7 @@ func newExporter(cfg component.Config, set exporter.Settings) (*faroExporter, er } func (fe *faroExporter) start(ctx context.Context, host component.Host) error { - client, err := fe.config.ClientConfig.ToClient(ctx, host, fe.settings) + client, err := fe.config.ToClient(ctx, host, fe.settings) if err != nil { return err } diff --git a/exporter/faroexporter/exporter_test.go b/exporter/faroexporter/exporter_test.go index a7cc32d4a9206..0961714086fe8 100644 --- a/exporter/faroexporter/exporter_test.go +++ b/exporter/faroexporter/exporter_test.go @@ -43,7 +43,7 @@ func TestExporter_BaseTest(t *testing.T) { name: "deflate", config: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Compression = configcompression.TypeDeflate + cfg.Compression = configcompression.TypeDeflate return cfg }(), }, @@ -51,7 +51,7 @@ func TestExporter_BaseTest(t *testing.T) { name: "no compression", config: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.CompressionParams = configcompression.CompressionParams{ + cfg.CompressionParams = configcompression.CompressionParams{ Level: configcompression.Level(0), } return cfg diff --git a/exporter/faroexporter/factory_test.go b/exporter/faroexporter/factory_test.go index f5b7eed738cbb..ca622ad63a853 100644 --- a/exporter/faroexporter/factory_test.go +++ b/exporter/faroexporter/factory_test.go @@ -25,7 +25,7 @@ func TestCreateDefaultConfig(t *testing.T) { func TestCreateExporter(t *testing.T) { factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "https://faro.example.com/collect" + cfg.Endpoint = "https://faro.example.com/collect" set := exporter.Settings{ TelemetrySettings: componenttest.NewNopTelemetrySettings(), diff --git a/exporter/googlecloudexporter/config_test.go b/exporter/googlecloudexporter/config_test.go index 2b9b79075b847..a67750b38da6f 100644 --- a/exporter/googlecloudexporter/config_test.go +++ b/exporter/googlecloudexporter/config_test.go @@ -75,8 +75,8 @@ func TestLoadConfig(t *testing.T) { } func sanitize(cfg *Config) *Config { - cfg.Config.MetricConfig.MapMonitoredResource = nil - cfg.Config.LogConfig.MapMonitoredResource = nil - cfg.Config.MetricConfig.GetMetricName = nil + cfg.MetricConfig.MapMonitoredResource = nil + cfg.LogConfig.MapMonitoredResource = nil + cfg.MetricConfig.GetMetricName = nil return cfg } diff --git a/exporter/googlemanagedprometheusexporter/factory.go b/exporter/googlemanagedprometheusexporter/factory.go index 06e1d04352aa3..16afa6b610eff 100644 --- a/exporter/googlemanagedprometheusexporter/factory.go +++ b/exporter/googlemanagedprometheusexporter/factory.go @@ -53,7 +53,7 @@ func createMetricsExporter( cfg component.Config, ) (exporter.Metrics, error) { eCfg := cfg.(*Config) - mExp, err := collector.NewGoogleCloudMetricsExporter(ctx, eCfg.GMPConfig.toCollectorConfig(), params, eCfg.TimeoutSettings.Timeout) + mExp, err := collector.NewGoogleCloudMetricsExporter(ctx, eCfg.toCollectorConfig(), params, eCfg.TimeoutSettings.Timeout) if err != nil { return nil, err } diff --git a/exporter/influxdbexporter/logger.go b/exporter/influxdbexporter/logger.go index c052366d7359f..a1a5ae322ee10 100644 --- a/exporter/influxdbexporter/logger.go +++ b/exporter/influxdbexporter/logger.go @@ -21,5 +21,5 @@ func newZapInfluxLogger(logger *zap.Logger) common.Logger { } func (l zapInfluxLogger) Debug(msg string, kv ...any) { - l.SugaredLogger.Debugw(msg, kv...) + l.Debugw(msg, kv...) } diff --git a/exporter/influxdbexporter/writer.go b/exporter/influxdbexporter/writer.go index f2235d21314b1..aa38c7983a46f 100644 --- a/exporter/influxdbexporter/writer.go +++ b/exporter/influxdbexporter/writer.go @@ -64,7 +64,7 @@ func newInfluxHTTPWriter(logger common.Logger, config *Config, telemetrySettings } func composeWriteURL(config *Config) (string, error) { - writeURL, err := url.Parse(config.ClientConfig.Endpoint) + writeURL, err := url.Parse(config.Endpoint) if err != nil { return "", err } @@ -90,20 +90,20 @@ func composeWriteURL(config *Config) (string, error) { if config.V1Compatibility.Username != "" && config.V1Compatibility.Password != "" { basicAuth := base64.StdEncoding.EncodeToString( []byte(config.V1Compatibility.Username + ":" + string(config.V1Compatibility.Password))) - if config.ClientConfig.Headers == nil { - config.ClientConfig.Headers = make(map[string]configopaque.String, 1) + if config.Headers == nil { + config.Headers = make(map[string]configopaque.String, 1) } - config.ClientConfig.Headers["Authorization"] = configopaque.String("Basic " + basicAuth) + config.Headers["Authorization"] = configopaque.String("Basic " + basicAuth) } } else { queryValues.Set("org", config.Org) queryValues.Set("bucket", config.Bucket) if config.Token != "" { - if config.ClientConfig.Headers == nil { - config.ClientConfig.Headers = make(map[string]configopaque.String, 1) + if config.Headers == nil { + config.Headers = make(map[string]configopaque.String, 1) } - config.ClientConfig.Headers["Authorization"] = "Token " + config.Token + config.Headers["Authorization"] = "Token " + config.Token } } diff --git a/exporter/loadbalancingexporter/factory.go b/exporter/loadbalancingexporter/factory.go index 78506d7a5829d..f288c844d4452 100644 --- a/exporter/loadbalancingexporter/factory.go +++ b/exporter/loadbalancingexporter/factory.go @@ -71,7 +71,7 @@ func buildExporterResilienceOptions(options []exporterhelper.Option, cfg *Config if cfg.QueueSettings.Enabled { options = append(options, exporterhelper.WithQueue(cfg.QueueSettings)) } - if cfg.BackOffConfig.Enabled { + if cfg.Enabled { options = append(options, exporterhelper.WithRetry(cfg.BackOffConfig)) } diff --git a/exporter/logicmonitorexporter/logs_exporter.go b/exporter/logicmonitorexporter/logs_exporter.go index 398281fc4cf9f..db3e98e11ab13 100644 --- a/exporter/logicmonitorexporter/logs_exporter.go +++ b/exporter/logicmonitorexporter/logs_exporter.go @@ -46,7 +46,7 @@ func newLogsExporter(_ context.Context, cfg component.Config, set exporter.Setti } func (e *logExporter) start(ctx context.Context, host component.Host) error { - client, err := e.config.ClientConfig.ToClient(ctx, host, e.settings) + client, err := e.config.ToClient(ctx, host, e.settings) if err != nil { return fmt.Errorf("failed to create http client: %w", err) } diff --git a/exporter/logicmonitorexporter/traces_exporter.go b/exporter/logicmonitorexporter/traces_exporter.go index 328c2cef44c89..05f7fda527dda 100644 --- a/exporter/logicmonitorexporter/traces_exporter.go +++ b/exporter/logicmonitorexporter/traces_exporter.go @@ -34,7 +34,7 @@ func newTracesExporter(_ context.Context, cfg component.Config, set exporter.Set } func (e *tracesExporter) start(ctx context.Context, host component.Host) error { - client, err := e.config.ClientConfig.ToClient(ctx, host, e.settings) + client, err := e.config.ToClient(ctx, host, e.settings) if err != nil { return fmt.Errorf("failed to create http client: %w", err) } diff --git a/exporter/logzioexporter/config.go b/exporter/logzioexporter/config.go index f11f7bc6ef956..763256b2b2923 100644 --- a/exporter/logzioexporter/config.go +++ b/exporter/logzioexporter/config.go @@ -52,6 +52,6 @@ func (c *Config) checkAndWarnDeprecatedOptions(logger hclog.Logger) { if c.CustomEndpoint != "" { logger.Warn("You are using the deprecated `custom_endpoint` option that will be removed in the next release; please use `endpoint` configuration instead: https://github.com/open-telemetry/opentelemetry-collector/tree/main/config/confighttp") logger.Warn("Mapping `custom_endpoint` -> `Endpoint`") - c.ClientConfig.Endpoint = c.CustomEndpoint + c.Endpoint = c.CustomEndpoint } } diff --git a/exporter/logzioexporter/config_test.go b/exporter/logzioexporter/config_test.go index c21a2c6726534..64b58515bfe47 100644 --- a/exporter/logzioexporter/config_test.go +++ b/exporter/logzioexporter/config_test.go @@ -36,7 +36,7 @@ func TestLoadConfig(t *testing.T) { Region: "eu", } expected.BackOffConfig = configretry.NewDefaultBackOffConfig() - expected.BackOffConfig.MaxInterval = 5 * time.Second + expected.MaxInterval = 5 * time.Second expected.QueueSettings = exporterhelper.NewDefaultQueueConfig() expected.QueueSettings.Enabled = false clientConfig := confighttp.NewDefaultClientConfig() diff --git a/exporter/logzioexporter/exporter.go b/exporter/logzioexporter/exporter.go index 546bb753e1a9a..9b153b3c4a99d 100644 --- a/exporter/logzioexporter/exporter.go +++ b/exporter/logzioexporter/exporter.go @@ -72,7 +72,7 @@ func newLogzioTracesExporter(config *Config, set exporter.Settings) (exporter.Tr if err != nil { return nil, err } - exporter.config.ClientConfig.Endpoint, err = generateEndpoint(config) + exporter.config.Endpoint, err = generateEndpoint(config) if err != nil { return nil, err } @@ -95,7 +95,7 @@ func newLogzioLogsExporter(config *Config, set exporter.Settings) (exporter.Logs if err != nil { return nil, err } - exporter.config.ClientConfig.Endpoint, err = generateEndpoint(config) + exporter.config.Endpoint, err = generateEndpoint(config) if err != nil { return nil, err } @@ -114,7 +114,7 @@ func newLogzioLogsExporter(config *Config, set exporter.Settings) (exporter.Logs } func (exporter *logzioExporter) start(ctx context.Context, host component.Host) error { - client, err := exporter.config.ClientConfig.ToClient(ctx, host, exporter.settings) + client, err := exporter.config.ToClient(ctx, host, exporter.settings) if err != nil { return err } @@ -146,7 +146,7 @@ func (exporter *logzioExporter) pushLogData(ctx context.Context, ld plog.Logs) e } } } - err := exporter.export(ctx, exporter.config.ClientConfig.Endpoint, dataBuffer.Bytes()) + err := exporter.export(ctx, exporter.config.Endpoint, dataBuffer.Bytes()) // reset the data buffer after each export to prevent duplicated data dataBuffer.Reset() return err @@ -218,7 +218,7 @@ func (exporter *logzioExporter) pushTraceData(ctx context.Context, traces ptrace } } } - err := exporter.export(ctx, exporter.config.ClientConfig.Endpoint, dataBuffer.Bytes()) + err := exporter.export(ctx, exporter.config.Endpoint, dataBuffer.Bytes()) // reset the data buffer after each export to prevent duplicated data dataBuffer.Reset() return err diff --git a/exporter/logzioexporter/factory.go b/exporter/logzioexporter/factory.go index 463cfeb0f55ee..e960319d410cd 100644 --- a/exporter/logzioexporter/factory.go +++ b/exporter/logzioexporter/factory.go @@ -74,11 +74,11 @@ func getListenerURL(region string) string { func generateEndpoint(cfg *Config) (string, error) { defaultURL := fmt.Sprintf("%s/?token=%s", getListenerURL(""), string(cfg.Token)) switch { - case cfg.ClientConfig.Endpoint != "": - return cfg.ClientConfig.Endpoint, nil + case cfg.Endpoint != "": + return cfg.Endpoint, nil case cfg.Region != "": return fmt.Sprintf("%s/?token=%s", getListenerURL(cfg.Region), string(cfg.Token)), nil - case cfg.ClientConfig.Endpoint == "" && cfg.Region == "": + case cfg.Endpoint == "" && cfg.Region == "": return defaultURL, errors.New("failed to generate endpoint, Endpoint or Region must be set") default: return defaultURL, nil diff --git a/exporter/lokiexporter/exporter.go b/exporter/lokiexporter/exporter.go index de3a6b792f525..d482c042e77f2 100644 --- a/exporter/lokiexporter/exporter.go +++ b/exporter/lokiexporter/exporter.go @@ -89,12 +89,12 @@ func (l *lokiExporter) sendPushRequest(ctx context.Context, tenant string, reque return consumererror.NewPermanent(err) } - req, err := http.NewRequestWithContext(ctx, http.MethodPost, l.config.ClientConfig.Endpoint, bytes.NewReader(buf)) + req, err := http.NewRequestWithContext(ctx, http.MethodPost, l.config.Endpoint, bytes.NewReader(buf)) if err != nil { return consumererror.NewPermanent(err) } - for k, v := range l.config.ClientConfig.Headers { + for k, v := range l.config.Headers { req.Header.Set(k, string(v)) } req.Header.Set("Content-Type", "application/x-protobuf") @@ -143,7 +143,7 @@ func encode(pb proto.Message) ([]byte, error) { } func (l *lokiExporter) start(ctx context.Context, host component.Host) (err error) { - client, err := l.config.ClientConfig.ToClient(ctx, host, l.settings) + client, err := l.config.ToClient(ctx, host, l.settings) if err != nil { return err } diff --git a/exporter/mezmoexporter/exporter.go b/exporter/mezmoexporter/exporter.go index 8daf74f7c7ac7..f8b58822e166b 100644 --- a/exporter/mezmoexporter/exporter.go +++ b/exporter/mezmoexporter/exporter.go @@ -58,7 +58,7 @@ func (m *mezmoExporter) pushLogData(_ context.Context, ld plog.Logs) error { } func (m *mezmoExporter) start(ctx context.Context, host component.Host) (err error) { - m.client, err = m.config.ClientConfig.ToClient(ctx, host, m.settings) + m.client, err = m.config.ToClient(ctx, host, m.settings) return err } diff --git a/exporter/opencensusexporter/opencensus.go b/exporter/opencensusexporter/opencensus.go index 0cedb5130ab65..aba28451f528b 100644 --- a/exporter/opencensusexporter/opencensus.go +++ b/exporter/opencensusexporter/opencensus.go @@ -64,7 +64,7 @@ func newOcExporter(_ context.Context, cfg *Config, settings component.TelemetryS metadata: metadata.New(nil), settings: settings, } - for k, v := range cfg.ClientConfig.Headers { + for k, v := range cfg.Headers { oce.metadata.Set(k, string(v)) } return oce, nil @@ -72,7 +72,7 @@ func newOcExporter(_ context.Context, cfg *Config, settings component.TelemetryS // start creates the gRPC client Connection func (oce *ocExporter) start(ctx context.Context, host component.Host) error { - clientConn, err := oce.cfg.ClientConfig.ToClientConn(ctx, host, oce.settings) + clientConn, err := oce.cfg.ToClientConn(ctx, host, oce.settings) if err != nil { return err } diff --git a/exporter/opencensusexporter/opencensus_test.go b/exporter/opencensusexporter/opencensus_test.go index 0467c67cb8c0f..5bfe5483eaf0e 100644 --- a/exporter/opencensusexporter/opencensus_test.go +++ b/exporter/opencensusexporter/opencensus_test.go @@ -29,7 +29,7 @@ func TestSendTraces(t *testing.T) { rFactory := opencensusreceiver.NewFactory() rCfg := rFactory.CreateDefaultConfig().(*opencensusreceiver.Config) endpoint := testutil.GetAvailableLocalAddress(t) - rCfg.ServerConfig.NetAddr.Endpoint = endpoint + rCfg.NetAddr.Endpoint = endpoint set := receivertest.NewNopSettings(metadata.Type) recv, err := rFactory.CreateTraces(context.Background(), set, rCfg, sink) assert.NoError(t, err) @@ -128,7 +128,7 @@ func TestSendMetrics(t *testing.T) { rFactory := opencensusreceiver.NewFactory() rCfg := rFactory.CreateDefaultConfig().(*opencensusreceiver.Config) endpoint := testutil.GetAvailableLocalAddress(t) - rCfg.ServerConfig.NetAddr.Endpoint = endpoint + rCfg.NetAddr.Endpoint = endpoint set := receivertest.NewNopSettings(metadata.Type) recv, err := rFactory.CreateMetrics(context.Background(), set, rCfg, sink) assert.NoError(t, err) diff --git a/exporter/opensearchexporter/config.go b/exporter/opensearchexporter/config.go index 6a74f44189651..8115532c713a2 100644 --- a/exporter/opensearchexporter/config.go +++ b/exporter/opensearchexporter/config.go @@ -145,7 +145,7 @@ func (cfg *Config) Validate() error { return errBulkActionInvalid } - if _, ok := mappingModes[cfg.MappingsSettings.Mode]; !ok { + if _, ok := mappingModes[cfg.Mode]; !ok { multiErr = append(multiErr, errMappingModeInvalid) } diff --git a/exporter/opensearchexporter/sso_log_exporter.go b/exporter/opensearchexporter/sso_log_exporter.go index fe3584f3e0f06..635786df3a21e 100644 --- a/exporter/opensearchexporter/sso_log_exporter.go +++ b/exporter/opensearchexporter/sso_log_exporter.go @@ -27,10 +27,10 @@ func newLogExporter(cfg *Config, set exporter.Settings) *logExporter { model := &encodeModel{ dedup: cfg.Dedup, dedot: cfg.Dedot, - sso: cfg.MappingsSettings.Mode == MappingSS4O.String(), - flattenAttributes: cfg.MappingsSettings.Mode == MappingFlattenAttributes.String(), - timestampField: cfg.MappingsSettings.TimestampField, - unixTime: cfg.MappingsSettings.UnixTimestamp, + sso: cfg.Mode == MappingSS4O.String(), + flattenAttributes: cfg.Mode == MappingFlattenAttributes.String(), + timestampField: cfg.TimestampField, + unixTime: cfg.UnixTimestamp, dataset: cfg.Dataset, namespace: cfg.Namespace, } diff --git a/exporter/otelarrowexporter/config_test.go b/exporter/otelarrowexporter/config_test.go index 1502b4f9110f9..cbbcc3ab48b46 100644 --- a/exporter/otelarrowexporter/config_test.go +++ b/exporter/otelarrowexporter/config_test.go @@ -34,7 +34,7 @@ func TestUnmarshalDefaultConfig(t *testing.T) { cfg := factory.CreateDefaultConfig() assert.NoError(t, cm.Unmarshal(cfg)) assert.Equal(t, factory.CreateDefaultConfig(), cfg) - assert.Equal(t, "round_robin", cfg.(*Config).ClientConfig.BalancerName) + assert.Equal(t, "round_robin", cfg.(*Config).BalancerName) assert.Equal(t, arrow.DefaultPrioritizer, cfg.(*Config).Arrow.Prioritizer) } diff --git a/exporter/otelarrowexporter/factory_test.go b/exporter/otelarrowexporter/factory_test.go index e3d9dbe1d5b56..b0173e482bbef 100644 --- a/exporter/otelarrowexporter/factory_test.go +++ b/exporter/otelarrowexporter/factory_test.go @@ -50,7 +50,7 @@ func TestCreateDefaultConfig(t *testing.T) { func TestCreateMetrics(t *testing.T) { factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = testutil.GetAvailableLocalAddress(t) + cfg.Endpoint = testutil.GetAvailableLocalAddress(t) set := exportertest.NewNopSettings(metadata.Type) oexp, err := factory.CreateMetrics(context.Background(), set, cfg) @@ -216,7 +216,7 @@ func TestCreateTraces(t *testing.T) { func TestCreateLogs(t *testing.T) { factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = testutil.GetAvailableLocalAddress(t) + cfg.Endpoint = testutil.GetAvailableLocalAddress(t) set := exportertest.NewNopSettings(metadata.Type) oexp, err := factory.CreateLogs(context.Background(), set, cfg) @@ -227,7 +227,7 @@ func TestCreateLogs(t *testing.T) { func TestCreateArrowTracesExporter(t *testing.T) { factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = testutil.GetAvailableLocalAddress(t) + cfg.Endpoint = testutil.GetAvailableLocalAddress(t) cfg.Arrow = ArrowConfig{ NumStreams: 1, } diff --git a/exporter/otelarrowexporter/otelarrow.go b/exporter/otelarrowexporter/otelarrow.go index e269dd71dc75b..a8a7af76ae4b0 100644 --- a/exporter/otelarrowexporter/otelarrow.go +++ b/exporter/otelarrowexporter/otelarrow.go @@ -117,20 +117,20 @@ func (e *baseExporter) start(ctx context.Context, host component.Host) (err erro dialOpts = append(dialOpts, configgrpc.WithGrpcDialOption(opt)) } - if e.clientConn, err = e.config.ClientConfig.ToClientConn(ctx, host, e.settings.TelemetrySettings, dialOpts...); err != nil { + if e.clientConn, err = e.config.ToClientConn(ctx, host, e.settings.TelemetrySettings, dialOpts...); err != nil { return err } e.traceExporter = ptraceotlp.NewGRPCClient(e.clientConn) e.metricExporter = pmetricotlp.NewGRPCClient(e.clientConn) e.logExporter = plogotlp.NewGRPCClient(e.clientConn) headers := map[string]string{} - for k, v := range e.config.ClientConfig.Headers { + for k, v := range e.config.Headers { headers[k] = string(v) } headerMetadata := metadata.New(headers) e.metadata = metadata.Join(e.metadata, headerMetadata) e.callOptions = []grpc.CallOption{ - grpc.WaitForReady(e.config.ClientConfig.WaitForReady), + grpc.WaitForReady(e.config.WaitForReady), } if !e.config.Arrow.Disabled { @@ -138,9 +138,9 @@ func (e *baseExporter) start(ctx context.Context, host component.Host) (err erro ctx := e.enhanceContext(context.Background()) var perRPCCreds credentials.PerRPCCredentials - if e.config.ClientConfig.Auth != nil { + if e.config.Auth != nil { // Get the auth extension, we'll use it to enrich the request context. - authClient, err := e.config.ClientConfig.Auth.GetGRPCClientAuthenticator(ctx, host.GetExtensions()) + authClient, err := e.config.Auth.GetGRPCClientAuthenticator(ctx, host.GetExtensions()) if err != nil { return err } @@ -155,7 +155,7 @@ func (e *baseExporter) start(ctx context.Context, host component.Host) (err erro arrowCallOpts := e.callOptions - if e.config.ClientConfig.Compression == configcompression.TypeZstd { + if e.config.Compression == configcompression.TypeZstd { // ignore the error below b/c Validate() was called _ = zstd.SetEncoderConfig(e.config.Arrow.Zstd) // use the configured compressor. @@ -215,7 +215,7 @@ func (e *baseExporter) pushTraces(ctx context.Context, td ptrace.Traces) error { return err } partialSuccess := resp.PartialSuccess() - if !(partialSuccess.ErrorMessage() == "" && partialSuccess.RejectedSpans() == 0) { + if partialSuccess.ErrorMessage() != "" || partialSuccess.RejectedSpans() != 0 { // TODO: These should be counted, similar to dropped items. e.settings.Logger.Warn("partial success", zap.String("message", resp.PartialSuccess().ErrorMessage()), @@ -237,7 +237,7 @@ func (e *baseExporter) pushMetrics(ctx context.Context, md pmetric.Metrics) erro return err } partialSuccess := resp.PartialSuccess() - if !(partialSuccess.ErrorMessage() == "" && partialSuccess.RejectedDataPoints() == 0) { + if partialSuccess.ErrorMessage() != "" || partialSuccess.RejectedDataPoints() != 0 { // TODO: These should be counted, similar to dropped items. e.settings.Logger.Warn("partial success", zap.String("message", resp.PartialSuccess().ErrorMessage()), @@ -259,7 +259,7 @@ func (e *baseExporter) pushLogs(ctx context.Context, ld plog.Logs) error { return err } partialSuccess := resp.PartialSuccess() - if !(partialSuccess.ErrorMessage() == "" && partialSuccess.RejectedLogRecords() == 0) { + if partialSuccess.ErrorMessage() != "" || partialSuccess.RejectedLogRecords() != 0 { // TODO: These should be counted, similar to dropped items. e.settings.Logger.Warn("partial success", zap.String("message", resp.PartialSuccess().ErrorMessage()), diff --git a/exporter/otelarrowexporter/otelarrow_test.go b/exporter/otelarrowexporter/otelarrow_test.go index dfe17ec4b3d5e..73b13ad0ada6c 100644 --- a/exporter/otelarrowexporter/otelarrow_test.go +++ b/exporter/otelarrowexporter/otelarrow_test.go @@ -463,10 +463,10 @@ func TestSendTracesWhenEndpointHasHttpScheme(t *testing.T) { factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) cfg.ClientConfig = test.gRPCClientSettings - cfg.ClientConfig.Endpoint = test.scheme + ln.Addr().String() + cfg.Endpoint = test.scheme + ln.Addr().String() cfg.Arrow.MaxStreamLifetime = 100 * time.Second if test.useTLS { - cfg.ClientConfig.TLSSetting.InsecureSkipVerify = true + cfg.TLSSetting.InsecureSkipVerify = true } set := exportertest.NewNopSettings(factory.Type()) exp, err := factory.CreateTraces(context.Background(), set, cfg) @@ -934,7 +934,7 @@ func testSendArrowTraces(t *testing.T, clientWaitForReady, streamServiceAvailabl cfg.QueueSettings.Enabled = false set := exportertest.NewNopSettings(factory.Type()) - set.TelemetrySettings.Logger = zaptest.NewLogger(t) + set.Logger = zaptest.NewLogger(t) exp, err := factory.CreateTraces(context.Background(), set, cfg) require.NoError(t, err) require.NotNil(t, exp) @@ -1103,7 +1103,7 @@ func TestSendArrowFailedTraces(t *testing.T) { cfg.QueueSettings.Enabled = false set := exportertest.NewNopSettings(factory.Type()) - set.TelemetrySettings.Logger = zaptest.NewLogger(t) + set.Logger = zaptest.NewLogger(t) exp, err := factory.CreateTraces(context.Background(), set, cfg) require.NoError(t, err) require.NotNil(t, exp) @@ -1167,7 +1167,7 @@ func TestUserDialOptions(t *testing.T) { } set := exportertest.NewNopSettings(factory.Type()) - set.TelemetrySettings.Logger = zaptest.NewLogger(t) + set.Logger = zaptest.NewLogger(t) exp, err := factory.CreateTraces(context.Background(), set, cfg) require.NoError(t, err) require.NotNil(t, exp) diff --git a/exporter/prometheusexporter/collector_test.go b/exporter/prometheusexporter/collector_test.go index d37a424c4712a..016ae7cd26a08 100644 --- a/exporter/prometheusexporter/collector_test.go +++ b/exporter/prometheusexporter/collector_test.go @@ -10,7 +10,6 @@ import ( "time" "github.com/prometheus/client_golang/prometheus" - dto "github.com/prometheus/client_model/go" io_prometheus_client "github.com/prometheus/client_model/go" "github.com/stretchr/testify/require" "go.opentelemetry.io/collector/pdata/pcommon" @@ -99,7 +98,7 @@ func TestConvertMetric(t *testing.T) { "testgauge": { mf: &io_prometheus_client.MetricFamily{ Name: proto.String("testgauge"), - Type: dto.MetricType_COUNTER.Enum(), + Type: io_prometheus_client.MetricType_COUNTER.Enum(), }, }, }, @@ -113,7 +112,7 @@ func TestConvertMetric(t *testing.T) { "testgauge": { mf: &io_prometheus_client.MetricFamily{ Name: proto.String("testgauge"), - Type: dto.MetricType_GAUGE.Enum(), + Type: io_prometheus_client.MetricType_GAUGE.Enum(), Help: proto.String("test help value"), }, }, diff --git a/exporter/sematextexporter/config.go b/exporter/sematextexporter/config.go index 008133cbc0edf..326e0ec7e1928 100644 --- a/exporter/sematextexporter/config.go +++ b/exporter/sematextexporter/config.go @@ -51,8 +51,8 @@ func (cfg *Config) Validate() error { if strings.ToLower(cfg.Region) != euRegion && strings.ToLower(cfg.Region) != usRegion && strings.ToLower(cfg.Region) != "" { return fmt.Errorf("invalid region: %s. please use either 'EU' or 'US'", cfg.Region) } - if !isValidUUID(cfg.MetricsConfig.AppToken) && cfg.MetricsConfig.AppToken != "" { - return fmt.Errorf("invalid metrics app_token: %s. app_token is not a valid UUID", cfg.MetricsConfig.AppToken) + if !isValidUUID(cfg.AppToken) && cfg.AppToken != "" { + return fmt.Errorf("invalid metrics app_token: %s. app_token is not a valid UUID", cfg.AppToken) } if strings.ToLower(cfg.Region) == euRegion { diff --git a/exporter/sematextexporter/logger.go b/exporter/sematextexporter/logger.go index a471220053143..584e703682da2 100644 --- a/exporter/sematextexporter/logger.go +++ b/exporter/sematextexporter/logger.go @@ -21,5 +21,5 @@ func newZapSematextLogger(logger *zap.Logger) common.Logger { } func (l zapSematextLogger) Debug(msg string, kv ...any) { - l.SugaredLogger.Debugw(msg, kv...) + l.Debugw(msg, kv...) } diff --git a/exporter/sematextexporter/writer.go b/exporter/sematextexporter/writer.go index 8bb63940c8e10..fbdc4ccbbb780 100644 --- a/exporter/sematextexporter/writer.go +++ b/exporter/sematextexporter/writer.go @@ -64,7 +64,7 @@ func newSematextHTTPWriter(logger common.Logger, config *Config, telemetrySettin payloadMaxBytes: config.PayloadMaxBytes, logger: logger, hostname: hostname, - token: config.MetricsConfig.AppToken, + token: config.AppToken, }, nil } diff --git a/exporter/sematextexporter/writer_test.go b/exporter/sematextexporter/writer_test.go index bcecf083f34a5..a7ed2018393d6 100644 --- a/exporter/sematextexporter/writer_test.go +++ b/exporter/sematextexporter/writer_test.go @@ -171,7 +171,7 @@ func TestSematextHTTPWriterBatchMaxPayload(t *testing.T) { token: "test-token", }, } - defer batch.sematextHTTPWriter.httpClient.CloseIdleConnections() + defer batch.httpClient.CloseIdleConnections() err := batch.EnqueuePoint(context.Background(), "m", map[string]string{"k": "v"}, map[string]any{"f": int64(1)}, time.Unix(1, 0), 0) require.NoError(t, err) diff --git a/exporter/signalfxexporter/config.go b/exporter/signalfxexporter/config.go index fc7254b52b81b..5b1f568ea9d68 100644 --- a/exporter/signalfxexporter/config.go +++ b/exporter/signalfxexporter/config.go @@ -220,7 +220,7 @@ func (cfg *Config) Validate() error { ` "ingest_url" and "api_url" should be explicitly set`) } - if cfg.ClientConfig.Timeout < 0 { + if cfg.Timeout < 0 { return errors.New(`cannot have a negative "timeout"`) } diff --git a/exporter/signalfxexporter/exporter.go b/exporter/signalfxexporter/exporter.go index 9d3c3c39c015f..92558f1d3ffb0 100644 --- a/exporter/signalfxexporter/exporter.go +++ b/exporter/signalfxexporter/exporter.go @@ -75,7 +75,7 @@ func newSignalFxExporter( } converter, err := translation.NewMetricsConverter( - createSettings.TelemetrySettings.Logger, + createSettings.Logger, metricTranslator, config.ExcludeMetrics, config.IncludeMetrics, @@ -212,7 +212,7 @@ func (se *signalfxExporter) startLogs(ctx context.Context, host component.Host) } func (se *signalfxExporter) createClient(ctx context.Context, host component.Host) (*http.Client, error) { - se.config.ClientConfig.TLSSetting = se.config.IngestTLSSettings + se.config.TLSSetting = se.config.IngestTLSSettings return se.config.ToClient(ctx, host, se.telemetrySettings) } @@ -262,11 +262,11 @@ func buildHeaders(config *Config, version string) map[string]string { // Add any custom headers from the config. They will override the pre-defined // ones above in case of conflict, but, not the content encoding one since // the latter one is defined according to the payload. - for k, v := range config.ClientConfig.Headers { + for k, v := range config.Headers { headers[k] = string(v) } // we want to control how headers are set, overriding user headers with our passthrough. - config.ClientConfig.Headers = nil + config.Headers = nil return headers } diff --git a/exporter/signalfxexporter/exporter_test.go b/exporter/signalfxexporter/exporter_test.go index 22d3fe4d9eddc..69afcac800be7 100644 --- a/exporter/signalfxexporter/exporter_test.go +++ b/exporter/signalfxexporter/exporter_test.go @@ -546,11 +546,11 @@ func TestConsumeMetricsWithAccessTokenPassthrough(t *testing.T) { cfg := factory.CreateDefaultConfig().(*Config) cfg.IngestURL = server.URL cfg.APIURL = server.URL - cfg.ClientConfig.Headers = make(map[string]configopaque.String) + cfg.Headers = make(map[string]configopaque.String) for k, v := range tt.additionalHeaders { - cfg.ClientConfig.Headers[k] = configopaque.String(v) + cfg.Headers[k] = configopaque.String(v) } - cfg.ClientConfig.Headers["test_header_"] = configopaque.String(tt.name) + cfg.Headers["test_header_"] = configopaque.String(tt.name) cfg.AccessToken = configopaque.String(fromHeaders) cfg.AccessTokenPassthrough = tt.accessTokenPassthrough cfg.SendOTLPHistograms = tt.sendOTLPHistograms @@ -669,11 +669,11 @@ func TestConsumeMetricsAccessTokenPassthroughPriorityToContext(t *testing.T) { cfg := factory.CreateDefaultConfig().(*Config) cfg.IngestURL = server.URL cfg.APIURL = server.URL - cfg.ClientConfig.Headers = make(map[string]configopaque.String) + cfg.Headers = make(map[string]configopaque.String) for k, v := range tt.additionalHeaders { - cfg.ClientConfig.Headers[k] = configopaque.String(v) + cfg.Headers[k] = configopaque.String(v) } - cfg.ClientConfig.Headers["test_header_"] = configopaque.String(tt.name) + cfg.Headers["test_header_"] = configopaque.String(tt.name) cfg.AccessToken = configopaque.String(fromHeaders) cfg.AccessTokenPassthrough = tt.accessTokenPassthrough cfg.SendOTLPHistograms = tt.sendOTLPHistograms diff --git a/exporter/signalfxexporter/factory.go b/exporter/signalfxexporter/factory.go index fdd21092f9d0d..a3c7e705632d1 100644 --- a/exporter/signalfxexporter/factory.go +++ b/exporter/signalfxexporter/factory.go @@ -88,17 +88,17 @@ func createTracesExporter( cfg := eCfg.(*Config) corrCfg := cfg.Correlation - if corrCfg.ClientConfig.Endpoint == "" { + if corrCfg.Endpoint == "" { apiURL, err := cfg.getAPIURL() if err != nil { return nil, fmt.Errorf("unable to create API URL: %w", err) } - corrCfg.ClientConfig.Endpoint = apiURL.String() + corrCfg.Endpoint = apiURL.String() } if cfg.AccessToken == "" { return nil, errors.New("access_token is required") } - set.Logger.Info("Correlation tracking enabled", zap.String("endpoint", corrCfg.ClientConfig.Endpoint)) + set.Logger.Info("Correlation tracking enabled", zap.String("endpoint", corrCfg.Endpoint)) tracker := correlation.NewTracker(corrCfg, cfg.AccessToken, set) return exporterhelper.NewTraces( diff --git a/exporter/signalfxexporter/internal/apm/tracetracker/cache.go b/exporter/signalfxexporter/internal/apm/tracetracker/cache.go index 432f82bab4311..0d27ea9f91cdd 100644 --- a/exporter/signalfxexporter/internal/apm/tracetracker/cache.go +++ b/exporter/signalfxexporter/internal/apm/tracetracker/cache.go @@ -110,11 +110,7 @@ func (t *TimeoutCache) GetPurgeable(now time.Time) []*CacheKey { var candidates []*CacheKey elm := t.keysByTime.Back() - for { - if elm == nil { - break - } - + for elm != nil { e := elm.Value.(*cacheElem) // If this one isn't timed out, nothing else in the list is either. if now.Sub(e.LastSeen) < t.timeout { diff --git a/exporter/signalfxexporter/internal/correlation/config.go b/exporter/signalfxexporter/internal/correlation/config.go index 0c69af3749a06..a8e3f2987b93e 100644 --- a/exporter/signalfxexporter/internal/correlation/config.go +++ b/exporter/signalfxexporter/internal/correlation/config.go @@ -49,11 +49,11 @@ type Config struct { } func (c *Config) validate() error { - if c.ClientConfig.Endpoint == "" { + if c.Endpoint == "" { return errors.New("`correlation.endpoint` not specified") } - _, err := url.Parse(c.ClientConfig.Endpoint) + _, err := url.Parse(c.Endpoint) if err != nil { return err } diff --git a/exporter/signalfxexporter/internal/correlation/config_test.go b/exporter/signalfxexporter/internal/correlation/config_test.go index 78e4ec71846f8..edd184174bb4b 100644 --- a/exporter/signalfxexporter/internal/correlation/config_test.go +++ b/exporter/signalfxexporter/internal/correlation/config_test.go @@ -12,7 +12,7 @@ import ( func TestValidConfig(t *testing.T) { config := DefaultConfig() - config.ClientConfig.Endpoint = "https://localhost" + config.Endpoint = "https://localhost" require.NoError(t, config.validate()) } diff --git a/exporter/signalfxexporter/internal/correlation/correlation.go b/exporter/signalfxexporter/internal/correlation/correlation.go index 590d1735b9012..b7678cc4e7871 100644 --- a/exporter/signalfxexporter/internal/correlation/correlation.go +++ b/exporter/signalfxexporter/internal/correlation/correlation.go @@ -51,9 +51,9 @@ func NewTracker(cfg *Config, accessToken configopaque.String, params exporter.Se func newCorrelationClient(ctx context.Context, cfg *Config, accessToken configopaque.String, params exporter.Settings, host component.Host) ( *correlationContext, error, ) { - corrURL, err := url.Parse(cfg.ClientConfig.Endpoint) + corrURL, err := url.Parse(cfg.Endpoint) if err != nil { - return nil, fmt.Errorf("failed to parse correlation endpoint URL %q: %w", cfg.ClientConfig.Endpoint, err) + return nil, fmt.Errorf("failed to parse correlation endpoint URL %q: %w", cfg.Endpoint, err) } httpClient, err := cfg.ToClient(ctx, host, params.TelemetrySettings) @@ -135,7 +135,7 @@ func (cor *Tracker) Shutdown(_ context.Context) error { if cor != nil { if cor.correlation != nil { cor.correlation.cancel() - cor.correlation.CorrelationClient.Shutdown() + cor.correlation.Shutdown() } if cor.pTicker != nil { diff --git a/exporter/signalfxexporter/internal/translation/translator.go b/exporter/signalfxexporter/internal/translation/translator.go index 5dd4b7dccb763..42c05a1e161a4 100644 --- a/exporter/signalfxexporter/internal/translation/translator.go +++ b/exporter/signalfxexporter/internal/translation/translator.go @@ -554,9 +554,10 @@ func (mp *MetricTranslator) Shutdown() { func calcNewMetricInputPairs(processedDataPoints []*sfxpb.DataPoint, tr Rule) [][2]*sfxpb.DataPoint { var operand1Pts, operand2Pts []*sfxpb.DataPoint for _, dp := range processedDataPoints { - if dp.Metric == tr.Operand1Metric { + switch dp.Metric { + case tr.Operand1Metric: operand1Pts = append(operand1Pts, dp) - } else if dp.Metric == tr.Operand2Metric { + case tr.Operand2Metric: operand2Pts = append(operand2Pts, dp) } } diff --git a/exporter/splunkhecexporter/client.go b/exporter/splunkhecexporter/client.go index 70e2e11bb6598..0e0f88932cb50 100644 --- a/exporter/splunkhecexporter/client.go +++ b/exporter/splunkhecexporter/client.go @@ -658,7 +658,7 @@ func checkHecHealth(ctx context.Context, client *http.Client, healthCheckURL *ur func buildHTTPClient(ctx context.Context, config *Config, host component.Host, telemetrySettings component.TelemetrySettings) (*http.Client, error) { // we handle compression explicitly. - config.ClientConfig.Compression = "" + config.Compression = "" return config.ToClient(ctx, host, telemetrySettings) } diff --git a/exporter/splunkhecexporter/client_test.go b/exporter/splunkhecexporter/client_test.go index 3716c140dcdad..c5b2056d104bb 100644 --- a/exporter/splunkhecexporter/client_test.go +++ b/exporter/splunkhecexporter/client_test.go @@ -218,7 +218,7 @@ func runMetricsExport(t *testing.T, cfg *Config, metrics pmetric.Metrics, expect } factory := NewFactory() - cfg.ClientConfig.Endpoint = "http://" + listener.Addr().String() + "/services/collector" + cfg.Endpoint = "http://" + listener.Addr().String() + "/services/collector" cfg.Token = "1234-1234" cfg.UseMultiMetricFormat = useMultiMetricsFormat @@ -270,7 +270,7 @@ func runTraceExport(t *testing.T, testConfig *Config, traces ptrace.Traces, expe factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://" + listener.Addr().String() + "/services/collector" + cfg.Endpoint = "http://" + listener.Addr().String() + "/services/collector" cfg.DisableCompression = testConfig.DisableCompression cfg.MaxContentLengthTraces = testConfig.MaxContentLengthTraces cfg.Token = "1234-1234" @@ -332,7 +332,7 @@ func runLogExport(t *testing.T, cfg *Config, ld plog.Logs, expectedBatchesNum in panic(err) } - cfg.ClientConfig.Endpoint = "http://" + listener.Addr().String() + "/services/collector" + cfg.Endpoint = "http://" + listener.Addr().String() + "/services/collector" cfg.Token = "1234-1234" rr := make(chan receivedRequest) @@ -1292,12 +1292,12 @@ func TestErrorReceived(t *testing.T) { factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://" + listener.Addr().String() + "/services/collector" + cfg.Endpoint = "http://" + listener.Addr().String() + "/services/collector" // Disable QueueSettings to ensure that we execute the request when calling ConsumeTraces // otherwise we will not see the error. cfg.QueueSettings.Enabled = false // Disable retries to not wait too much time for the return error. - cfg.BackOffConfig.Enabled = false + cfg.Enabled = false cfg.DisableCompression = true cfg.Token = "1234-1234" @@ -1350,8 +1350,8 @@ func TestInvalidURL(t *testing.T) { // otherwise we will not see the error. cfg.QueueSettings.Enabled = false // Disable retries to not wait too much time for the return error. - cfg.BackOffConfig.Enabled = false - cfg.ClientConfig.Endpoint = "ftp://example.com:134" + cfg.Enabled = false + cfg.Endpoint = "ftp://example.com:134" cfg.Token = "1234-1234" params := exportertest.NewNopSettings(metadata.Type) exporter, err := factory.CreateTraces(context.Background(), params, cfg) @@ -1384,12 +1384,12 @@ func TestHeartbeatStartupFailed(t *testing.T) { }() factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://" + listener.Addr().String() + "/services/collector" + cfg.Endpoint = "http://" + listener.Addr().String() + "/services/collector" // Disable QueueSettings to ensure that we execute the request when calling ConsumeTraces // otherwise we will not see the error. cfg.QueueSettings.Enabled = false // Disable retries to not wait too much time for the return error. - cfg.BackOffConfig.Enabled = false + cfg.Enabled = false cfg.DisableCompression = true cfg.Token = "1234-1234" cfg.Heartbeat.Startup = true @@ -1425,12 +1425,12 @@ func TestHeartbeatStartupPass_Disabled(t *testing.T) { }() factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://" + listener.Addr().String() + "/services/collector" + cfg.Endpoint = "http://" + listener.Addr().String() + "/services/collector" // Disable QueueSettings to ensure that we execute the request when calling ConsumeTraces // otherwise we will not see the error. cfg.QueueSettings.Enabled = false // Disable retries to not wait too much time for the return error. - cfg.BackOffConfig.Enabled = false + cfg.Enabled = false cfg.DisableCompression = true cfg.Token = "1234-1234" cfg.Heartbeat.Startup = false @@ -1460,12 +1460,12 @@ func TestHeartbeatStartupPass(t *testing.T) { }() factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://" + listener.Addr().String() + "/services/collector" + cfg.Endpoint = "http://" + listener.Addr().String() + "/services/collector" // Disable QueueSettings to ensure that we execute the request when calling ConsumeTraces // otherwise we will not see the error. cfg.QueueSettings.Enabled = false // Disable retries to not wait too much time for the return error. - cfg.BackOffConfig.Enabled = false + cfg.Enabled = false cfg.DisableCompression = true cfg.Token = "1234-1234" cfg.Heartbeat.Startup = true diff --git a/exporter/splunkhecexporter/config.go b/exporter/splunkhecexporter/config.go index cd08cb600becc..0ffc15a46697e 100644 --- a/exporter/splunkhecexporter/config.go +++ b/exporter/splunkhecexporter/config.go @@ -145,7 +145,7 @@ type Config struct { } func (cfg *Config) getURL() (out *url.URL, err error) { - out, err = url.Parse(cfg.ClientConfig.Endpoint) + out, err = url.Parse(cfg.Endpoint) if err != nil { return out, err } @@ -161,7 +161,7 @@ func (cfg *Config) Validate() error { if !cfg.LogDataEnabled && !cfg.ProfilingDataEnabled { return errors.New(`either "log_data_enabled" or "profiling_data_enabled" has to be true`) } - if cfg.ClientConfig.Endpoint == "" { + if cfg.Endpoint == "" { return errors.New(`requires a non-empty "endpoint"`) } _, err := cfg.getURL() diff --git a/exporter/splunkhecexporter/config_test.go b/exporter/splunkhecexporter/config_test.go index 86a0336f24b1f..a0eb85d83d2c7 100644 --- a/exporter/splunkhecexporter/config_test.go +++ b/exporter/splunkhecexporter/config_test.go @@ -32,7 +32,7 @@ func TestLoadConfig(t *testing.T) { // Endpoint and Token do not have a default value so set them directly. defaultCfg := createDefaultConfig().(*Config) defaultCfg.Token = "00000000-0000-0000-0000-0000000000000" - defaultCfg.ClientConfig.Endpoint = "https://splunk:8088/services/collector" + defaultCfg.Endpoint = "https://splunk:8088/services/collector" hundred := 100 idleConnTimeout := 10 * time.Second @@ -167,7 +167,7 @@ func TestConfig_Validate(t *testing.T) { name: "bad url", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "cache_object:foo/bar" + cfg.Endpoint = "cache_object:foo/bar" cfg.Token = "foo" return cfg }(), @@ -177,7 +177,7 @@ func TestConfig_Validate(t *testing.T) { name: "missing token", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://example.com" + cfg.Endpoint = "http://example.com" return cfg }(), wantErr: "requires a non-empty \"token\"", @@ -186,7 +186,7 @@ func TestConfig_Validate(t *testing.T) { name: "max default content-length for logs", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://foo_bar.com" + cfg.Endpoint = "http://foo_bar.com" cfg.MaxContentLengthLogs = maxContentLengthLogsLimit + 1 cfg.Token = "foo" return cfg @@ -197,7 +197,7 @@ func TestConfig_Validate(t *testing.T) { name: "max default content-length for metrics", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://foo_bar.com" + cfg.Endpoint = "http://foo_bar.com" cfg.MaxContentLengthMetrics = maxContentLengthMetricsLimit + 1 cfg.Token = "foo" return cfg @@ -208,7 +208,7 @@ func TestConfig_Validate(t *testing.T) { name: "max default content-length for traces", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://foo_bar.com" + cfg.Endpoint = "http://foo_bar.com" cfg.MaxContentLengthTraces = maxContentLengthTracesLimit + 1 cfg.Token = "foo" return cfg @@ -219,7 +219,7 @@ func TestConfig_Validate(t *testing.T) { name: "max default event-size", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://foo_bar.com" + cfg.Endpoint = "http://foo_bar.com" cfg.MaxEventSize = maxMaxEventSize + 1 cfg.Token = "foo" return cfg @@ -230,7 +230,7 @@ func TestConfig_Validate(t *testing.T) { name: "negative queue size", cfg: func() *Config { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "http://foo_bar.com" + cfg.Endpoint = "http://foo_bar.com" cfg.QueueSettings.Enabled = true cfg.QueueSettings.QueueSize = -5 cfg.Token = "foo" diff --git a/exporter/splunkhecexporter/factory_test.go b/exporter/splunkhecexporter/factory_test.go index 30171f79cc823..c097aab9b479d 100644 --- a/exporter/splunkhecexporter/factory_test.go +++ b/exporter/splunkhecexporter/factory_test.go @@ -25,7 +25,7 @@ func TestCreateDefaultConfig(t *testing.T) { func TestCreateMetrics(t *testing.T) { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "https://example.com:8088/services/collector" + cfg.Endpoint = "https://example.com:8088/services/collector" cfg.Token = "1234-1234" params := exportertest.NewNopSettings(metadata.Type) @@ -35,7 +35,7 @@ func TestCreateMetrics(t *testing.T) { func TestCreateTraces(t *testing.T) { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "https://example.com:8088/services/collector" + cfg.Endpoint = "https://example.com:8088/services/collector" cfg.Token = "1234-1234" params := exportertest.NewNopSettings(metadata.Type) @@ -45,7 +45,7 @@ func TestCreateTraces(t *testing.T) { func TestCreateLogs(t *testing.T) { cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "https://example.com:8088/services/collector" + cfg.Endpoint = "https://example.com:8088/services/collector" cfg.Token = "1234-1234" params := exportertest.NewNopSettings(metadata.Type) @@ -57,7 +57,7 @@ func TestCreateInstanceViaFactory(t *testing.T) { factory := NewFactory() cfg := factory.CreateDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = "https://example.com:8088/services/collector" + cfg.Endpoint = "https://example.com:8088/services/collector" cfg.Token = "1234-1234" params := exportertest.NewNopSettings(metadata.Type) exp, err := factory.CreateMetrics( @@ -68,7 +68,7 @@ func TestCreateInstanceViaFactory(t *testing.T) { // Set values that don't have a valid default. cfg.Token = "testToken" - cfg.ClientConfig.Endpoint = "https://example.com" + cfg.Endpoint = "https://example.com" exp, err = factory.CreateMetrics( context.Background(), params, cfg) diff --git a/exporter/splunkhecexporter/integration_test.go b/exporter/splunkhecexporter/integration_test.go index 02ee065155745..2382f3686efd7 100644 --- a/exporter/splunkhecexporter/integration_test.go +++ b/exporter/splunkhecexporter/integration_test.go @@ -364,7 +364,7 @@ func TestSplunkHecExporter(t *testing.T) { // Endpoint and Token do not have a default value so set them directly. config := NewFactory().CreateDefaultConfig().(*Config) config.Token = configopaque.String(integrationtestutils.GetConfigVariable("HEC_TOKEN")) - config.ClientConfig.Endpoint = "https://" + integrationtestutils.GetConfigVariable("HOST") + ":" + integrationtestutils.GetConfigVariable("HEC_PORT") + "/services/collector" + config.Endpoint = "https://" + integrationtestutils.GetConfigVariable("HOST") + ":" + integrationtestutils.GetConfigVariable("HEC_PORT") + "/services/collector" config.Source = "otel" config.SourceType = "st-otel" diff --git a/exporter/stefexporter/exporter.go b/exporter/stefexporter/exporter.go index 8acbbd8fdbf57..5b182d7e67b5d 100644 --- a/exporter/stefexporter/exporter.go +++ b/exporter/stefexporter/exporter.go @@ -86,7 +86,7 @@ func newStefExporter(set component.TelemetrySettings, cfg *Config) *stefExporter func (s *stefExporter) Start(ctx context.Context, host component.Host) error { // Prepare gRPC connection. var err error - s.grpcConn, err = s.cfg.ClientConfig.ToClientConn(ctx, host, s.set) + s.grpcConn, err = s.cfg.ToClientConn(ctx, host, s.set) if err != nil { return err } diff --git a/exporter/stefexporter/exporter_test.go b/exporter/stefexporter/exporter_test.go index e5595db41247e..af14bed338bf3 100644 --- a/exporter/stefexporter/exporter_test.go +++ b/exporter/stefexporter/exporter_test.go @@ -141,7 +141,7 @@ func runTest( if cfg == nil { cfg = factory.CreateDefaultConfig().(*Config) } - cfg.ClientConfig.Endpoint = mockSrv.endpoint + cfg.Endpoint = mockSrv.endpoint // Use insecure mode for tests so that we don't bother with certificates. cfg.TLSSetting.Insecure = true @@ -149,7 +149,7 @@ func runTest( cfg.RetryConfig.InitialInterval = 10 * time.Millisecond set := exportertest.NewNopSettings(metadata.Type) - set.TelemetrySettings.Logger = logger + set.Logger = logger exp, err := factory.CreateMetrics(context.Background(), set, cfg) require.NoError(t, err) @@ -212,7 +212,7 @@ func TestReconnect(t *testing.T) { // Shorten max ack waiting time so that the attempt to send on a failed // connection times out quickly and attempt to send again is tried // until the broken connection is detected and reconnection happens. - cfg.TimeoutConfig.Timeout = 300 * time.Millisecond + cfg.Timeout = 300 * time.Millisecond runTest( t, @@ -261,7 +261,7 @@ func TestAckTimeout(t *testing.T) { // Shorten max ack waiting time so that tests run fast. // Increase this if the second eventually() below fails sporadically. - cfg.TimeoutConfig.Timeout = 300 * time.Millisecond + cfg.Timeout = 300 * time.Millisecond runTest( t, @@ -316,7 +316,7 @@ func TestStartServerAfterClient(t *testing.T) { } set := exportertest.NewNopSettings(metadata.Type) - set.TelemetrySettings.Logger = logger + set.Logger = logger exp := newStefExporter(set.TelemetrySettings, cfg) require.NotNil(t, exp) @@ -372,7 +372,7 @@ func TestCancelBlockedExport(t *testing.T) { } set := exportertest.NewNopSettings(exportertest.NopType) - set.TelemetrySettings.Logger = logger + set.Logger = logger exp := newStefExporter(set.TelemetrySettings, cfg) require.NotNil(t, exp) @@ -426,7 +426,7 @@ func TestCancelAfterExport(t *testing.T) { } set := exportertest.NewNopSettings(exportertest.NopType) - set.TelemetrySettings.Logger = logger + set.Logger = logger exp := newStefExporter(set.TelemetrySettings, cfg) require.NotNil(t, exp) diff --git a/exporter/sumologicexporter/config.go b/exporter/sumologicexporter/config.go index e86f447e1fe49..2f515f5b93a19 100644 --- a/exporter/sumologicexporter/config.go +++ b/exporter/sumologicexporter/config.go @@ -77,18 +77,18 @@ func (cfg *Config) Validate() error { return errors.New("support for compress_encoding configuration has been removed, in favor of compression") } - if cfg.ClientConfig.Timeout < 1 || cfg.ClientConfig.Timeout > maxTimeout { - return fmt.Errorf("timeout must be between 1 and 55 seconds, got %v", cfg.ClientConfig.Timeout) + if cfg.Timeout < 1 || cfg.Timeout > maxTimeout { + return fmt.Errorf("timeout must be between 1 and 55 seconds, got %v", cfg.Timeout) } - switch cfg.ClientConfig.Compression { + switch cfg.Compression { case configcompression.TypeGzip: case configcompression.TypeDeflate: case configcompression.TypeZstd: case NoCompression: default: - return fmt.Errorf("invalid compression encoding type: %v", cfg.ClientConfig.Compression) + return fmt.Errorf("invalid compression encoding type: %v", cfg.Compression) } switch cfg.LogFormat { @@ -110,13 +110,13 @@ func (cfg *Config) Validate() error { return fmt.Errorf("unexpected metric format: %s", cfg.MetricFormat) } - if len(cfg.ClientConfig.Endpoint) == 0 && cfg.ClientConfig.Auth == nil { + if len(cfg.Endpoint) == 0 && cfg.Auth == nil { return errors.New("no endpoint and no auth extension specified") } - if _, err := url.Parse(cfg.ClientConfig.Endpoint); err != nil { + if _, err := url.Parse(cfg.Endpoint); err != nil { return fmt.Errorf("failed parsing endpoint URL: %s; err: %w", - cfg.ClientConfig.Endpoint, err, + cfg.Endpoint, err, ) } diff --git a/exporter/sumologicexporter/exporter_test.go b/exporter/sumologicexporter/exporter_test.go index 6b2a0e3e39545..845913269537b 100644 --- a/exporter/sumologicexporter/exporter_test.go +++ b/exporter/sumologicexporter/exporter_test.go @@ -47,7 +47,7 @@ type exporterTest struct { func createTestConfig() *Config { config := createDefaultConfig().(*Config) - config.ClientConfig.Compression = NoCompression + config.Compression = NoCompression config.LogFormat = TextFormat config.MaxRequestBodySize = 20_971_520 config.MetricFormat = OTLPMetricFormat @@ -80,8 +80,8 @@ func prepareExporterTest(t *testing.T, cfg *Config, cb []func(w http.ResponseWri ) }) - cfg.ClientConfig.Endpoint = testServer.URL - cfg.ClientConfig.Auth = nil + cfg.Endpoint = testServer.URL + cfg.Auth = nil exp, err := initExporter(cfg, exportertest.NewNopSettings(metadata.Type)) require.NoError(t, err) @@ -498,8 +498,8 @@ func Benchmark_ExporterPushLogs(b *testing.B) { config := createDefaultConfig().(*Config) config.MetricFormat = PrometheusFormat config.LogFormat = TextFormat - config.ClientConfig.Auth = nil - config.ClientConfig.Compression = configcompression.TypeGzip + config.Auth = nil + config.Compression = configcompression.TypeGzip return config } @@ -508,7 +508,7 @@ func Benchmark_ExporterPushLogs(b *testing.B) { b.Cleanup(func() { testServer.Close() }) cfg := createConfig() - cfg.ClientConfig.Endpoint = testServer.URL + cfg.Endpoint = testServer.URL exp, err := initExporter(cfg, exportertest.NewNopSettings(metadata.Type)) require.NoError(b, err) diff --git a/exporter/sumologicexporter/sender_test.go b/exporter/sumologicexporter/sender_test.go index 6b2ce1820e19e..a0db3f7f306cc 100644 --- a/exporter/sumologicexporter/sender_test.go +++ b/exporter/sumologicexporter/sender_test.go @@ -60,20 +60,20 @@ func prepareSenderTest(t *testing.T, compression configcompression.Type, cb []fu t.Cleanup(func() { testServer.Close() }) cfg := createDefaultConfig().(*Config) - cfg.ClientConfig.Endpoint = testServer.URL + cfg.Endpoint = testServer.URL switch compression { case configcompression.TypeGzip: - cfg.ClientConfig.Compression = configcompression.TypeGzip + cfg.Compression = configcompression.TypeGzip case configcompression.TypeZstd: - cfg.ClientConfig.Compression = configcompression.TypeZstd + cfg.Compression = configcompression.TypeZstd case NoCompression: - cfg.ClientConfig.Compression = NoCompression + cfg.Compression = NoCompression case configcompression.TypeDeflate: - cfg.ClientConfig.Compression = configcompression.TypeDeflate + cfg.Compression = configcompression.TypeDeflate default: - cfg.ClientConfig.Compression = configcompression.TypeGzip + cfg.Compression = configcompression.TypeGzip } - cfg.ClientConfig.Auth = nil + cfg.Auth = nil httpSettings := cfg.ClientConfig host := componenttest.NewNopHost() client, err := httpSettings.ToClient(context.Background(), host, componenttest.NewNopTelemetrySettings()) diff --git a/exporter/zipkinexporter/config.go b/exporter/zipkinexporter/config.go index 27d8d936c229e..48038c17b6bd8 100644 --- a/exporter/zipkinexporter/config.go +++ b/exporter/zipkinexporter/config.go @@ -30,7 +30,7 @@ var _ component.Config = (*Config)(nil) // Validate checks if the exporter configuration is valid func (cfg *Config) Validate() error { - if cfg.ClientConfig.Endpoint == "" { + if cfg.Endpoint == "" { return errors.New("endpoint required") } return nil