You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: src/connections/destinations/catalog/amazon-eventbridge/index.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,9 +3,9 @@ rewrite: true
3
3
title: Amazon EventBridge Destination
4
4
id: 5d1994fb320116000112aa12
5
5
---
6
-
[Amazon EventBridge](https://aws.amazon.com/eventbridge/) is the easiest way to onboard your Segment data into the AWS ecosystem.
6
+
[Amazon EventBridge](https://aws.amazon.com/eventbridge/){:target="_blank”} is the easiest way to onboard your Segment data into the AWS ecosystem.
7
7
8
-
In addition to already supported destinations like Kinesis, S3, and Redshift, you can use EventBridge to selectively route streaming data into Amazon SQS, SNS, and any service supported by [AWS CloudWatch Events](https://docs.aws.amazon.com/AmazonCloudWatch/latest/events/WhatIsCloudWatchEvents.html).
8
+
In addition to already supported destinations like Kinesis, S3, and Redshift, you can use EventBridge to selectively route streaming data into Amazon SQS, SNS, and any service supported by [AWS CloudWatch Events](https://docs.aws.amazon.com/AmazonCloudWatch/latest/events/WhatIsCloudWatchEvents.html){:target="_blank”}.
9
9
10
10
11
11
## Getting Started
@@ -14,7 +14,7 @@ In addition to already supported destinations like Kinesis, S3, and Redshift, yo
14
14
15
15
1. Provide Segment your AWS Account ID and the region you'd like us to configure the Segment Partner Event Source in. Ensure you've provided the same region in Segment where you'd like to configure your Event Bus.
16
16
2. Once you send an event through with the destination enabled, we'll create a Partner Event Source in Amazon EventBridge, which you can activate in the AWS Console.
17
-
3. Use the [AWS Console](http://console.aws.amazon.com/events/) to configure rules and destinations for the events in your Segment Partner Event Source.
17
+
3. Use the [AWS Console](http://console.aws.amazon.com/events/){:target="_blank”} to configure rules and destinations for the events in your Segment Partner Event Source.
18
18
19
19
The Event Source will be denoted by your Segment Source ID, which you can find in your Source Settings page under API Keys.
Copy file name to clipboardExpand all lines: src/connections/destinations/catalog/amazon-kinesis-firehose/index.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,7 @@ rewrite: true
3
3
title: Amazon Kinesis Firehose Destination
4
4
id: 59022a2270a3e552b955caa9
5
5
---
6
-
[Amazon Kinesis Firehose](https://aws.amazon.com/kinesis/data-firehose/) provides way to load streaming data into AWS. It can capture, transform, and load streaming data into Amazon Kinesis Analytics, Amazon S3, Amazon Redshift, and Amazon Elasticsearch Service, enabling near real-time analytics with existing business intelligence tools and dashboards you're already using today. It's a fully managed service that automatically scales to match the throughput of your data and requires no ongoing administration. It can also batch, compress, and encrypt the data before loading it, minimizing the amount of storage used at the destination and increasing security.
6
+
[Amazon Kinesis Firehose](https://aws.amazon.com/kinesis/data-firehose/){:target="_blank”} provides way to load streaming data into AWS. It can capture, transform, and load streaming data into Amazon Kinesis Analytics, Amazon S3, Amazon Redshift, and Amazon Elasticsearch Service, enabling near real-time analytics with existing business intelligence tools and dashboards you're already using today. It's a fully managed service that automatically scales to match the throughput of your data and requires no ongoing administration. It can also batch, compress, and encrypt the data before loading it, minimizing the amount of storage used at the destination and increasing security.
7
7
8
8
## Getting Started
9
9
@@ -36,7 +36,7 @@ To get started:
36
36
37
37
38
38
3. Create an IAM role.
39
-
1. Follow [these instructions](http://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_create_for-user.html#roles-creatingrole-user-console) to create an IAM role to allow Segment permission to write to your Kinesis Firehose Stream.
39
+
1. Follow [these instructions](http://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_create_for-user.html#roles-creatingrole-user-console){:target="_blank”} to create an IAM role to allow Segment permission to write to your Kinesis Firehose Stream.
40
40
2. When prompted to enter an Account ID, enter `595280932656`.
41
41
3. Select the checkbox to enable **Require External ID**.
42
42
4. Enter your Secret ID as the **External ID**. This can be found in Segment by navigating to your Amazon Kinesis Firehose destination in Segment, going to the Settings tab, and clicking the Secret ID setting.
Copy file name to clipboardExpand all lines: src/connections/destinations/catalog/amazon-kinesis/index.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,7 @@ rewrite: true
3
3
title: Amazon Kinesis Destination
4
4
id: 57da359580412f644ff33fb9
5
5
---
6
-
[Amazon Kinesis](https://aws.amazon.com/kinesis/) enables you to build custom applications that process or analyze streaming data for specialized needs. Amazon Kinesis Streams can continuously capture and store terabytes of data per hour from hundreds of thousands of sources such as website clickstreams, financial transactions, social media feeds, IT logs, and location-tracking events.
6
+
[Amazon Kinesis](https://aws.amazon.com/kinesis/){:target="_blank”} enables you to build custom applications that process or analyze streaming data for specialized needs. Amazon Kinesis Streams can continuously capture and store terabytes of data per hour from hundreds of thousands of sources such as website clickstreams, financial transactions, social media feeds, IT logs, and location-tracking events.
Copy file name to clipboardExpand all lines: src/connections/destinations/catalog/amazon-lambda/index.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -186,11 +186,11 @@ To configure your Segment Lambda destination:
186
186
187
187
**What is the Log Type Setting?**
188
188
189
-
This setting controls the [Log Type](https://docs.aws.amazon.com/lambda/latest/dg/API_Invoke.html#API_Invoke_RequestSyntax) for your Lambda function using Cloud Watch. Select option `Tail` if you would like to see [detailed logs](https://docs.aws.amazon.com/lambda/latest/dg/monitoring-functions.html) in Cloud Watch.
189
+
This setting controls the [Log Type](https://docs.aws.amazon.com/lambda/latest/dg/API_Invoke.html#API_Invoke_RequestSyntax){:target="_blank”} for your Lambda function using Cloud Watch. Select option `Tail` if you would like to see [detailed logs](https://docs.aws.amazon.com/lambda/latest/dg/monitoring-functions.html){:target="_blank”} in Cloud Watch.
190
190
191
191
**My Lambda <> Segment connection is timing out, what do I do?**
192
192
193
-
Due to how the event delivery system, [Centrifuge](https://segment.com/blog/introducing-centrifuge/), works, your Lambda can't take more than 5 seconds to run per message. If you're consistently running into timeout issues, you should consult the [AWS Lambda docs](https://docs.aws.amazon.com/lambda/index.html#lang/en_us), as well as docs for your language of choice, for tips on optimizing performance.
193
+
Due to how the event delivery system, [Centrifuge](https://segment.com/blog/introducing-centrifuge/){:target="_blank”}, works, your Lambda can't take more than 5 seconds to run per message. If you're consistently running into timeout issues, you should consult the [AWS Lambda docs](https://docs.aws.amazon.com/lambda/index.html#lang/en_us){:target="_blank”}, as well as docs for your language of choice, for tips on optimizing performance.
194
194
195
195
**Handling Common Errors**
196
196
You can find delivery logs in Destination > [Event Delivery](/docs/connections/event-delivery/).
Copy file name to clipboardExpand all lines: src/connections/destinations/catalog/amazon-personalize/index.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -762,7 +762,7 @@ import of import init_personalize_api as api_helper
762
762
api_helper.init()
763
763
```
764
764
765
-
This `import` and function call uses some boilerplate code, packaged as a [Lambda Layer](https://docs.aws.amazon.com/lambda/latest/dg/configuration-layers.html) needed to configure the Personalize API with the AWS Python SDK. This is only necessary while Personalize is in Preview. Once Personalize is GA and the API is bundled with the Python SDK, as well as other language SDKs, this supporting Layer will no longer be needed.
765
+
This `import` and function call uses some boilerplate code, packaged as a [Lambda Layer](https://docs.aws.amazon.com/lambda/latest/dg/configuration-layers.html){:target="_blank”} needed to configure the Personalize API with the AWS Python SDK. This is only necessary while Personalize is in Preview. Once Personalize is GA and the API is bundled with the Python SDK, as well as other language SDKs, this supporting Layer will no longer be needed.
766
766
767
767
To install Segment's Layer:
768
768
@@ -829,7 +829,7 @@ You need to modify the IAM Role & Policy originally created with this Lambda to
829
829
830
830
**Wire-up Personalize Event Tracker**
831
831
832
-
Another dependency in the function is the ability to call the Personalize [PutEvents API](https://docs.aws.amazon.com/personalize/latest/dg/API_UBS_PutEvents.html) endpoint as shown in the following excerpt.
832
+
Another dependency in the function is the ability to call the Personalize [PutEvents API](https://docs.aws.amazon.com/personalize/latest/dg/API_UBS_PutEvents.html){:target="_blank”} endpoint as shown in the following excerpt.
833
833
834
834
```js
835
835
personalize_events.put_events(
@@ -1005,8 +1005,8 @@ There are two settings relevant for track calls:
1005
1005
1006
1006
**What is the Log Type Setting?**
1007
1007
1008
-
This setting controls the [Log Type](https://docs.aws.amazon.com/lambda/latest/dg/API_Invoke.html#API_Invoke_RequestSyntax) for your Lambda function using Cloud Watch. Select option `Tail` if you would like to see [detailed logs](https://docs.aws.amazon.com/lambda/latest/dg/monitoring-functions.html)in Cloud Watch.
1008
+
This setting controls the [Log Type](https://docs.aws.amazon.com/lambda/latest/dg/API_Invoke.html#API_Invoke_RequestSyntax){:target="_blank”} for your Lambda function using Cloud Watch. Select option `Tail` if you would like to see [detailed logs](https://docs.aws.amazon.com/lambda/latest/dg/monitoring-functions.html){:target="_blank”}in Cloud Watch.
1009
1009
1010
1010
**My Lambda <> Segment connection is timing out, what do I do?**
1011
1011
1012
-
Due to how Segment's event delivery system, [Centrifuge](https://segment.com/blog/introducing-centrifuge/), works, your Lambda can't take more than five seconds to run per message. If you're consistently running into timeout issues, you should consult the [AWS Lambda docs](https://docs.aws.amazon.com/lambda/index.html#lang/en_us), as well as docs for your language of choice, for tips on optimizing performance.
1012
+
Due to how Segment's event delivery system, [Centrifuge](https://segment.com/blog/introducing-centrifuge/){:target="_blank”}, works, your Lambda can't take more than five seconds to run per message. If you're consistently running into timeout issues, you should consult the [AWS Lambda docs](https://docs.aws.amazon.com/lambda/index.html#lang/en_us){:target="_blank”}, as well as docs for your language of choice, for tips on optimizing performance.
Copy file name to clipboardExpand all lines: src/connections/destinations/catalog/amberflo/index.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,7 +5,7 @@ id: 62274854b16140600b51d1cd
5
5
---
6
6
## Amberflo Destination
7
7
8
-
[Amberflo](https://www.amberflo.io/) provides cloud based usage metering, pricing, and billing. Meter any infrastructure, platform, application, custom resource, event, or feature. Amberflo provides an end-to-end usage platform engine that serves as the system of records and single source of truth. It's platform is built on top of the metering service, Amberflo Metering Cloud. It is built on cloud platform design principles of durability, availability, scalability, and cost-effectiveness with specialized logic built-in to ensure accuracy of the metering system - that is that each record is processed once, and once only, and that duplicate records sent are automatically de-duped. Amberflo Billing Cloud is a decoupled (yet integrated) application that is built on the Metering Cloud. It allows users to create, model, and manage usage-based pricing plans with full flexibility over modern sales artifacts such as prepaid credits, rewards, promotions, and custom currency creation.
8
+
[Amberflo](https://www.amberflo.io/){:target="_blank”} provides cloud based usage metering, pricing, and billing. Meter any infrastructure, platform, application, custom resource, event, or feature. Amberflo provides an end-to-end usage platform engine that serves as the system of records and single source of truth. It's platform is built on top of the metering service, Amberflo Metering Cloud. It is built on cloud platform design principles of durability, availability, scalability, and cost-effectiveness with specialized logic built-in to ensure accuracy of the metering system - that is that each record is processed once, and once only, and that duplicate records sent are automatically de-duped. Amberflo Billing Cloud is a decoupled (yet integrated) application that is built on the Metering Cloud. It allows users to create, model, and manage usage-based pricing plans with full flexibility over modern sales artifacts such as prepaid credits, rewards, promotions, and custom currency creation.
9
9
10
10
This destination is maintained by Amberflo. For any issues with the destination, [contact the Amberflo Support team](mailto:[email protected]).
11
11
@@ -16,7 +16,7 @@ This destination is maintained by Amberflo. For any issues with the destination,
16
16
1. From the Destinations catalog page in the Segment App, click **Add Destination**.
17
17
2. Search for "Amberflo" in the Destinations Catalog, and select the "Amberflo" destination.
18
18
3. Choose which Source should send data to the "Amberflo" destination.
19
-
4. Go to the [Amberflo dashboard](https://ui.amberflo.io/settings/account/api-keys), find and copy the "API key".
19
+
4. Go to the [Amberflo dashboard](https://ui.amberflo.io/settings/account/api-keys){:target="_blank”}, find and copy the "API key".
20
20
5. Enter the "API Key" in the "Amberflo" destination settings in Segment.
21
21
22
22
## Supported methods
@@ -25,7 +25,7 @@ Amberflo supports the following methods, as specified in the [Segment Spec](/doc
25
25
26
26
### Page
27
27
28
-
Send [Page](/docs/connections/spec/page) calls the [Amberflo Ingestion API](https://docs.amberflo.io/reference/post_ingest) to ingest as a meter with value 1. For example:
28
+
Send [Page](/docs/connections/spec/page) calls the [Amberflo Ingestion API](https://docs.amberflo.io/reference/post_ingest){:target="_blank”} to ingest as a meter with value 1. For example:
29
29
```js
30
30
analytics.page({
31
31
userId:"some_user_id",
@@ -72,7 +72,7 @@ curl --request POST \
72
72
73
73
### Screen
74
74
75
-
Send [Screen](/docs/connections/spec/screen) calls the [Amberflo Ingestion API](https://docs.amberflo.io/reference/post_ingest) to ingest as a meter with value 1. For example:
75
+
Send [Screen](/docs/connections/spec/screen) calls the [Amberflo Ingestion API](https://docs.amberflo.io/reference/post_ingest){:target="_blank”} to ingest as a meter with value 1. For example:
76
76
77
77
```js
78
78
analytics.screen({
@@ -109,7 +109,7 @@ curl --request POST \
109
109
110
110
### Identify
111
111
112
-
Send [Identify](/docs/connections/spec/identify) calls the [Amberflo Customers API](https://docs.amberflo.io/reference/post_customers) to create a customer in Amberflo or update if the customer already exists. For example:
112
+
Send [Identify](/docs/connections/spec/identify) calls the [Amberflo Customers API](https://docs.amberflo.io/reference/post_customers){:target="_blank”} to create a customer in Amberflo or update if the customer already exists. For example:
113
113
114
114
```js
115
115
analytics.identify({
@@ -155,7 +155,7 @@ curl --request PUT \
155
155
156
156
### Track
157
157
158
-
Send [Track](/docs/connections/spec/track) calls [Amberflo Ingestion API](https://docs.amberflo.io/reference/post_ingest) to ingest as a meter with value of `properties.value` or 1 if value is not set. For example:
158
+
Send [Track](/docs/connections/spec/track) calls [Amberflo Ingestion API](https://docs.amberflo.io/reference/post_ingest){:target="_blank”} to ingest as a meter with value of `properties.value` or 1 if value is not set. For example:
0 commit comments