You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The recommendation is to set up Azure Data Explorer cluster with a new consumer group from the Event Hub or IoT Hub and wait for retention period to pass and fill Azure Data Explorer with the same data as Time Series Insights environment.
19
-
If telemetry data is required to be exported from Time Series Insights environment, the suggestion is to use Time Series Insights Query API to download the events in batches and serialize in required format.
20
-
For reference data, Time Series Insights Explorer or Reference Data API can be used to download reference data set and upload it into Azure Data Explorer as another table. Then, materialized views in Azure Data Explorer can be used to join reference data with telemetry data. Use materialized view with arg_max() aggregation function which will get the latest record per entity, as demonstrated in the following example. For more information about materialized views, read the following documentation: [Materialized views use cases](/azure/data-explorer/kusto/management/materialized-views/materialized-view-overview#materialized-views-use-cases).
18
+
[Eventhouse](/fabric/real-time-intelligence/eventhouse) is the time series database in Fabric Real-Time Intelligence. It serves as the target for migrating data away from Time Series Insights.
19
+
20
+
### Prerequisites
21
+
22
+
To begin the migration process, ensure that you have created a new Eventhouse in Fabric Real-Time Intelligence. Follow the documentation on how to [create an Eventhouse](/fabric/real-time-intelligence/create-eventhouse).
23
+
24
+
### Ingest new data
25
+
26
+
To start ingesting new data into your Eventhouse, follow these steps:
27
+
28
+
1. Configure your [Azure Event Hub](/azure/event-hubs/event-hubs-about) with a new consumer group.
29
+
30
+
2. Consume data from the data source and ingest it into your Eventhouse. Refer to the documentation on how to [ingest data from your EventHub](/fabric/real-time-intelligence/get-data-event-hub).
31
+
32
+
### Migrate historical data from Time Series Insights
33
+
34
+
If you need to export telemetry data from your Time Series Insights environment, you can use the Time Series Insights Query API to download the events in batches and serialize them in the required format. Depending on where you stored the exported data, you can ingest the data from [Azure Storage](/real-time-intelligence/get-data-azure-storage), [local files](/fabric/real-time-intelligence/get-data-local-file), or [OneLake](/fabric/real-time-intelligence/get-data-onelake).
35
+
36
+
### Migrate reference data
37
+
38
+
For migrating reference data, you can follow these steps:
39
+
40
+
1. Use Time Series Insights Explorer or the Reference Data API to download the reference data set. This will allow you to retrieve the necessary data for migration.
41
+
42
+
2. Once you have the reference data set, you can upload it into your Eventhouse as another table. This can be done using the [Upload](/fabric/real-time-intelligence/get-data-local-file) feature in Fabric Real-Time Intelligence. By uploading the reference data set, you will be able to access and utilize it within your Eventhouse environment.
43
+
44
+
Please note that these steps assume you have already created a new Eventhouse in Fabric Real-Time Intelligence as a prerequisite for the migration process.
21
45
22
-
```
23
-
.create materialized-view MVName on table T
24
-
{
25
-
T
26
-
| summarize arg_max(Column1,*) by Column2
27
-
}
28
-
```
29
46
## Translate Time Series Insights Queries to KQL
30
47
31
-
For queries, the recommendation is to use KQL in Azure Data Explorer.
48
+
For queries, the recommendation is to use KQL in Eventhouse.
0 commit comments