You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/integrations/data-ingestion/data-ingestion-index.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,7 @@ For more information check out the pages below:
18
18
|[Azure Synapse](/integrations/azure-synapse)| A fully managed, cloud-based analytics service provided by Microsoft Azure, combining big data and data warehousing to simplify data integration, transformation, and analytics at scale using SQL, Apache Spark, and data pipelines. |
19
19
|[Azure Data Factory](/integrations/azure-data-factory)| A cloud-based data integration service that enables you to create, schedule, and orchestrate data workflows at scale. |
20
20
|[Apache Beam](/integrations/apache-beam)| An open-source, unified programming model that enables developers to define and execute both batch and stream (continuous) data processing pipelines. |
21
-
|[BladePipe](/integrations/bladepipe)| A real-time end-to-end data integration tool, boosting seamless data flow across platforms. |
21
+
|[BladePipe](/integrations/bladepipe)| A real-time end-to-end data integration tool with sub-second latency, boosting seamless data flow across platforms. |
22
22
|[dbt](/integrations/dbt)| Enables analytics engineers to transform data in their warehouses by simply writing select statements. |
23
23
|[dlt](/integrations/data-ingestion/etl-tools/dlt-and-clickhouse)| An open-source library that you can add to your Python scripts to load data from various and often messy data sources into well-structured, live datasets. |
24
24
|[Fivetran](/integrations/fivetran)| An automated data movement platform moving data out of, into and across your cloud data platforms. |
Copy file name to clipboardExpand all lines: docs/integrations/data-ingestion/etl-tools/bladepipe-and-clickhouse.md
+15-16Lines changed: 15 additions & 16 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -24,18 +24,18 @@ import ClickHouseSupportedBadge from '@theme/badges/ClickHouseSupported';
24
24
<ClickHouseSupportedBadge/>
25
25
26
26
27
-
<ahref="https://www.bladepipe.com/"target="_blank">BladePipe</a> is a real-time end-to-end data integration tool, boosting seamless data flow across platforms.
27
+
<ahref="https://www.bladepipe.com/"target="_blank">BladePipe</a> is a real-time end-to-end data integration tool with sub-second latency, boosting seamless data flow across platforms.
28
28
29
-
ClickHouse is one of BladePipe's pre-built connectors, allowing users to integrate data from various sources into ClickHouse. This page will show how to load data into ClickHouse in real time step by step.
29
+
ClickHouse is one of BladePipe's pre-built connectors, allowing users to integrate data from various sources into ClickHouse automatically. This page will show how to load data into ClickHouse in real time step by step.
30
30
31
31
32
-
## 1. Run BladePipe {#1-run-bladepipe}
32
+
## 1. Download and run BladePipe {#1-run-bladepipe}
33
33
1. Log in to <ahref="https://www.bladepipe.com/"target="_blank">BladePipe Cloud</a>.
34
34
35
35
2. Follow the instructions in <ahref="https://doc.bladepipe.com/productOP/byoc/installation/install_worker_docker"target="_blank">Install Worker (Docker)</a> or <ahref="https://doc.bladepipe.com/productOP/byoc/installation/install_worker_binary"target="_blank">Install Worker (Binary)</a> to download and install a BladePipe Worker.
36
36
37
37
:::note
38
-
Alternatively, you can download and deploy <ahref="https://doc.bladepipe.com/productOP/onPremise/installation/install_all_in_one_binary"target="_blank">BladePipe Enterprise</a>
38
+
Alternatively, you can download and deploy <ahref="https://doc.bladepipe.com/productOP/onPremise/installation/install_all_in_one_binary"target="_blank">BladePipe Enterprise</a>.
39
39
:::
40
40
41
41
## 2. Add ClickHouse as a target {#2-add-clickhouse-as-a-target}
@@ -45,38 +45,38 @@ ClickHouse is one of BladePipe's pre-built connectors, allowing users to integra
45
45
2. To use ClickHouse as a target, make sure that the user has SELECT, INSERT and common DDL permissions.
46
46
:::
47
47
48
-
1. In BladePipe, click **DataSource** > **Add DataSource**.
48
+
1. In BladePipe, click "DataSource" > "Add DataSource".
49
49
50
-
2. Select ClickHouse, and fill out the settings by providing your ClickHouse host and port, username and password, and click **Test Connection**.
50
+
2. Select `ClickHouse`, and fill out the settings by providing your ClickHouse host and port, username and password, and click "Test Connection".
51
51
52
52
<Imageimg={bp_ck_1}size="lg"borderalt="Add ClickHouse as a target" />
53
53
54
-
3. Click **Add DataSource** at the bottom, and a ClickHouse instance is added.
54
+
3. Click "Add DataSource" at the bottom, and a ClickHouse instance is added.
55
55
56
56
## 3. Add MySQL as a source {#3-add-mysql-as-a-source}
57
-
In this tutorial, we use a MySQL instance as the source, and expalin the process of loading MySQL data to ClickHouse.
57
+
In this tutorial, we use a MySQL instance as the source, and explain the process of loading MySQL data to ClickHouse.
58
58
59
59
:::note
60
60
To use MySQL as a source, make sure that the user has the <ahref="https://doc.bladepipe.com/dataMigrationAndSync/datasource_func/MySQL/privs_for_mysql"target="_blank">requried permissions</a>.
61
61
:::
62
62
63
-
1. In BladePipe, click **DataSource** > **Add DataSource**.
63
+
1. In BladePipe, click "DataSource" > "Add DataSource".
64
64
65
-
2. Select MySQL, and fill out the settings by providing your MySQL host and port, username and password, and click **Test Connection**.
65
+
2. Select `MySQL`, and fill out the settings by providing your MySQL host and port, username and password, and click "Test Connection".
66
66
67
67
<Imageimg={bp_ck_2}size="lg"borderalt="Add MySQL as a source" />
68
68
69
-
3. Click **Add DataSource** at the bottom, and a MySQL instance is added.
69
+
3. Click "Add DataSource" at the bottom, and a MySQL instance is added.
70
70
71
71
72
72
## 4. Create a pipeline {#4-create-a-pipeline}
73
73
74
-
1. In BladePipe, click **DataJob** > **Create DataJob**.
74
+
1. In BladePipe, click "DataJob" > "Create DataJob".
75
75
76
-
2. Select the added MySQL and ClickHouse instances and click **Test Connection** to ensure BladePipe is connected to the instances. Then, select the databases to be moved.
76
+
2. Select the added MySQL and ClickHouse instances and click "Test Connection" to ensure BladePipe is connected to the instances. Then, select the databases to be moved.
77
77
<Imageimg={bp_ck_3}size="lg"borderalt="Select source and target" />
78
78
79
-
3. Select **Incremental** for DataJob Type, together with the **Full Data** option.
79
+
3. Select "Incremental" for DataJob Type, together with the "Full Data" option.
@@ -93,12 +93,11 @@ To use MySQL as a source, make sure that the user has the <a href="https://doc.b
93
93
94
94
## 5. Verify the data {#5-verify-the-data}
95
95
1. Stop data write in MySQL instance and wait for ClickHouse to merge data.
96
-
97
96
:::note
98
97
Due to the unpredictable timing of ClickHouse's automatic merging, you can manually trigger a merging by running the `OPTIMIZE TABLE xxx FINAL;` command. Note that there is a chance that this manual merging may not always succeed.
99
98
100
99
Alternatively, you can run the `CREATE VIEW xxx_v AS SELECT * FROM xxx FINAL;` command to create a view and perform queries on the view to ensure the data is fully merged.
101
100
:::
102
101
103
-
2. Create a <ahref="https://doc.bladepipe.com/operation/job_manage/create_job/create_period_verification_correction_job"target="_blank">Verification DataJob</a>. Once the Verification DataJob is completed, review the results to confirm that the data in ClickHouse are the same as the data in MySQL.
102
+
2. Create a <ahref="https://doc.bladepipe.com/operation/job_manage/create_job/create_period_verification_correction_job"target="_blank">Verification DataJob</a>. Once the Verification DataJob is completed, review the results to confirm that the data in ClickHouse is the same as the data in MySQL.
0 commit comments