Skip to content

Commit dc8801c

Browse files
committed
more spellings
1 parent 524d702 commit dc8801c

File tree

32 files changed

+153
-155
lines changed

32 files changed

+153
-155
lines changed

docs/en/chdb/install/python.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,7 @@ res = chdb.query('select * from file("data.csv", CSV)', 'CSV'); print(res)
6767
print(f"SQL read {res.rows_read()} rows, {res.bytes_read()} bytes, elapsed {res.elapsed()} seconds")
6868
```
6969

70-
**Pandas dataframe output**
70+
**Pandas DataFrame output**
7171
```python
7272
# See more in https://clickhouse.com/docs/en/interfaces/formats
7373
chdb.query('select * from file("data.parquet", Parquet)', 'Dataframe')

docs/en/cloud/reference/changelog.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -966,7 +966,7 @@ Adds support for a subset of features in ClickHouse 23.1, for example:
966966
- Added a feature to explore all available databases
967967
- Fixed synchronization of database with AggregationFunction type
968968
- [DBT-clickhouse](/docs/en/integrations/data-ingestion/etl-tools/dbt/index.md): Added support for the latest DBT version v1.4.1
969-
- [Python client](/docs/en/integrations/language-clients/python/index.md): Improved proxy and ssh tunneling support; added a number of fixes and performance optimizations for Pandas dataframes
969+
- [Python client](/docs/en/integrations/language-clients/python/index.md): Improved proxy and ssh tunneling support; added a number of fixes and performance optimizations for Pandas DataFrames
970970
- [Nodejs client](/docs/en/integrations/language-clients/js.md): Released ability to attach `query_id` to query result, which can be used to retrieve query metrics from the `system.query_log`
971971
- [Golang client](/docs/en/integrations/language-clients/go/index.md): Optimized network connection with ClickHouse Cloud
972972

docs/en/cloud/security/cmek.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ Coming soon:
1818

1919
- [Azure Key Vault](https://azure.microsoft.com/en-us/products/key-vault) for services hosted on: Azure
2020
- [GCP Cloud Key Management](https://cloud.google.com/security-key-management) for services hosted on: GCP
21-
- [Hashicorp Vault](https://www.hashicorp.com/products/vault) for services hosted on: AWS, Azure, GCP
21+
- [HashiCorp Vault](https://www.hashicorp.com/products/vault) for services hosted on: AWS, Azure, GCP
2222

2323
:::warning
2424
Deleting a KMS key used to encrypt a ClickHouse Cloud service will cause your ClickHouse service to be stopped and its data will be unretrievable, along with existing backups.

docs/en/cloud/security/gcp-private-service-connect.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -108,7 +108,7 @@ In the Google Cloud console, navigate to **Network services -> Private Service C
108108
Open the Private Service Connect creation dialog by clicking on the **Connect Endpoint** button.
109109

110110
- **Target**: use **Published service**
111-
- **Target service**: use **endpointServiceId** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step.
111+
- **Target service**: use `endpointServiceId` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step.
112112
- **Endpoint name**: set a name for the PSC **Endpoint name**.
113113
- **Network/Subnetwork/IP address**: Choose the network you want to use for the connection. You will need to create an IP address or use an existing one for the Private Service Connect endpoint. In our example, we pre-created an address with the name **your-ip-address** and assigned IP address `10.128.0.2`
114114
- To make the endpoint available from any region, you can enable the **Enable global access** checkbox.
@@ -172,7 +172,7 @@ output "psc_connection_id" {
172172
```
173173

174174
:::note
175-
TARGET - Use **endpointServiceId** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
175+
TARGET - Use `endpointServiceId` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
176176
:::
177177

178178
## Setting up DNS
@@ -228,7 +228,7 @@ gcloud dns \
228228
--rrdatas="10.128.0.2"
229229
```
230230
:::note
231-
DNS_RECORD - use **privateDnsHostname** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
231+
DNS_RECORD - use `privateDnsHostname` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
232232
:::
233233

234234
### Option 3: Using Terraform
@@ -256,12 +256,12 @@ resource "google_dns_record_set" "psc_dns_record" {
256256
```
257257

258258
:::note
259-
DNS_NAME - Use **privateDnsHostname** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
259+
DNS_NAME - Use `privateDnsHostname` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
260260
:::
261261

262262
## Verify DNS setup
263263

264-
DNS_RECORD - Use **privateDnsHostname** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
264+
DNS_RECORD - Use `privateDnsHostname` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
265265

266266
```bash
267267
ping $DNS_RECORD
@@ -387,7 +387,7 @@ curl --silent --user ${KEY_ID:?}:${KEY_SECRET:?} -X PATCH -H "Content-Type: appl
387387

388388
## Accessing instance using Private Service Connect
389389

390-
Each instance with configured Private Service Connect filters has two endpoints: public and private. In order to connect using Private Service Connect, you need to use a private endpoint, see use **endpointServiceId** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
390+
Each instance with configured Private Service Connect filters has two endpoints: public and private. In order to connect using Private Service Connect, you need to use a private endpoint, see use `endpointServiceId` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
391391

392392
:::note
393393
Private DNS hostname is only available from your GCP VPC. Do not try to resolve the DNS host from a machine that resides outside of GCP VPC.
@@ -421,7 +421,7 @@ In this example, connection to the `xxxxxxx.yy-xxxxN.p.gcp.clickhouse.cloud` hos
421421

422422
### Test DNS setup
423423

424-
DNS_NAME - Use **privateDnsHostname** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
424+
DNS_NAME - Use `privateDnsHostname` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
425425

426426
```bash
427427
nslookup $DNS_NAME
@@ -443,7 +443,7 @@ If you have problems with connecting using PSC link, check your connectivity usi
443443

444444
OpenSSL should be able to connect (see CONNECTED in the output). `errno=104` is expected.
445445

446-
DNS_NAME - Use **privateDnsHostname** from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
446+
DNS_NAME - Use `privateDnsHostname` from [Obtain GCP service attachment for Private Service Connect](#obtain-gcp-service-attachment-and-dns-name-for-private-service-connect) step
447447

448448
```bash
449449
openssl s_client -connect ${DNS_NAME}:9440

docs/en/integrations/data-ingestion/clickpipes/postgres/faq.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ sidebar_position: 2
99

1010
### How does idling affect my Postgres CDC ClickPipe?
1111

12-
If your ClickHouse Cloud service is idling, your Postgres CDC clickpipe will continue to sync data, your service will wake-up at the next sync interval to handle the incoming data. Once the sync is finished and the idle period is reached, your service will go back to idling.
12+
If your ClickHouse Cloud service is idling, your Postgres CDC ClickPipe will continue to sync data, your service will wake-up at the next sync interval to handle the incoming data. Once the sync is finished and the idle period is reached, your service will go back to idling.
1313

1414
As an example, if your sync interval is set to 30 mins and your service idle time is set to 10 mins, Your service will wake-up every 30 mins and be active for 10 mins, then go back to idling.
1515

@@ -48,9 +48,9 @@ Yes! ClickPipes for Postgres offers two ways to connect to databases in private
4848

4949
### How do you handle UPDATEs and DELETEs?
5050

51-
ClickPipes for Postgres captures both INSERTs and UPDATEs from Postgres as new rows with different versions (using the _peerdb_version column) in ClickHouse. The ReplacingMergeTree table engine periodically performs deduplication in the background based on the ordering key (ORDER BY columns), retaining only the row with the latest _peerdb_version.
51+
ClickPipes for Postgres captures both INSERTs and UPDATEs from Postgres as new rows with different versions (using the `_peerdb_` version column) in ClickHouse. The ReplacingMergeTree table engine periodically performs deduplication in the background based on the ordering key (ORDER BY columns), retaining only the row with the latest `_peerdb_` version.
5252

53-
DELETEs from Postgres are propagated as new rows marked as deleted (using the _peerdb_is_deleted column). Since the deduplication process is asynchronous, you might temporarily see duplicates. To address this, you need to handle deduplication at the query layer.
53+
DELETEs from Postgres are propagated as new rows marked as deleted (using the `_peerdb_is_deleted` column). Since the deduplication process is asynchronous, you might temporarily see duplicates. To address this, you need to handle deduplication at the query layer.
5454

5555
For more details, refer to:
5656

@@ -97,7 +97,7 @@ If you're noticing that the size of your Postgres replication slot keeps increas
9797

9898
4. **VACUUM and VACUUM ANALYZE**
9999
- Although necessary for database health, these operations can create extra WAL traffic—especially if they scan large tables.
100-
- Consider using autovacuum tuning parameters or scheduling manual VACUUMs during off-peak hours.
100+
- Consider using autovacuum tuning parameters or scheduling manual VACUUM operations during off-peak hours.
101101

102102
5. **Replication Consumer Not Actively Reading the Slot**
103103
- If your CDC pipeline (e.g., ClickPipes) or another replication consumer stops, pauses, or crashes, WAL data will accumulate in the slot.
@@ -136,13 +136,13 @@ As of now, you can create a ClickPipe only via the UI. However, we are actively
136136
137137
You cannot speed up an already running initial load. However, you can optimize future initial loads by adjusting certain settings. By default, the settings are configured with 4 parallel threads and a snapshot number of rows per partition set to 100,000. These are advanced settings and are generally sufficient for most use cases.
138138
139-
For Postgres versions 13 or lower, ctid range scans are slower, and these settings become more critical. In such cases, consider the following process to improve performance:
139+
For Postgres versions 13 or lower, CTID range scans are slower, and these settings become more critical. In such cases, consider the following process to improve performance:
140140
141141
1. **Drop the existing pipe**: This is necessary to apply new settings.
142142
2. **Delete destination tables on ClickHouse**: Ensure that the tables created by the previous pipe are removed.
143143
3. **Create a new pipe with optimized settings**: Typically, increase the snapshot number of rows per partition to between 1 million and 10 million, depending on your specific requirements and the load your Postgres instance can handle.
144144
145-
These adjustments should significantly enhance the performance of the initial load, especially for older Postgres versions. If you are using Postgres 14 or later, these settings are less impactful due to improved support for ctid range scans.
145+
These adjustments should significantly enhance the performance of the initial load, especially for older Postgres versions. If you are using Postgres 14 or later, these settings are less impactful due to improved support for CTID range scans.
146146
147147
### How should I scope my publications when setting up replication?
148148

docs/en/integrations/data-ingestion/clickpipes/postgres/postgres_generated_columns.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ When using PostgreSQL's generated columns in tables that are being replicated, t
77

88
## The Problem with Generated Columns
99

10-
1. **Not Published via pgoutput:** Generated columns are not published through the pgoutput logical replication plugin. This means that when you're replicating data from PostgreSQL to another system, the values of generated columns are not included in the replication stream.
10+
1. **Not Published via `pgoutput`:** Generated columns are not published through the `pgoutput` logical replication plugin. This means that when you're replicating data from PostgreSQL to another system, the values of generated columns are not included in the replication stream.
1111

1212
2. **Issues with Primary Keys:** If a generated column is part of your primary key, it can cause deduplication problems on the destination. Since the generated column values are not replicated, the destination system won't have the necessary information to properly identify and deduplicate rows.
1313

docs/en/integrations/data-ingestion/clickpipes/postgres/source/google-cloudsql.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
---
2-
sidebar_label: Google CloudSQL
3-
description: Set up Google CloudSQL Postgres instance as a source for ClickPipes
2+
sidebar_label: Google Cloud SQL
3+
description: Set up Google Cloud SQL Postgres instance as a source for ClickPipes
44
slug: /en/integrations/clickpipes/postgres/source/google-cloudsql
55
---
66

7-
# Google CloudSQL Postgres Source Setup Guide
7+
# Google Cloud SQL Postgres Source Setup Guide
88

99
:::info
1010

@@ -23,19 +23,19 @@ Anything on or after Postgres 12
2323

2424
1. Click on **Edit** button on the Overview page.
2525

26-
![Edit Button in CloudSQL Postgres](images/setup/google-cloudsql/edit.png)
26+
![Edit Button in Cloud SQL Postgres](images/setup/google-cloudsql/edit.png)
2727

2828

2929
2. Go to Flags and change `cloudsql.logical_decoding` to on and `wal_sender_timeout` to 0. These changes will need restarting your Postgres server.
3030

31-
![Change cloudsql.logical_decoding to on](images/setup/google-cloudsql/cloudsql_logical_decoding1.png)
32-
![Changed cloudsql.logical_decoding and wal_sender_timeout](images/setup/google-cloudsql/cloudsql_logical_decoding2.png)
31+
![Change `cloudsql.logical_decoding` to on](images/setup/google-cloudsql/cloudsql_logical_decoding1.png)
32+
![Changed `cloudsql.logical_decoding` and `wal_sender_timeout`](images/setup/google-cloudsql/cloudsql_logical_decoding2.png)
3333
![Restart Server](images/setup/google-cloudsql/cloudsql_logical_decoding3.png)
3434

3535

3636
## Creating ClickPipes User and Granting permissions
3737

38-
Connect to your CloudSQL Postgres through the admin user and run the below commands:
38+
Connect to your Cloud SQL Postgres through the admin user and run the below commands:
3939

4040
1. Create a Postgres user for exclusively ClickPipes.
4141

docs/en/integrations/data-ingestion/clickpipes/postgres/source/supabase.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ This is a guide on how to setup Supabase Postgres for usage in ClickPipes.
1010

1111
:::note
1212

13-
ClickPipes supports Supabase via IPv6 natively for seemless replication.
13+
ClickPipes supports Supabase via IPv6 natively for seamless replication.
1414

1515
:::
1616

docs/en/integrations/data-ingestion/data-formats/binary.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -197,7 +197,7 @@ This saves data to the [proto.bin](assets/proto.bin) file. ClickHouse also suppo
197197

198198
<CloudNotSupportedBadge/>
199199

200-
Another popular binary serialization format supported by ClickHouse is [Cap’n Proto](https://capnproto.org/). Similarly to `Protobuf` format, we have to define a schema file ([schema.capnp](assets/schema.capnp)) in our example:
200+
Another popular binary serialization format supported by ClickHouse is [Cap’n Proto](https://capnproto.org/). Similarly to `Protobuf` format, we have to define a schema file ([`schema.capnp`](assets/schema.capnp)) in our example:
201201

202202
```
203203
@0xec8ff1a10aa10dbe;

0 commit comments

Comments
 (0)