Skip to content

Commit fdc400b

Browse files
committed
PCX review
1 parent 4826f79 commit fdc400b

File tree

8 files changed

+60
-53
lines changed

8 files changed

+60
-53
lines changed

src/content/docs/r2/api/tokens.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -47,14 +47,14 @@ Jurisdictional buckets can only be accessed via the corresponding jurisdictional
4747

4848
| Permission | Description |
4949
| ------------------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
50-
| Admin Read & Write | Allows the ability to create, list, and delete buckets, edit bucket configuration, read, write, and list objects, and read and write access to data catalog tables and associated metadata. |
51-
| Admin Read only | Allows the ability to list buckets and view bucket configuration, read and list objects, and read access to data catalog tables and associated metadata. |
50+
| Admin Read & Write | Allows the ability to create, list, and delete buckets, edit bucket configuration, read, write, and list objects, and read and write to data catalog tables and associated metadata. |
51+
| Admin Read only | Allows the ability to list buckets and view bucket configuration, read and list objects, and read from the data catalog tables and associated metadata. |
5252
| Object Read & Write | Allows the ability to read, write, and list objects in specific buckets. |
5353
| Object Read only | Allows the ability to read and list objects in specific buckets. |
5454

5555
:::note
5656

57-
Currently Admin Read & Write or Admin Read only permission is required to interact with and query [R2 Data Catalog](/r2/data-catalog/).
57+
Currently **Admin Read & Write** or **Admin Read only** permission is required to interact with [R2 Data Catalog](/r2/data-catalog/).
5858

5959
:::
6060

src/content/docs/r2/data-catalog/config-examples/pyiceberg.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,8 @@ Below is an example of using [PyIceberg](https://py.iceberg.apache.org/) to conn
88
## Prerequisites
99

1010
- Sign up for a [Cloudflare account](https://dash.cloudflare.com/sign-up/workers-and-pages).
11-
- Create an [R2 bucket](/r2/buckets/) and enable the data catalog.
12-
- Create an [R2 API token](/r2/api/tokens/) with both [R2 and data catalog permissions](/r2/api/tokens/#permissions).
11+
- [Create an R2 bucket](/r2/buckets/create-buckets/) and enable the data catalog.
12+
- [Create an R2 API token](/r2/api/tokens/) with both [R2 and data catalog permissions](/r2/api/tokens/#permissions).
1313
- Install the [PyIceberg](https://py.iceberg.apache.org/#installation) and [PyArrow](https://arrow.apache.org/docs/python/install.html) libraries.
1414

1515
## Example usage

src/content/docs/r2/data-catalog/config-examples/snowflake.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,13 +8,13 @@ Below is an example of using [Snowflake](https://docs.snowflake.com/en/user-guid
88
## Prerequisites
99

1010
- Sign up for a [Cloudflare account](https://dash.cloudflare.com/sign-up/workers-and-pages).
11-
- Create an [R2 bucket](/r2/buckets/) and enable the data catalog.
12-
- Create an [R2 API token](/r2/api/tokens/) with both [R2 and data catalog permissions](/r2/api/tokens/#permissions).
11+
- [Create an R2 bucket](/r2/buckets/create-buckets/) and enable the data catalog.
12+
- [Create an R2 API token](/r2/api/tokens/) with both [R2 and data catalog permissions](/r2/api/tokens/#permissions).
1313
- A [Snowflake](https://www.snowflake.com/) account with the necessary privileges to create external volumes and catalog integrations.
1414

1515
## Example usage
1616

17-
In your Snowflake [SQL worksheet](https://docs.snowflake.com/en/user-guide/ui-snowsight-worksheets-gs) or [notebook](https://docs.snowflake.com/en/user-guide/ui-snowsight/notebooks) run the following commands:
17+
In your Snowflake [SQL worksheet](https://docs.snowflake.com/en/user-guide/ui-snowsight-worksheets-gs) or [notebook](https://docs.snowflake.com/en/user-guide/ui-snowsight/notebooks), run the following commands:
1818

1919
```sql
2020
-- Create a database (if you don't already have one) to organize your external data

src/content/docs/r2/data-catalog/config-examples/spark.mdx

Lines changed: 28 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -3,20 +3,25 @@ title: Spark
33
pcx_content_type: example
44
---
55

6-
Below is an example of how you can build an [Apache Spark](https://spark.apache.org/) application (with Scala) which connects to the R2 Data Catalog. This application is built to run locally, but it can be adapted to run on a cluster.
6+
import { FileTree } from "~/components"
7+
8+
9+
Below is an example of how you can build an [Apache Spark](https://spark.apache.org/) application (with Scala) which connects to R2 Data Catalog. This application is built to run locally, but it can be adapted to run on a cluster.
710

811
## Prerequisites
912

1013
- Sign up for a [Cloudflare account](https://dash.cloudflare.com/sign-up/workers-and-pages).
11-
- Create an [R2 bucket](/r2/buckets/) and enable the data catalog.
12-
- Create an [R2 API token](/r2/api/tokens/) with both [R2 and data catalog permissions](/r2/api/tokens/#permissions).
14+
- [Create an R2 bucket](/r2/buckets/create-buckets/) and enable the data catalog.
15+
- [Create an R2 API token](/r2/api/tokens/) with both [R2 and data catalog permissions](/r2/api/tokens/#permissions).
1316
- Install Java 17, Spark 3.5.3, and SBT 1.10.11
1417
- Note: The specific versions of tools are critical for getting things to work in this example.
1518
- Tip: [“SDKMAN”](https://sdkman.io/) is a convenient package manager for installing SDKs.
1619

1720
## Example usage
1821

19-
To start, create a new empty project directory somewhere on your machine. Inside that directory, create the following file at `src/main/scala/com/example/R2DataCatalogDemo.scala`. This will serve as the main entry point for your Spark application.
22+
To start, create a new empty project directory somewhere on your machine.
23+
24+
Inside that directory, create the following file at `src/main/scala/com/example/R2DataCatalogDemo.scala`. This will serve as the main entry point for your Spark application.
2025

2126
```java
2227
package com.example
@@ -105,7 +110,7 @@ To enable the [sbt-assembly plugin](https://github.com/sbt/sbt-assembly?tab=read
105110
addSbtPlugin("com.eed3si9n" % "sbt-assembly" % "1.2.0")
106111
```
107112

108-
Make sure Java, Spark, and sbt are installed and available in your shell. If you're using SDKMAN, you can install them as shown below:
113+
Make sure Java, Spark, and sbt are installed and available in your shell. If you are using SDKMAN, you can install them as shown below:
109114

110115
```bash
111116
sdk install java 17.0.14-amzn
@@ -121,7 +126,7 @@ sbt clean assembly
121126

122127
After building, the output JAR should be located at `target/scala-2.12/R2DataCatalogDemo-assembly-1.0.jar`.
123128

124-
To run the application, you'll use `spark-submit`. Below is an example shell script (`submit.sh`) that includes the necessary Java compatability flags for Spark on Java 17:
129+
To run the application, you will use `spark-submit`. Below is an example shell script (`submit.sh`) that includes the necessary Java compatability flags for Spark on Java 17:
125130

126131
```
127132
# We need to set these "--add-opens" so that Spark can run on Java 17 (it needs access to
@@ -142,23 +147,22 @@ chmod +x submit.sh
142147

143148
At this point, your project directory should be structured like this:
144149

145-
```
146-
.
147-
├── Makefile
148-
├── README.md
149-
├── build.sbt
150-
├── project
151-
│ ├── assembly.sbt
152-
│ ├── build.properties
153-
│ └── project
154-
├── spark-submit.sh
155-
└── src
156-
└── main
157-
└── scala
158-
└── com
159-
└── example
160-
└── R2DataCatalogDemo.scala
161-
```
150+
<FileTree>
151+
- Makefile
152+
- README.md
153+
- build.sbt
154+
- project
155+
- assembly.sbt
156+
- build.properties
157+
- project
158+
- spark-submit.sh
159+
- src
160+
- main
161+
- scala
162+
- com
163+
- example
164+
- R2DataCatalogDemo.scala
165+
</FileTree>
162166

163167
Before submitting the job, make sure you have the required environment variable set for your catalog URI, warehouse, and [Cloudflare API token](/r2/api/tokens/).
164168

@@ -168,7 +172,7 @@ export WAREHOUSE=
168172
export TOKEN=
169173
```
170174

171-
You're now ready to run the job:
175+
You are now ready to run the job:
172176

173177
```bash
174178
./submit.sh

src/content/docs/r2/data-catalog/get-started.mdx

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
---
22
pcx_content_type: get-started
3-
title: Get started
3+
title: Getting started
44
head: []
55
sidebar:
66
order: 2
@@ -44,7 +44,7 @@ This guide will instruct you through:
4444
npx wrangler login
4545
```
4646

47-
2. Then, enable the catalog on your chosen R2 bucket:
47+
2. Enable the catalog on your chosen R2 bucket:
4848

4949
```
5050
npx wrangler r2 bucket r2-data-catalog-tutorial
@@ -104,20 +104,20 @@ Iceberg clients (including [PyIceberg](https://py.iceberg.apache.org/)) must aut
104104

105105
6. Select **Create API Token**.
106106

107-
7. Note the **Token value**, you will need this.
107+
7. Note the **Token value**.
108108

109109
</Steps>
110110

111111
## 4. Install uv
112112

113-
Next, you'll need to install a Python package manager, in this guide we'll be using [uv](https://docs.astral.sh/uv/). If you don't already have uv installed, follow the [installing uv guide](https://docs.astral.sh/uv/getting-started/installation/).
113+
You need to install a Python package manager. In this guide, use [uv](https://docs.astral.sh/uv/). If you do not already have uv installed, follow the [installing uv guide](https://docs.astral.sh/uv/getting-started/installation/).
114114

115115
## 5. Install marimo
116116

117-
We'll be using [marimo](https://github.com/marimo-team/marimo) as a Python notebook.
117+
We will use [marimo](https://github.com/marimo-team/marimo) as a Python notebook.
118118

119119
<Steps>
120-
1. Create a directory where our notebook will live:
120+
1. Create a directory where our notebook will be stored:
121121

122122
```
123123
mkdir r2-data-catalog-notebook
@@ -269,7 +269,7 @@ We'll be using [marimo](https://github.com/marimo-team/marimo) as a Python noteb
269269
app.run()
270270
```
271271

272-
3. Replace the `CATALOG_URI`, `WAREHOUSE` and `TOKEN` variables with your values from sections **2** and **3** respectively.
272+
3. Replace the `CATALOG_URI`, `WAREHOUSE`, and `TOKEN` variables with your values from sections **2** and **3** respectively.
273273

274274
</Steps>
275275
In the Python notebook above, you:
@@ -286,7 +286,7 @@ In the Python notebook above, you:
286286

287287
<LinkCard
288288
title="Managing catalogs"
289-
href="/r2/data-catalog/managing-catalogs/"
289+
href="/r2/data-catalog/manage-catalogs/"
290290
description="Enable or disable R2 Data Catalog on your bucket, retrieve configuration details, and authenticate your Iceberg engine."
291291
/>
292292

src/content/docs/r2/data-catalog/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ R2 Data Catalog is a managed [Apache Iceberg](https://iceberg.apache.org/) data
1919

2020
R2 Data Catalog makes it easy to turn an R2 bucket into a data warehouse or lakehouse for a variety of analytical workloads including log analytics, business intelligence, and data pipelines. R2's zero-egress fee model means that data users and consumers can access and analyze data from different clouds, data platforms, or regions without incurring transfer costs.
2121

22-
Refer to the [get started guide](/r2/data-catalog/get-started/) to start with R2 Data Catalog.
22+
To get started with R2 Data Catalog, refer to the [R2 Data Catalog: Getting started](/r2/data-catalog/get-started/).
2323

2424
## What is Apache Iceberg?
2525

@@ -49,7 +49,7 @@ Similarly, data catalogs ensure consistent, coordinated access, which allows mul
4949

5050
<LinkCard
5151
title="Managing catalogs"
52-
href="/r2/data-catalog/managing-catalogs/"
52+
href="/r2/data-catalog/manage-catalogs/"
5353
description="Enable or disable R2 Data Catalog on your bucket, retrieve configuration details, and authenticate your Iceberg engine."
5454
/>
5555

src/content/docs/r2/data-catalog/managing-catalogs.mdx renamed to src/content/docs/r2/data-catalog/manage-catalogs.mdx

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
---
22
pcx_content_type: configuration
3-
title: Managing catalogs
3+
title: Manage catalogs
44
description: Understand how to manage Iceberg REST catalogs associated with R2 buckets
55
sidebar:
66
order: 3
@@ -18,7 +18,10 @@ import {
1818
LinkCard,
1919
} from "~/components";
2020

21-
Learn how to enable and disable [R2 Data Catalog](/r2/data-catalog/) on your buckets and authenticate Iceberg engines using API tokens.
21+
Learn how to:
22+
23+
- Enable and disable [R2 Data Catalog](/r2/data-catalog/) on your buckets.
24+
- Authenticate Iceberg engines using API tokens.
2225

2326
## Enable R2 Data Catalog on a bucket
2427

@@ -72,7 +75,7 @@ To connect your Iceberg engine to R2 Data Catalog, you will need a Cloudflare AP
7275
2. Copy the **Token value** from your new API token.
7376
3. In your engine configuration, provide this token as a bearer token.
7477
Internally, this token will be sent as:
75-
78+
7679
```
7780
Authorization: Bearer <TOKEN_VALUE>
7881
```

src/content/docs/r2/pricing.mdx

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -24,13 +24,13 @@ To learn about potential cost savings from using R2, refer to the [R2 pricing ca
2424

2525
## R2 pricing
2626

27-
| | Standard storage | Infrequent Access storage<InlineBadge preset="beta" /> |
28-
| ---------------------------------- | ------------------------ | ------------------------------------------------------ |
29-
| Storage | $0.015 / GB-month | $0.01 / GB-month |
30-
| Class A Operations | $4.50 / million requests | $9.00 / million requests |
31-
| Class B Operations | $0.36 / million requests | $0.90 / million requests |
32-
| Data Retrieval (processing) | None | $0.01 / GB |
33-
| Egress (data transfer to Internet) | Free [^1] | Free [^1] |
27+
| | Standard storage | Infrequent Access storage <InlineBadge preset="beta" /> |
28+
| ---------------------------------- | ------------------------ | ------------------------------------------------------- |
29+
| Storage | $0.015 / GB-month | $0.01 / GB-month |
30+
| Class A Operations | $4.50 / million requests | $9.00 / million requests |
31+
| Class B Operations | $0.36 / million requests | $0.90 / million requests |
32+
| Data Retrieval (processing) | None | $0.01 / GB |
33+
| Egress (data transfer to Internet) | Free [^1] | Free [^1] |
3434

3535
### Free tier
3636

@@ -82,7 +82,7 @@ For objects stored in Infrequent Access storage, you will be charged for the obj
8282

8383
## R2 Data Catalog pricing
8484

85-
R2 Data Catalog is in **public beta**, and any developer with [R2 subscription](/r2/pricing/) can start using it. Currently, outside of standard R2 storage and operations, you will not be billed for your use of R2 Data Catalog. We'll provide at least 30 days notice before we make any changes or start charging for usage
85+
R2 Data Catalog is in **public beta**, and any developer with an [R2 subscription](/r2/pricing/) can start using it. Currently, outside of standard R2 storage and operations, you will not be billed for your use of R2 Data Catalog. We will provide at least 30 days' notice before we make any changes or start charging for usage.
8686

8787
To learn more about our thinking on future pricing, refer to the [R2 Data Catalog announcement blog](https://blog.cloudflare.com/r2-data-catalog-public-beta).
8888

0 commit comments

Comments
 (0)