You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/modules/hdfs/pages/getting_started/first_steps.adoc
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
= First steps
2
2
:description: Deploy and verify an HDFS cluster with Stackable by setting up Zookeeper and HDFS components, then test file operations using WebHDFS API.
3
3
4
-
Once you have followed the steps in the xref:getting_started/installation.adoc[] section to install the operator and its dependencies, you will now deploy an HDFS cluster and its dependencies.
4
+
Once you have followed the steps in the xref:getting_started/installation.adoc[] section to install the operator and its dependencies, now deploy an HDFS cluster and its dependencies.
5
5
Afterward, you can <<_verify_that_it_works, verify that it works>> by creating, verifying and deleting a test file in HDFS.
6
6
7
7
== Setup
@@ -13,7 +13,7 @@ To deploy a Zookeeper cluster create one file called `zk.yaml`:
13
13
[source,yaml]
14
14
include::example$getting_started/zk.yaml[]
15
15
16
-
We also need to define a ZNode that will be used by the HDFS cluster to reference Zookeeper.
16
+
Define a ZNode that is used by the HDFS cluster to reference Zookeeper.
17
17
Create another file called `znode.yaml`:
18
18
19
19
[source,yaml]
@@ -94,7 +94,7 @@ Then use `curl` to issue a `PUT` command:
Copy file name to clipboardExpand all lines: docs/modules/hdfs/pages/getting_started/index.adoc
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,18 +1,18 @@
1
1
= Getting started
2
2
:description: Start with HDFS using the Stackable Operator. Install the Operator, set up your HDFS cluster, and verify its operation with this guide.
3
3
4
-
This guide will get you started with HDFS using the Stackable Operator.
5
-
It will guide you through the installation of the Operator and its dependencies, setting up your first HDFS cluster and verifying its operation.
4
+
This guide gets you started with HDFS using the Stackable operator.
5
+
It guides you through the installation of the operator and its dependencies, setting up your first HDFS cluster and verifying its operation.
6
6
7
7
== Prerequisites
8
8
9
-
You will need:
9
+
You need:
10
10
11
11
* a Kubernetes cluster
12
12
* kubectl
13
13
* optional: Helm
14
14
15
-
Resource sizing depends on cluster type(s), usage and scope, but as a starting point we recommend a minimum of the following resources for this operator:
15
+
Resource sizing depends on cluster type(s), usage and scope, but as a starting point the following resources are recommended as a minium requirement for this operator:
Copy file name to clipboardExpand all lines: docs/modules/hdfs/pages/index.adoc
+1-3Lines changed: 1 addition & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,9 +18,7 @@ The operator depends on the xref:zookeeper:index.adoc[] to operate a ZooKeeper c
18
18
19
19
== Getting started
20
20
21
-
Follow the xref:getting_started/index.adoc[Getting started guide] which will guide you through installing the Stackable
22
-
HDFS and ZooKeeper operators, setting up ZooKeeper and HDFS and writing a file to HDFS to verify that everything is set
23
-
up correctly.
21
+
Follow the xref:getting_started/index.adoc[Getting started guide] which guides you through installing the Stackable HDFS and ZooKeeper operators, setting up ZooKeeper and HDFS and writing a file to HDFS to verify that everything is set up correctly.
24
22
25
23
Afterwards you can consult the xref:usage-guide/index.adoc[] to learn more about tailoring your HDFS configuration to
26
24
your needs, or have a look at the <<demos, demos>> for some example setups.
Copy file name to clipboardExpand all lines: docs/modules/hdfs/pages/usage-guide/configuration-environment-overrides.adoc
+10-3Lines changed: 10 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -50,9 +50,10 @@ nameNodes:
50
50
replicas: 2
51
51
----
52
52
53
-
All override property values must be strings. The properties will be formatted and escaped correctly into the XML file.
53
+
All override property values must be strings.
54
+
The properties are formatted and escaped correctly into the XML file.
54
55
55
-
For a full list of configuration options we refer to the Apache Hdfs documentation for https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/hdfs-default.xml[hdfs-site.xml] and https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/core-default.xml[core-site.xml].
56
+
For a full list of configuration options refer to the Apache Hdfs documentation for https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/hdfs-default.xml[hdfs-site.xml] and https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/core-default.xml[core-site.xml].
56
57
57
58
=== The security.properties file
58
59
@@ -117,4 +118,10 @@ nameNodes:
117
118
replicas: 1
118
119
----
119
120
120
-
IMPORTANT: Some environment variables will be overriden by the operator and cannot be set manually by the user. These are `HADOOP_HOME`, `HADOOP_CONF_DIR`, `POD_NAME` and `ZOOKEEPER`.
121
+
IMPORTANT: Some environment variables are overridden by the operator and cannot be set manually by the user.
122
+
These are `HADOOP_HOME`, `HADOOP_CONF_DIR`, `POD_NAME` and `ZOOKEEPER`.
123
+
124
+
== Pod overrides
125
+
126
+
The HDFS operator also supports Pod overrides, allowing you to override any property that you can set on a Kubernetes Pod.
127
+
Read the xref:concepts:overrides.adoc#pod-overrides[Pod overrides documentation] to learn more about this feature.
Copy file name to clipboardExpand all lines: docs/modules/hdfs/pages/usage-guide/fuse.adoc
+11-12Lines changed: 11 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,15 +1,13 @@
1
1
= FUSE
2
2
:description: Use HDFS FUSE driver to mount HDFS filesystems into Linux environments via a Kubernetes Pod with necessary privileges and configurations.
3
3
4
-
Our images of Apache Hadoop do contain the necessary binaries and libraries to use the HDFS FUSE driver.
5
-
6
-
FUSE is short for _Filesystem in Userspace_ and allows a user to export a filesystem into the Linux kernel, which can then be mounted.
7
-
HDFS contains a native FUSE driver/application, which means that an existing HDFS filesystem can be mounted into a Linux environment.
4
+
HDFS contains a native FUSE driver/application, enabling you to mout an existing HDFS filesystem into a Linux environment.
8
5
6
+
Stackable images of Apache Hadoop contain the necessary binaries and libraries to use the HDFS FUSE driver.
9
7
To use the FUSE driver you can either copy the required files out of the image and run it on a host outside of Kubernetes or you can run it in a Pod.
10
-
This Pod, however, will need some extra capabilities.
8
+
This Pod, however, needs some extra capabilities.
11
9
12
-
This is an example Pod that will work_as long as the host system that is running the kubelet does support FUSE_:
10
+
This is an example Pod that works_as long as the host system that is running the kubelet does support FUSE_:
13
11
14
12
[source,yaml]
15
13
----
@@ -39,8 +37,9 @@ spec:
39
37
configMap:
40
38
name: <your hdfs here> <2>
41
39
----
42
-
<1> Ideally use the same version your HDFS is using. FUSE is baked in to our images as of SDP 23.11.
43
-
<2> This needs to be a reference to a discovery ConfigMap as written by our HDFS operator.
40
+
<1> Ideally use the same version your HDFS is using.
41
+
Stackable HDFS images contain the FUSE driver since 23.11.
42
+
<2> This needs to be a reference to a discovery ConfigMap as written by the HDFS operator.
44
43
45
44
[TIP]
46
45
.Privileged Pods
@@ -57,7 +56,7 @@ securityContext:
57
56
----
58
57
59
58
Unfortunately, there is no way around some extra privileges.
60
-
In Kubernetes the Pods usually share the Kernel with the host running the Kubelet, which means a Pod wanting to use FUSE will need access to the underlying Kernel modules.
59
+
In Kubernetes the Pods usually share the Kernel with the host running the Kubelet, which means a Pod wanting to use FUSE needs access to the underlying Kernel modules.
61
60
====
62
61
63
62
Inside this Pod you can get a shell (e.g. using `kubectl exec --stdin --tty hdfs-fuse -- /bin/bash`) to get access to a script called `fuse_dfs_wrapper` (it is in the `PATH` of our Hadoop images).
@@ -70,14 +69,14 @@ To mount HDFS call the script like this:
Copy file name to clipboardExpand all lines: docs/modules/hdfs/pages/usage-guide/index.adoc
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,6 @@
2
2
:description: Learn to configure and use the Stackable Operator for Apache HDFS. Ensure basic setup knowledge from the Getting Started guide before proceeding.
3
3
:page-aliases: ROOT:usage.adoc
4
4
5
-
This Section will help you to use and configure the Stackable Operator for Apache HDFS in various ways.
5
+
This Section helps you to use and configure the Stackable operator for Apache HDFS in various ways.
6
6
You should already be familiar with how to set up a basic instance.
7
7
Follow the xref:getting_started/index.adoc[] guide to learn how to set up a basic instance with all the required dependencies (for example ZooKeeper).
0 commit comments