Skip to content

Commit b258117

Browse files
Merge pull request #48331 from jldohmann/OSDOCS-3867
OSDOCS-3867: add references to throughput issues doc
2 parents 5cc9c94 + fcb79e9 commit b258117

File tree

2 files changed

+23
-27
lines changed

2 files changed

+23
-27
lines changed
Lines changed: 16 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -1,50 +1,39 @@
11
// Module filename: nw-throughput-troubleshoot.adoc
22
// Module included in the following assemblies:
3-
// * networking/configuring-routing.adoc
3+
// * networking/routes/route-configuration.adoc
44

5+
:_content-type: CONCEPT
56
[id="nw-throughput-troubleshoot_{context}"]
6-
= Troubleshooting throughput issues
7-
Sometimes applications deployed through {product-title} can cause
8-
network throughput issues such as unusually high latency between
9-
specific services.
7+
= Throughput issue troubleshooting methods
108

11-
Use the following methods to analyze performance issues if pod logs do not
12-
reveal any cause of the problem:
9+
Sometimes applications deployed by using {product-title} can cause network throughput issues, such as unusually high latency between specific services.
1310

14-
* Use a packet analyzer, such as ping or link:http://www.tcpdump.org/[tcpdump]
15-
to analyze traffic between a pod and its node.
11+
If pod logs do not reveal any cause of the problem, use the following methods to analyze performance issues:
12+
13+
* Use a packet analyzer, such as `ping` or `tcpdump` to analyze traffic between a pod and its node.
1614
+
17-
For example, run the tcpdump tool on each pod while reproducing the behavior
18-
that led to the issue.
19-
Review the captures on both sides to compare send and receive timestamps to
20-
analyze the latency of traffic to and from a pod.
21-
Latency can occur in {product-title} if a node interface is overloaded with
22-
traffic from other pods, storage devices, or the data plane.
15+
For example, link:https://access.redhat.com/solutions/4569211[run the `tcpdump` tool on each pod] while reproducing the behavior that led to the issue. Review the captures on both sides to compare send and receive timestamps to analyze the latency of traffic to and from a pod. Latency can occur in {product-title} if a node interface is overloaded with traffic from other pods, storage devices, or the data plane.
2316
+
2417
[source,terminal]
2518
----
2619
$ tcpdump -s 0 -i any -w /tmp/dump.pcap host <podip 1> && host <podip 2> <1>
2720
----
2821
+
29-
<1> `podip` is the IP address for the pod. Run the `oc get pod <pod_name> -o wide` command to get
30-
the IP address of a pod.
22+
<1> `podip` is the IP address for the pod. Run the `oc get pod <pod_name> -o wide` command to get the IP address of a pod.
3123
+
32-
tcpdump generates a file at `/tmp/dump.pcap` containing all traffic between
33-
these two pods. Ideally, run the analyzer shortly
34-
before the issue is reproduced and stop the analyzer shortly after the issue
35-
is finished reproducing to minimize the size of the file.
36-
You can also run a packet analyzer between the nodes (eliminating the SDN from
37-
the equation) with:
24+
The `tcpdump` command generates a file at `/tmp/dump.pcap` containing all traffic between these two pods. You can run the analyzer shortly before the issue is reproduced and stop the analyzer shortly after the issue is finished reproducing to minimize the size of the file. You can also link:https://access.redhat.com/solutions/5074041[run a packet analyzer between the nodes] (eliminating the SDN from the equation) with:
3825
+
3926
[source,terminal]
4027
----
4128
$ tcpdump -s 0 -i any -w /tmp/dump.pcap port 4789
4229
----
4330
44-
* Use a bandwidth measuring tool, such as iperf, to measure streaming throughput
45-
and UDP throughput. Run the tool from the pods first, then from the nodes,
46-
to locate any bottlenecks.
31+
* Use a bandwidth measuring tool, such as link:https://access.redhat.com/solutions/6129701[`iperf`], to measure streaming throughput and UDP throughput. Locate any bottlenecks by running the tool from the pods first, and then running it from the nodes.
4732
4833
ifdef::openshift-enterprise,openshift-webscale[]
49-
** For information on installing and using iperf, see this link:https://access.redhat.com/solutions/33103[Red Hat Solution].
34+
** For information on installing and using `iperf`, see this link:https://access.redhat.com/solutions/33103[Red Hat Solution].
5035
endif::openshift-enterprise,openshift-webscale[]
36+
37+
* In some cases, the cluster may mark the node with the router pod as unhealthy due to latency issues. Use worker latency profiles to adjust the frequency that the cluster waits for a status update from the node before taking action.
38+
39+
* If your cluster has designated lower-latency and higher-latency nodes, configure the `spec.nodePlacement` field in the Ingress Controller to control the placement of the router pod.

networking/routes/route-configuration.adoc

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,13 @@ include::modules/nw-enforcing-hsts-per-domain.adoc[leveloffset=+2]
3333
//Troubleshooting Throughput Issues
3434
include::modules/nw-throughput-troubleshoot.adoc[leveloffset=+1]
3535

36+
[role="_additional-resources"]
37+
.Additional resources
38+
39+
* xref:../../nodes/edge/nodes-edge-remote-workers.html#nodes-edge-remote-workers-latency[Latency spikes or temporary reduction in throughput to remote workers]
40+
41+
* xref:../ingress-operator.html#nw-ingress-controller-configuration-parameters_configuring-ingress[Ingress Controller configuration parameters]
42+
3643
//Using cookies to keep route statefulness
3744
include::modules/nw-using-cookies-keep-route-statefulness.adoc[leveloffset=+1]
3845

0 commit comments

Comments
 (0)