Skip to content

Commit aac6dd6

Browse files
authored
Update platform-tips.md
1 parent 659438b commit aac6dd6

File tree

1 file changed

+4
-35
lines changed

1 file changed

+4
-35
lines changed

user/enterprise/platform-tips.md

Lines changed: 4 additions & 35 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,6 @@ In order to obtain live logs from a specific running pod, one can run *on your l
2424

2525
> We strongly recommend setting up an instance of grabbing live logs from pods stdout and storing them in the logging storage of your choice. These stored logs can be useful when diagnosing or troubleshooting situations for pods that were killed and/or re-deployed. The size of the logs depends strictly on your usage, thus please adjust to your needs. As a rule of thumb, a 4-weeks of log storage would be recommended.
2626
27-
2827
### Worker logs
2928

3029
This section describes how to obtain worker logs with Ubuntu as the host operating system.
@@ -47,20 +46,17 @@ On the Worker, you can find the main log file at
4746
### Console access in TCIE 3.x
4847

4948
For TCIE 3.x, you gain access to individual pods through the `kubectl` command (The equivalent to `travis bash` in Enterprise 2.x versions)
50-
In order to run console commands, run console in `travis-api-pod`:
49+
In order to run console commands, run the console in `travis-api-pod`:
5150

5251
`kubectl exec -it [travis-api-pod] /app/script/console`
5352

54-
5553
## Cancel or Reset Stuck Jobs
5654

5755
Occasionally, jobs can get stuck in a `queued` state on the worker. To cancel or
5856
reset a large number of jobs, please execute the following steps:
5957

6058
**TCIE 3.x**: `kubectl exec -it [travis-api-pod]j /app/script/console` *on your local machine*
6159

62-
**TCIE 2.x**: `$ travis console` *on Platform host*
63-
6460
Then, please run:
6561

6662
```
@@ -96,46 +92,19 @@ Then, please run:
9692

9793
## Manage RabbitMQ in TCIE 3.x
9894

99-
RabbitMQ is now deployed in separate pod named `travisci-platform-rabbitmq-ha-0` and all Rabbit-related maintenance should be done there.
100-
In order to access RabbitMQ pod execute
95+
RabbitMQ is now deployed in a separate pod named `travisci-platform-rabbitmq-ha-0`, and all Rabbit-related maintenance should be done there.
96+
In order to access the RabbitMQ pod, execute
10197

10298
`kubectl exec -it travisci-platform-rabbitmq-ha-0 bash`
10399

104100
and perform any necessary actions.
105101

106102
The RabbitMQ management UI is available under `https://[platform-hostname]/amqp_ui`.
107103

108-
## Reset the RabbitMQ Certificate in TCIE 2.x
109-
110-
After an upgrade of Replicated 2.8.0 to a newer version, occasionally the service
111-
restarts with the following error:
112-
113-
```
114-
$ docker inspect --format '{% raw %}{{.State.Error}}{% endraw %}' focused_yalow
115-
oci runtime error: container_linux.go:247: starting container process
116-
caused "process_linux.go:359: container init caused
117-
\"rootfs_linux.go:54: mounting
118-
\\\"/var/lib/replicated-operator/44c648980d1e4b1c5a97167046f32f11/etc/travis/ssl/rabbitmq.cert\\\"
119-
to rootfs
120-
\\\"/var/lib/docker/aufs/mnt/a00833d25e72b761e2a0e72b1015dd2b2f3a32cafd2851ba408b298f73b37d37\\\"
121-
at
122-
\\\"/var/lib/docker/aufs/mnt/a00833d25e72b761e2a0e72b1015dd2b2f3a32cafd2851ba408b298f73b37d37/etc/travis/ssl/rabbitmq.cert\\\"
123-
caused \\\"not a directory\\\"\""
124-
: Are you trying to mount a directory onto a file (or vice-versa)? Check
125-
if the specified host path exists and is the expected type
126-
```
127-
128-
To address this, remove the RabbitMQ cert from `/etc/travis/ssl/`:
129-
130-
```
131-
$ sudo rm -r /etc/travis/ssl/rabbitmq.cert
132-
```
133-
After this, do a full reboot of the system and everything should start again properly.
134-
135104

136105
## View Sidekiq Queue Statistics
137106

138-
In the past there have been reported cases where the system became unresponsive. It took quite a while until jobs where worked off or they weren't picked up at all. We found out that often full Sidekiq queues played a part in this. To get some insight, it helps to retrieve some basics statistics in the Ruby console:
107+
In the past, there have been reported cases where the system became unresponsive. It took quite a while until jobs were worked off, or they weren't picked up at all. We found out that full Sidekiq queues often played a part in this. To get some insight, it helps to retrieve some basic statistics in the Ruby console:
139108

140109
**TCIE 3.x**: `kubectl exec -it [travis-api-pod]j /app/script/console` *on your local machine*
141110

0 commit comments

Comments
 (0)