Elasticsearch Status Fault and Stenographer 90%+ packet loss after 2.4.141 upgrade #14505
-
Version2.4.140 Installation MethodSecurity Onion ISO image Descriptionupgrading Installation TypeStandalone Locationon-prem with Internet access Hardware SpecsExceeds minimum requirements CPU32 RAM128 Storage for /32 TB Storage for /nsm28 TB Network Traffic Collectionspan port Network Traffic SpeedsLess than 1Gbps StatusYes, all services on all nodes are running OK Salt StatusYes, there are salt failures (please provide detail below) LogsYes, there are additional clues in /opt/so/log/ (please provide detail below) DetailSince upgrading to Security Onion 2.4.141, my standalone installation maintains an Elasticsearch Status of Fault and Stenographer consistently has over 90% packet loss. so-status shows all containers are running OK. There is one fault out of 711 that shows when running a high state. /opt/so/log/salt/master shows the following error: Name: /usr/sbin/so-elasticsearch-templates-load - Function: cmd.run - Result: Failed - Started: 14:24:57.622941 - Duration: 20014.427 ms Is anyone else experiencing this problem? Guidelines
|
Beta Was this translation helpful? Give feedback.
Replies: 4 comments 17 replies
-
Do you see errors in the |
Beta Was this translation helpful? Give feedback.
-
I was able to figure out the Elasticsearch Status Fault issue. I had reached the 1000/1000 shard maximum. After cleaning up my indices and shards, the Fault issue has been corrected. However I am still having problems with over 90% Stenographer Packet Loss as well as 30%-40% I/O Wait times. I did not experience any of these prior to the 2.4.141 upgrade. How would I go about troubleshooting this particular matter? |
Beta Was this translation helpful? Give feedback.
-
I am beginning to suspect this issue may be related to #14545, which is a log ingestion problem related to a template failure load. I am experiencing the same error during a high state call. |
Beta Was this translation helpful? Give feedback.
-
I think I finally figured out the Stenographer issue. I believe there was some seasonal internal traffic spikes that I was not accounting for. We'll find out on the next go round lol smh. Thank you for all of your help!! |
Beta Was this translation helpful? Give feedback.
I think I finally figured out the Stenographer issue. I believe there was some seasonal internal traffic spikes that I was not accounting for. We'll find out on the next go round lol smh.
Thank you for all of your help!!