-
Notifications
You must be signed in to change notification settings - Fork 935
Ignore mgmt_ds_lock error in loganalyzer due to expected behavior #21856
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ignore mgmt_ds_lock error in loganalyzer due to expected behavior #21856
Conversation
Ignore expected FRR backend datastore lock contention during initialization - backends retry indefinitely until successful Signed-off-by: markxiao <markxiao@arista.com>
|
/azp run |
|
Azure Pipelines successfully started running 1 pipeline(s). |
…rror-in-loganalyzer
|
/azp run |
|
Azure Pipelines successfully started running 1 pipeline(s). |
StormLiangMS
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If this is by design, should FRR downgrade the log level to "notice" instead of "error"?
As we're moving from separate mode to unified mode for config initialization, I expect the race condition between staticd and zebra will be resolved once the transition is complete.
For now, I think it is ok to ignore this error logs.
@lipxu could you help to take a look?
|
@markx-arista PR conflicts with 202511 branch |
…nic-net#21856) Ignore expected FRR backend datastore lock contention during initialization - backends retry indefinitely until successful Signed-off-by: markxiao <markxiao@arista.com> Signed-off-by: Priyansh Tratiya <ptratiya@microsoft.com>
…nic-net#21856) Ignore expected FRR backend datastore lock contention during initialization - backends retry indefinitely until successful Signed-off-by: markxiao <markxiao@arista.com> Signed-off-by: Andoni Sanguesa <andoniesanguesa@gmail.com>
…nic-net#21856) Ignore expected FRR backend datastore lock contention during initialization - backends retry indefinitely until successful Signed-off-by: markxiao <markxiao@arista.com> Signed-off-by: Andoni Sanguesa <andoniesanguesa@gmail.com>
…nic-net#21856) Ignore expected FRR backend datastore lock contention during initialization - backends retry indefinitely until successful Signed-off-by: markxiao <markxiao@arista.com> Signed-off-by: Saravanan Sellappa <saravanan@nexthop.ai>
|
@markx-arista could you help to resolve 202511 conflict issue? |
Cherry-pick to 202511: #22099 |
Description of PR
Summary:
Ignore expected FRR backend datastore lock contention during initialization - backends retry indefinitely until successful
Backend daemons like zebra and staticd both try to acquire the lock during initialization, and zebra starts first, then staticd retries every 50ms until zebra finishes initialization and release the lock.
Zebra takes time to initialize on some SKUs, and causes the staticd to retry. However, lock contention and retry mechanism is by design. This log is expected and it only triggers loganalyzer error because its log level is ERR.
sonic-buildimage issue: sonic-net/sonic-buildimage#24966
Fixes # (issue)
Type of change
Back port request
Approach
What is the motivation for this PR?
test_bgp_router_id.py failed due to loganalyzer during bgp service restart
How did you do it?
Ignore the expected log
How did you verify/test it?
test_bgp_router_id.py passed
Any platform specific information?
We only saw the problem on Arista 7050CX3-32c because zebra takes a little bit longer time to initialize on it. Could happen to other SKUs.
Supported testbed topology if it's a new test case?
Documentation