-
Notifications
You must be signed in to change notification settings - Fork 166
[release-4.16] OCPBUGS-63155: Fix EgressIP stale GARP post reboot + pod restart #2809
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: release-4.16
Are you sure you want to change the base?
Conversation
Currently, we are force exiting with the trap before the background processes can end, container is removed and the orphaned processes end early causing our config to go into an unknown state because we dont end in an orderly manner. Wait until the pid file for ovnkube controller with node is removed which shows the process has completed. Signed-off-by: Martin Kennelly <[email protected]> (cherry picked from commit 8b29419) (cherry picked from commit d65ec5c) (cherry picked from commit d3ae338) (cherry picked from commit 7057948) (cherry picked from commit e86fca4)
Prevent ovn-controller from sending stale GARP by adding drop flows on external bridge patch ports until ovnkube-controller synchronizes the southbound database - henceforth known as "drop flows". This addresses race conditions where ovn-controller processes outdated SB DB state before ovnkube-controller updates it, particularly affecting EIP SNAT configurations attached to logical router ports. Fixes: https://issues.redhat.com/browse/FDP-1537 ovnkube-controller controls the lifecycle of the drop flows. ovs / ovn-controller running is required to configure external bridge. Downstream, the external bridge maybe precreated and ovn-controller will use this. This fix considers three primary scenarios: node, container and pod restart. On Node restart means the ovs flows installed priotior to reboot on the node are cleared but the external bridge exists. Add the flows before ovnkube controller with node starts. The reason to add it here is that our gateway code depends on ovn-controller started and running... There is now a race here between ovn-controller starting (and garping) before we set this flow but I think the risk is low however it needs serious testing. The reason I did not naturally at the drop flows before ovn-controller started is because I have no way to detect if its a node reboot or pod reboot and i dont want to inject drop flows for simple ovn-controller container restart which could disrupt traffic. ovnkube-controller starts, we create a new gateway and apply flows the same flows in-order to ensure we always drop GARP when ovnkube controller hasn't sync. Remove the flows when ovnkube-controller has syncd. There is also a race here between ovnkube-controller removing the flows and ovn-controller GARPing with stale SB DB info. There is no easy way to detect what SB DB data ovn-controller has consumed. On Pod restart, we add the drop flows before exit. ovnkube-controller-with-node will also add it before it starts the go code. Container restart: - ovnkube-controller: adds flows upon start and exit - ovn-controller: no changes While the drop flows are set, OVN may not be able to resolve IPs it doesn't know about in its Logical Router pipelines generation. Following removal of the drop flows, OVN may resolve the IPs using GARP requests. OVN-Controller always sends out GARPs with op code 1 on startup. Conflicts: Within allowNodeIPGARPFlows & dropGARPFlows, removed the iteration over multiple networks and only generate the config for the default network. Exported private DefaultNodeNetworkController field gateway in-order to access gateway methods SetDefaultBridgeGARPDropFlows & Reconcile. Within Start for type nodeNetworkControllerManager, return if error found when starting the nadController: - err = ncm.nadController.Start() + if err = ncm.nadController.Start(); err != nil { + return fmt.Errorf("failed to start NAD controller: %v", err) + } - return err Signed-off-by: Martin Kennelly <[email protected]> (cherry picked from commit 82fc3bf) (cherry picked from commit 50a94e1) (cherry picked from commit 37dd4e2) (cherry picked from commit 5b53803) (cherry picked from commit 7da434d)
PR 5373 to drop the GARP flows didnt consider that we set the default network controller and later we set the gateway obj. In-between this period, ovnkube node may receive a stop signal and we do not guard against accessing the gateway if its not yet set. OVNKube controller may have sync'd before the gateway obj is set. There is nothing to reconcile if the gateway is not set. Conflict: Needed to cast defaultNodeNetworkController from interface BaseNetworkController to the concrete type DefaultNodeNetworkController. We can be sure its only set to this type because its set in one location in func initDefaultNodeNetworkController. Signed-off-by: Martin Kennelly <[email protected]> (cherry picked from commit e60220a) (cherry picked from commit a7869b2) (cherry picked from commit 2ac68e4) (cherry picked from commit 3b039fe) (cherry picked from commit 547aaca)
Ensure ovn-controller has processed the SB DB updates before removing the GARP drop flows by utilizing the hv_cfg field in NB_Global [1] OVNKube controller increments the nb_cfg value post sync, which is copied to SB DB by northd. OVN-Controllers copy this nb_cfg value from SB DB and write it to their chassis_private tables nb_cfg field after they have processed the SB DB changes. Northd will then look at all the chassis_private tables nb_cfg value and set the NB DBs Nb_global hv_cfg value to the min integer found. Since IC currently only supports one node per zone, we can be sure ovn-controller is running locally and therefore its ok to block removing the drop GARP flows. [1] https://man7.org/linux/man-pages/man5/ovn-nb.5.html Signed-off-by: Martin Kennelly <[email protected]> (cherry picked from commit 3b5da01) (cherry picked from commit a4776fb) (cherry picked from commit f7c67b7) (cherry picked from commit 2396130) (cherry picked from commit f22003c)
|
@martinkennelly: This pull request references Jira Issue OCPBUGS-63155, which is invalid:
Comment The bug has been updated to refer to the pull request using the external bug tracker. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the openshift-eng/jira-lifecycle-plugin repository. |
|
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: martinkennelly The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
|
/test unit Unit test for ovspinning |
|
/test 4.16-upgrade-from-stable-4.15-local-gateway-e2e-aws-ovn-upgrade Slight disruption increase over the limit (+3s). Unrelated to this PR. |
|
/verified by 'pre-merge testing' |
|
/retest |
|
@jechen0648: This PR has been marked as verified by In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the openshift-eng/jira-lifecycle-plugin repository. |
|
/retest |
|
@martinkennelly: The following tests failed, say
Full PR test history. Your PR dashboard. Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. I understand the commands that are listed here. |
/hold
Depends on #2807
Conflicts documented in each commit.
Requires Surya ack + QE premerge ack.