-
Notifications
You must be signed in to change notification settings - Fork 1.1k
PYTHON-5089 Convert test.test_mongos_load_balancing to async #2107
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 7 commits
ff0d5dc
690dcf7
988663f
4b8f936
5cd8be8
499fe42
41d2aab
3c7c2cb
5cd7d02
8b3f1a9
b0c878c
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,221 @@ | ||
# Copyright 2015-present MongoDB, Inc. | ||
# | ||
# Licensed under the Apache License, Version 2.0 (the "License"); | ||
# you may not use this file except in compliance with the License. | ||
# You may obtain a copy of the License at | ||
# | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# | ||
# Unless required by applicable law or agreed to in writing, software | ||
# distributed under the License is distributed on an "AS IS" BASIS, | ||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
# See the License for the specific language governing permissions and | ||
# limitations under the License. | ||
|
||
"""Test AsyncMongoClient's mongos load balancing using a mock.""" | ||
from __future__ import annotations | ||
|
||
import asyncio | ||
import sys | ||
import threading | ||
|
||
from pymongo.operations import _Op | ||
|
||
sys.path[0:0] = [""] | ||
|
||
from test.asynchronous import AsyncMockClientTest, async_client_context, connected, unittest | ||
from test.asynchronous.pymongo_mocks import AsyncMockClient | ||
from test.utils import async_wait_until | ||
|
||
from pymongo.errors import AutoReconnect, InvalidOperation | ||
from pymongo.server_selectors import writable_server_selector | ||
from pymongo.topology_description import TOPOLOGY_TYPE | ||
|
||
_IS_SYNC = False | ||
|
||
if _IS_SYNC: | ||
PARENT = threading.Thread | ||
else: | ||
PARENT = object | ||
|
||
|
||
class SimpleOp(PARENT): | ||
def __init__(self, client): | ||
super().__init__() | ||
self.client = client | ||
self.passed = False | ||
self.task = None | ||
|
||
async def run(self): | ||
await self.client.db.command("ping") | ||
self.passed = True # No exception raised. | ||
|
||
def start(self): | ||
if _IS_SYNC: | ||
super().start() | ||
else: | ||
self.task = asyncio.create_task(self.run()) | ||
|
||
async def join(self): | ||
if _IS_SYNC: | ||
super().join() | ||
else: | ||
await self.task | ||
|
||
|
||
async def do_simple_op(client, ntasks): | ||
tasks = [SimpleOp(client) for _ in range(ntasks)] | ||
for t in tasks: | ||
t.start() | ||
|
||
for t in tasks: | ||
await t.join() | ||
|
||
for t in tasks: | ||
assert t.passed | ||
|
||
|
||
async def writable_addresses(topology): | ||
return { | ||
server.description.address | ||
for server in await topology.select_servers(writable_server_selector, _Op.TEST) | ||
} | ||
|
||
|
||
class TestMongosLoadBalancing(AsyncMockClientTest): | ||
@async_client_context.require_connection | ||
@async_client_context.require_no_load_balancer | ||
async def asyncSetUp(self): | ||
await super().asyncSetUp() | ||
|
||
def mock_client(self, **kwargs): | ||
mock_client = AsyncMockClient( | ||
standalones=[], | ||
members=[], | ||
mongoses=["a:1", "b:2", "c:3"], | ||
host="a:1,b:2,c:3", | ||
connect=False, | ||
**kwargs, | ||
) | ||
self.addAsyncCleanup(mock_client.aclose) | ||
|
||
# Latencies in seconds. | ||
mock_client.mock_rtts["a:1"] = 0.020 | ||
mock_client.mock_rtts["b:2"] = 0.025 | ||
mock_client.mock_rtts["c:3"] = 0.045 | ||
return mock_client | ||
|
||
async def test_lazy_connect(self): | ||
# While connected() ensures we can trigger connection from the main | ||
# thread and wait for the monitors, this test triggers connection from | ||
# several threads at once to check for data races. | ||
nthreads = 10 | ||
client = self.mock_client() | ||
self.assertEqual(0, len(client.nodes)) | ||
|
||
# Trigger initial connection. | ||
await do_simple_op(client, nthreads) | ||
await async_wait_until(lambda: len(client.nodes) == 3, "connect to all mongoses") | ||
|
||
async def test_failover(self): | ||
nthreads = 10 | ||
client = await connected(self.mock_client(localThresholdMS=0.001)) | ||
await async_wait_until(lambda: len(client.nodes) == 3, "connect to all mongoses") | ||
|
||
# Our chosen mongos goes down. | ||
client.kill_host("a:1") | ||
|
||
# Trigger failover to higher-latency nodes. AutoReconnect should be | ||
# raised at most once in each thread. | ||
passed = [] | ||
|
||
async def f(): | ||
try: | ||
await client.db.command("ping") | ||
except AutoReconnect: | ||
# Second attempt succeeds. | ||
await client.db.command("ping") | ||
|
||
passed.append(True) | ||
|
||
if _IS_SYNC: | ||
threads = [threading.Thread(target=f) for _ in range(nthreads)] | ||
for t in threads: | ||
t.start() | ||
|
||
for t in threads: | ||
t.join() | ||
else: | ||
tasks = [asyncio.create_task(f()) for _ in range(nthreads)] | ||
for t in tasks: | ||
await t | ||
|
||
self.assertEqual(nthreads, len(passed)) | ||
|
||
# Down host removed from list. | ||
self.assertEqual(2, len(client.nodes)) | ||
|
||
async def test_local_threshold(self): | ||
client = await connected(self.mock_client(localThresholdMS=30)) | ||
self.assertEqual(30, client.options.local_threshold_ms) | ||
await async_wait_until(lambda: len(client.nodes) == 3, "connect to all mongoses") | ||
topology = client._topology | ||
|
||
# All are within a 30-ms latency window, see self.mock_client(). | ||
self.assertEqual({("a", 1), ("b", 2), ("c", 3)}, await writable_addresses(topology)) | ||
|
||
# No error | ||
await client.admin.command("ping") | ||
|
||
client = await connected(self.mock_client(localThresholdMS=0)) | ||
self.assertEqual(0, client.options.local_threshold_ms) | ||
# No error | ||
await client.db.command("ping") | ||
# Our chosen mongos goes down. | ||
client.kill_host("{}:{}".format(*next(iter(client.nodes)))) | ||
try: | ||
await client.db.command("ping") | ||
except: | ||
pass | ||
|
||
# We eventually connect to a new mongos. | ||
async def connect_to_new_mongos(): | ||
try: | ||
return await client.db.command("ping") | ||
except AutoReconnect: | ||
pass | ||
|
||
await async_wait_until(connect_to_new_mongos, "connect to a new mongos") | ||
|
||
async def test_load_balancing(self): | ||
# Although the server selection JSON tests already prove that | ||
# select_servers works for sharded topologies, here we do an end-to-end | ||
# test of discovering servers' round trip times and configuring | ||
# localThresholdMS. | ||
client = await connected(self.mock_client()) | ||
await async_wait_until(lambda: len(client.nodes) == 3, "connect to all mongoses") | ||
|
||
# Prohibited for topology type Sharded. | ||
with self.assertRaises(InvalidOperation): | ||
await client.address | ||
|
||
topology = client._topology | ||
self.assertEqual(TOPOLOGY_TYPE.Sharded, topology.description.topology_type) | ||
|
||
# a and b are within the 15-ms latency window, see self.mock_client(). | ||
self.assertEqual({("a", 1), ("b", 2)}, await writable_addresses(topology)) | ||
|
||
client.mock_rtts["a:1"] = 0.045 | ||
|
||
# Discover only b is within latency window. | ||
async def predicate(): | ||
return {("b", 2)} == await writable_addresses(topology) | ||
|
||
await async_wait_until( | ||
predicate, | ||
'discover server "a" is too far', | ||
) | ||
|
||
|
||
if __name__ == "__main__": | ||
unittest.main() |
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -15,6 +15,7 @@ | |
"""Test MongoClient's mongos load balancing using a mock.""" | ||
from __future__ import annotations | ||
|
||
import asyncio | ||
import sys | ||
import threading | ||
|
||
|
@@ -30,33 +31,47 @@ | |
from pymongo.server_selectors import writable_server_selector | ||
from pymongo.topology_description import TOPOLOGY_TYPE | ||
|
||
_IS_SYNC = True | ||
|
||
@client_context.require_connection | ||
@client_context.require_no_load_balancer | ||
def setUpModule(): | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. We can keep these to improve performance when skipping this test suite. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I moved it to the class's There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Oh whoops my bad, forgot how the wrappers interact with async. You're right! There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. All good, wasn't sure if I was missing / forgetting what the trick was! |
||
pass | ||
if _IS_SYNC: | ||
PARENT = threading.Thread | ||
else: | ||
PARENT = object | ||
|
||
|
||
class SimpleOp(threading.Thread): | ||
class SimpleOp(PARENT): | ||
def __init__(self, client): | ||
super().__init__() | ||
self.client = client | ||
self.passed = False | ||
self.task = None | ||
|
||
def run(self): | ||
self.client.db.command("ping") | ||
self.passed = True # No exception raised. | ||
|
||
def start(self): | ||
if _IS_SYNC: | ||
super().start() | ||
else: | ||
self.task = asyncio.create_task(self.run()) | ||
|
||
def do_simple_op(client, nthreads): | ||
threads = [SimpleOp(client) for _ in range(nthreads)] | ||
for t in threads: | ||
def join(self): | ||
if _IS_SYNC: | ||
super().join() | ||
else: | ||
self.task | ||
|
||
|
||
def do_simple_op(client, ntasks): | ||
tasks = [SimpleOp(client) for _ in range(ntasks)] | ||
for t in tasks: | ||
t.start() | ||
|
||
for t in threads: | ||
for t in tasks: | ||
t.join() | ||
|
||
for t in threads: | ||
for t in tasks: | ||
assert t.passed | ||
|
||
|
||
|
@@ -68,6 +83,11 @@ def writable_addresses(topology): | |
|
||
|
||
class TestMongosLoadBalancing(MockClientTest): | ||
@client_context.require_connection | ||
@client_context.require_no_load_balancer | ||
def setUp(self): | ||
super().setUp() | ||
|
||
def mock_client(self, **kwargs): | ||
mock_client = MockClient( | ||
standalones=[], | ||
|
@@ -118,12 +138,17 @@ def f(): | |
|
||
passed.append(True) | ||
|
||
threads = [threading.Thread(target=f) for _ in range(nthreads)] | ||
for t in threads: | ||
t.start() | ||
if _IS_SYNC: | ||
threads = [threading.Thread(target=f) for _ in range(nthreads)] | ||
for t in threads: | ||
t.start() | ||
|
||
for t in threads: | ||
t.join() | ||
for t in threads: | ||
t.join() | ||
else: | ||
tasks = [asyncio.create_task(f()) for _ in range(nthreads)] | ||
for t in tasks: | ||
t | ||
|
||
self.assertEqual(nthreads, len(passed)) | ||
|
||
|
@@ -183,8 +208,11 @@ def test_load_balancing(self): | |
client.mock_rtts["a:1"] = 0.045 | ||
|
||
# Discover only b is within latency window. | ||
def predicate(): | ||
return {("b", 2)} == writable_addresses(topology) | ||
|
||
wait_until( | ||
lambda: {("b", 2)} == writable_addresses(topology), | ||
predicate, | ||
'discover server "a" is too far', | ||
) | ||
|
||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we use
ConcurrentRunner
for these as well?