Skip to content

Commit e397ce2

Browse files
committed
Updated Yamux impl.,added tests for yamux and mplex
1 parent 286752c commit e397ce2

File tree

5 files changed

+459
-68
lines changed

5 files changed

+459
-68
lines changed

libp2p/stream_muxer/mplex/mplex_stream.py

Lines changed: 53 additions & 48 deletions
Original file line numberDiff line numberDiff line change
@@ -46,8 +46,9 @@ class MplexStream(IMuxedStream):
4646
read_deadline: int | None
4747
write_deadline: int | None
4848

49-
# TODO: Add lock for read/write to avoid interleaving receiving messages?
5049
close_lock: trio.Lock
50+
read_lock: trio.Lock
51+
write_lock: trio.Lock
5152

5253
# NOTE: `dataIn` is size of 8 in Go implementation.
5354
incoming_data_channel: "trio.MemoryReceiveChannel[bytes]"
@@ -80,6 +81,8 @@ def __init__(
8081
self.event_remote_closed = trio.Event()
8182
self.event_reset = trio.Event()
8283
self.close_lock = trio.Lock()
84+
self.read_lock = trio.Lock()
85+
self.write_lock = trio.Lock()
8386
self.incoming_data_channel = incoming_data_channel
8487
self._buf = bytearray()
8588

@@ -113,63 +116,65 @@ async def read(self, n: int | None = None) -> bytes:
113116
:param n: number of bytes to read
114117
:return: bytes actually read
115118
"""
116-
if n is not None and n < 0:
117-
raise ValueError(
118-
"the number of bytes to read `n` must be non-negative or "
119-
f"`None` to indicate read until EOF, got n={n}"
120-
)
121-
if self.event_reset.is_set():
122-
raise MplexStreamReset
123-
if n is None:
124-
return await self._read_until_eof()
125-
if len(self._buf) == 0:
126-
data: bytes
127-
# Peek whether there is data available. If yes, we just read until there is
128-
# no data, then return.
129-
try:
130-
data = self.incoming_data_channel.receive_nowait()
131-
self._buf.extend(data)
132-
except trio.EndOfChannel:
133-
raise MplexStreamEOF
134-
except trio.WouldBlock:
135-
# We know `receive` will be blocked here. Wait for data here with
136-
# `receive` and catch all kinds of errors here.
119+
async with self.read_lock:
120+
if n is not None and n < 0:
121+
raise ValueError(
122+
"the number of bytes to read `n` must be non-negative or "
123+
f"`None` to indicate read until EOF, got n={n}"
124+
)
125+
if self.event_reset.is_set():
126+
raise MplexStreamReset
127+
if n is None:
128+
return await self._read_until_eof()
129+
if len(self._buf) == 0:
130+
data: bytes
131+
# Peek whether there is data available. If yes, we just read until
132+
# there is no data, then return.
137133
try:
138-
data = await self.incoming_data_channel.receive()
134+
data = self.incoming_data_channel.receive_nowait()
139135
self._buf.extend(data)
140136
except trio.EndOfChannel:
141-
if self.event_reset.is_set():
142-
raise MplexStreamReset
143-
if self.event_remote_closed.is_set():
144-
raise MplexStreamEOF
145-
except trio.ClosedResourceError as error:
146-
# Probably `incoming_data_channel` is closed in `reset` when we are
147-
# waiting for `receive`.
148-
if self.event_reset.is_set():
149-
raise MplexStreamReset
150-
raise Exception(
151-
"`incoming_data_channel` is closed but stream is not reset. "
152-
"This should never happen."
153-
) from error
154-
self._buf.extend(self._read_return_when_blocked())
155-
payload = self._buf[:n]
156-
self._buf = self._buf[len(payload) :]
157-
return bytes(payload)
137+
raise MplexStreamEOF
138+
except trio.WouldBlock:
139+
# We know `receive` will be blocked here. Wait for data here with
140+
# `receive` and catch all kinds of errors here.
141+
try:
142+
data = await self.incoming_data_channel.receive()
143+
self._buf.extend(data)
144+
except trio.EndOfChannel:
145+
if self.event_reset.is_set():
146+
raise MplexStreamReset
147+
if self.event_remote_closed.is_set():
148+
raise MplexStreamEOF
149+
except trio.ClosedResourceError as error:
150+
# Probably `incoming_data_channel` is closed in `reset` when
151+
# we are waiting for `receive`.
152+
if self.event_reset.is_set():
153+
raise MplexStreamReset
154+
raise Exception(
155+
"`incoming_data_channel` is closed but stream is not reset."
156+
"This should never happen."
157+
) from error
158+
self._buf.extend(self._read_return_when_blocked())
159+
payload = self._buf[:n]
160+
self._buf = self._buf[len(payload) :]
161+
return bytes(payload)
158162

159163
async def write(self, data: bytes) -> None:
160164
"""
161165
Write to stream.
162166
163167
:return: number of bytes written
164168
"""
165-
if self.event_local_closed.is_set():
166-
raise MplexStreamClosed(f"cannot write to closed stream: data={data!r}")
167-
flag = (
168-
HeaderTags.MessageInitiator
169-
if self.is_initiator
170-
else HeaderTags.MessageReceiver
171-
)
172-
await self.muxed_conn.send_message(flag, data, self.stream_id)
169+
async with self.write_lock:
170+
if self.event_local_closed.is_set():
171+
raise MplexStreamClosed(f"cannot write to closed stream: data={data!r}")
172+
flag = (
173+
HeaderTags.MessageInitiator
174+
if self.is_initiator
175+
else HeaderTags.MessageReceiver
176+
)
177+
await self.muxed_conn.send_message(flag, data, self.stream_id)
173178

174179
async def close(self) -> None:
175180
"""

libp2p/stream_muxer/yamux/yamux.py

Lines changed: 74 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -77,6 +77,8 @@ def __init__(self, stream_id: int, conn: "Yamux", is_initiator: bool) -> None:
7777
self.send_window = DEFAULT_WINDOW_SIZE
7878
self.recv_window = DEFAULT_WINDOW_SIZE
7979
self.window_lock = trio.Lock()
80+
self.read_lock = trio.Lock()
81+
self.write_lock = trio.Lock()
8082

8183
async def __aenter__(self) -> "YamuxStream":
8284
"""Enter the async context manager."""
@@ -98,16 +100,32 @@ async def write(self, data: bytes) -> None:
98100
# Flow control: Check if we have enough send window
99101
total_len = len(data)
100102
sent = 0
101-
103+
logging.debug(f"Stream {self.stream_id}: Starts writing {total_len} bytes ")
102104
while sent < total_len:
105+
# Wait for available window with timeout
106+
timeout = False
103107
async with self.window_lock:
104-
# Wait for available window
105-
while self.send_window == 0 and not self.closed:
106-
# Release lock while waiting
108+
if self.send_window == 0:
109+
logging.debug(
110+
f"Stream {self.stream_id}: Window is zero, waiting for update"
111+
)
112+
# Release lock and wait with timeout
107113
self.window_lock.release()
108-
await trio.sleep(0.01)
114+
# To avoid re-acquiring the lock immediately,
115+
with trio.move_on_after(5.0) as cancel_scope:
116+
while self.send_window == 0 and not self.closed:
117+
await trio.sleep(0.01)
118+
# If we timed out, cancel the scope
119+
timeout = cancel_scope.cancelled_caught
120+
# Re-acquire lock
109121
await self.window_lock.acquire()
110122

123+
# If we timed out waiting for window update, raise an error
124+
if timeout:
125+
raise MuxedStreamError(
126+
"Timed out waiting for window update after 5 seconds."
127+
)
128+
111129
if self.closed:
112130
raise MuxedStreamError("Stream is closed")
113131

@@ -123,25 +141,53 @@ async def write(self, data: bytes) -> None:
123141
await self.conn.secured_conn.write(header + chunk)
124142
sent += to_send
125143

126-
# If window is getting low, consider updating
127-
if self.send_window < DEFAULT_WINDOW_SIZE // 2:
128-
await self.send_window_update()
129-
130-
async def send_window_update(self, increment: int | None = None) -> None:
131-
"""Send a window update to peer."""
144+
async def send_window_update(
145+
self, increment: int | None, skip_lock: bool = False
146+
) -> None:
147+
"""
148+
Send a window update to peer.
149+
150+
param:increment: The amount to increment the window size by.
151+
If None, uses the difference between DEFAULT_WINDOW_SIZE
152+
and current receive window.
153+
param:skip_lock (bool): If True, skips acquiring window_lock.
154+
This should only be used when calling from a context
155+
that already holds the lock.
156+
"""
157+
increment_value = 0
132158
if increment is None:
133-
increment = DEFAULT_WINDOW_SIZE - self.recv_window
134-
135-
if increment <= 0:
159+
increment_value = DEFAULT_WINDOW_SIZE - self.recv_window
160+
else:
161+
increment_value = increment
162+
if increment_value <= 0:
163+
# If increment is zero or negative, skip sending update
164+
logging.debug(
165+
f"Stream {self.stream_id}: Skipping window update"
166+
f"(increment={increment})"
167+
)
136168
return
169+
logging.debug(
170+
f"Stream {self.stream_id}: Sending window update with increment={increment}"
171+
)
137172

138-
async with self.window_lock:
139-
self.recv_window += increment
173+
async def _do_window_update() -> None:
174+
self.recv_window += increment_value
140175
header = struct.pack(
141-
YAMUX_HEADER_FORMAT, 0, TYPE_WINDOW_UPDATE, 0, self.stream_id, increment
176+
YAMUX_HEADER_FORMAT,
177+
0,
178+
TYPE_WINDOW_UPDATE,
179+
0,
180+
self.stream_id,
181+
increment_value,
142182
)
143183
await self.conn.secured_conn.write(header)
144184

185+
if skip_lock:
186+
await _do_window_update()
187+
else:
188+
async with self.window_lock:
189+
await _do_window_update()
190+
145191
async def read(self, n: int | None = -1) -> bytes:
146192
# Handle None value for n by converting it to -1
147193
if n is None:
@@ -198,11 +244,19 @@ async def read(self, n: int | None = -1) -> bytes:
198244
# Return all buffered data
199245
data = bytes(buffer)
200246
buffer.clear()
201-
logging.debug(f"Stream {self.stream_id}: Returning {len(data)} bytes")
202247
return data
203248

204-
# For specific size read (n > 0), return available data immediately
205-
return await self.conn.read_stream(self.stream_id, n)
249+
data = await self.conn.read_stream(self.stream_id, n)
250+
async with self.window_lock:
251+
self.recv_window -= len(data)
252+
# Automatically send a window update if recv_window is low
253+
if self.recv_window <= DEFAULT_WINDOW_SIZE // 2:
254+
logging.debug(
255+
f"Stream {self.stream_id}: "
256+
f"Low recv_window ({self.recv_window}), sending update"
257+
)
258+
await self.send_window_update(None, skip_lock=True)
259+
return data
206260

207261
async def close(self) -> None:
208262
if not self.send_closed:

newsfragments/639.feature.rst

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
Added separate read and write locks to the `MplexStream` & `YamuxStream` class.This ensures thread-safe access and data integrity when multiple coroutines interact with the same MplexStream instance.
Lines changed: 124 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,124 @@
1+
import pytest
2+
import trio
3+
4+
from libp2p.abc import ISecureConn
5+
from libp2p.crypto.keys import PrivateKey, PublicKey
6+
from libp2p.peer.id import ID
7+
from libp2p.stream_muxer.mplex.constants import (
8+
HeaderTags,
9+
)
10+
from libp2p.stream_muxer.mplex.datastructures import (
11+
StreamID,
12+
)
13+
from libp2p.stream_muxer.mplex.mplex import (
14+
Mplex,
15+
)
16+
from libp2p.stream_muxer.mplex.mplex_stream import (
17+
MplexStream,
18+
)
19+
20+
21+
class DummySecureConn(ISecureConn):
22+
"""A minimal implementation of ISecureConn for testing."""
23+
24+
async def write(self, data: bytes) -> None:
25+
pass
26+
27+
async def read(self, n: int | None = -1) -> bytes:
28+
return b""
29+
30+
async def close(self) -> None:
31+
pass
32+
33+
def get_remote_address(self) -> tuple[str, int] | None:
34+
return None
35+
36+
def get_local_peer(self) -> ID:
37+
return ID(b"local")
38+
39+
def get_local_private_key(self) -> PrivateKey:
40+
return PrivateKey() # Dummy key for testing
41+
42+
def get_remote_peer(self) -> ID:
43+
return ID(b"remote")
44+
45+
def get_remote_public_key(self) -> PublicKey:
46+
return PublicKey() # Dummy key for testing
47+
48+
49+
class DummyMuxedConn(Mplex):
50+
"""A minimal mock of Mplex for testing read/write locks."""
51+
52+
def __init__(self) -> None:
53+
self.secured_conn = DummySecureConn()
54+
self.peer_id = ID(b"dummy")
55+
self.streams = {}
56+
self.streams_lock = trio.Lock()
57+
self.event_shutting_down = trio.Event()
58+
self.event_closed = trio.Event()
59+
self.event_started = trio.Event()
60+
self.stream_backlog_limit = 256
61+
self.stream_backlog_semaphore = trio.Semaphore(256)
62+
channels = trio.open_memory_channel[MplexStream](0)
63+
self.new_stream_send_channel, self.new_stream_receive_channel = channels
64+
65+
async def send_message(
66+
self, flag: HeaderTags, data: bytes, stream_id: StreamID
67+
) -> None:
68+
await trio.sleep(0.01)
69+
70+
71+
@pytest.mark.trio
72+
async def test_concurrent_writes_are_serialized():
73+
stream_id = StreamID(1, True)
74+
send_log = []
75+
76+
class LoggingMuxedConn(DummyMuxedConn):
77+
async def send_message(
78+
self, flag: HeaderTags, data: bytes, stream_id: StreamID
79+
) -> None:
80+
send_log.append(data)
81+
await trio.sleep(0.01)
82+
83+
memory_send, memory_recv = trio.open_memory_channel(8)
84+
stream = MplexStream(
85+
name="test",
86+
stream_id=stream_id,
87+
muxed_conn=LoggingMuxedConn(),
88+
incoming_data_channel=memory_recv,
89+
)
90+
91+
async def writer(data):
92+
await stream.write(data)
93+
94+
async with trio.open_nursery() as nursery:
95+
for i in range(5):
96+
nursery.start_soon(writer, f"msg-{i}".encode())
97+
# Order doesn't matter due to concurrent execution
98+
assert sorted(send_log) == sorted([f"msg-{i}".encode() for i in range(5)])
99+
100+
101+
@pytest.mark.trio
102+
async def test_concurrent_reads_are_serialized():
103+
stream_id = StreamID(2, True)
104+
muxed_conn = DummyMuxedConn()
105+
memory_send, memory_recv = trio.open_memory_channel(8)
106+
results = []
107+
stream = MplexStream(
108+
name="test",
109+
stream_id=stream_id,
110+
muxed_conn=muxed_conn,
111+
incoming_data_channel=memory_recv,
112+
)
113+
for i in range(5):
114+
await memory_send.send(f"data-{i}".encode())
115+
await memory_send.aclose()
116+
117+
async def reader():
118+
data = await stream.read(6)
119+
results.append(data)
120+
121+
async with trio.open_nursery() as nursery:
122+
for _ in range(5):
123+
nursery.start_soon(reader)
124+
assert sorted(results) == [f"data-{i}".encode() for i in range(5)]

0 commit comments

Comments
 (0)