@@ -1596,7 +1596,7 @@ def test_iter_batches_dataframe(fail_collect):
1596
1596
assert got .equals (exp )
1597
1597
1598
1598
1599
- def test_arrow_c_stream_to_table (fail_collect ):
1599
+ def test_arrow_c_stream_to_table_and_reader (fail_collect ):
1600
1600
ctx = SessionContext ()
1601
1601
1602
1602
# Create a DataFrame with two separate record batches
@@ -1613,6 +1613,12 @@ def test_arrow_c_stream_to_table(fail_collect):
1613
1613
assert table .schema == df .schema ()
1614
1614
assert table .column ("a" ).num_chunks == 2
1615
1615
1616
+ reader = pa .RecordBatchReader ._import_from_c_capsule (df .__arrow_c_stream__ ())
1617
+ assert isinstance (reader , pa .RecordBatchReader )
1618
+ reader_table = pa .Table .from_batches (reader )
1619
+ expected = pa .Table .from_batches ([batch1 , batch2 ])
1620
+ assert reader_table .equals (expected )
1621
+
1616
1622
1617
1623
def test_arrow_c_stream_order ():
1618
1624
ctx = SessionContext ()
@@ -1631,14 +1637,6 @@ def test_arrow_c_stream_order():
1631
1637
assert col .chunk (1 )[0 ].as_py () == 2
1632
1638
1633
1639
1634
- def test_arrow_c_stream_reader (df ):
1635
- reader = pa .RecordBatchReader ._import_from_c_capsule (df .__arrow_c_stream__ ())
1636
- assert isinstance (reader , pa .RecordBatchReader )
1637
- table = pa .Table .from_batches (reader )
1638
- expected = pa .Table .from_batches (df .collect ())
1639
- assert table .equals (expected )
1640
-
1641
-
1642
1640
def test_arrow_c_stream_schema_selection (fail_collect ):
1643
1641
ctx = SessionContext ()
1644
1642
0 commit comments