@@ -553,6 +553,7 @@ mod tests {
553553 use std:: sync:: Arc ;
554554
555555 /// Helper function to create a test RecordBatch with 1000 rows of (int, string) data
556+ /// Example batch_id 1 -> 0..1000, 2 -> 1001..2000
556557 fn create_test_record_batch ( batch_id : i32 ) -> Result < RecordBatch > {
557558 assert ! ( batch_id > 0 , "batch_id must be greater than 0" ) ;
558559 let num_rows = batch_id * 1000 ;
@@ -577,6 +578,7 @@ mod tests {
577578
578579 #[ tokio:: test]
579580 #[ cfg( feature = "hdfs-opendal" ) ]
581+ #[ ignore = "This test requires a running HDFS cluster" ]
580582 async fn test_write_to_hdfs_sync ( ) -> Result < ( ) > {
581583 use opendal:: services:: Hdfs ;
582584 use opendal:: Operator ;
@@ -627,6 +629,7 @@ mod tests {
627629
628630 #[ tokio:: test]
629631 #[ cfg( feature = "hdfs-opendal" ) ]
632+ #[ ignore = "This test requires a running HDFS cluster" ]
630633 async fn test_write_to_hdfs_streaming ( ) -> Result < ( ) > {
631634 use opendal:: services:: Hdfs ;
632635 use opendal:: Operator ;
@@ -707,6 +710,7 @@ mod tests {
707710
708711 #[ tokio:: test]
709712 #[ cfg( feature = "hdfs-opendal" ) ]
713+ #[ ignore = "This test requires a running HDFS cluster" ]
710714 async fn test_parquet_writer_streaming ( ) -> Result < ( ) > {
711715 // Configure output path
712716 let output_path = "/user/test_parquet_writer_streaming/data.parquet" ;
@@ -755,6 +759,7 @@ mod tests {
755759
756760 #[ tokio:: test]
757761 #[ cfg( feature = "hdfs-opendal" ) ]
762+ #[ ignore = "This test requires a running HDFS cluster" ]
758763 async fn test_parquet_writer_exec_with_memory_input ( ) -> Result < ( ) > {
759764 use datafusion:: datasource:: memory:: MemorySourceConfig ;
760765 use datafusion:: datasource:: source:: DataSourceExec ;
0 commit comments