@@ -33,7 +33,7 @@ use vortex::error::VortexResult;
3333use vortex:: file:: { LayoutContext , LayoutDeserializer , VortexFileWriter , VortexReadBuilder } ;
3434use vortex:: sampling_compressor:: compressors:: fsst:: FSSTCompressor ;
3535use vortex:: sampling_compressor:: { SamplingCompressor , ALL_ENCODINGS_CONTEXT } ;
36- use vortex:: { Array , ArrayDType , IntoArray , IntoCanonical } ;
36+ use vortex:: { ArrayDType , ArrayData , IntoArrayData , IntoCanonical } ;
3737
3838use crate :: tokio_runtime:: TOKIO_RUNTIME ;
3939
@@ -100,7 +100,7 @@ fn parquet_decompress_read(buf: bytes::Bytes) -> usize {
100100 nbytes
101101}
102102
103- fn parquet_compressed_written_size ( array : & Array , compression : Compression ) -> usize {
103+ fn parquet_compressed_written_size ( array : & ArrayData , compression : Compression ) -> usize {
104104 let chunked = ChunkedArray :: try_from ( array) . unwrap ( ) ;
105105 let ( batches, schema) = chunked_to_vec_record_batch ( chunked) ;
106106 parquet_compress_write ( batches, schema, compression, & mut Vec :: new ( ) )
@@ -109,10 +109,10 @@ fn parquet_compressed_written_size(array: &Array, compression: Compression) -> u
109109fn vortex_compress_write (
110110 runtime : & Runtime ,
111111 compressor : & SamplingCompressor < ' _ > ,
112- array : & Array ,
112+ array : & ArrayData ,
113113 buf : & mut Vec < u8 > ,
114114) -> VortexResult < u64 > {
115- async fn async_write ( array : & Array , cursor : & mut Cursor < & mut Vec < u8 > > ) -> VortexResult < ( ) > {
115+ async fn async_write ( array : & ArrayData , cursor : & mut Cursor < & mut Vec < u8 > > ) -> VortexResult < ( ) > {
116116 let mut writer = VortexFileWriter :: new ( cursor) ;
117117
118118 writer = writer. write_array_columns ( array. clone ( ) ) . await ?;
@@ -129,7 +129,7 @@ fn vortex_compress_write(
129129}
130130
131131fn vortex_decompress_read ( runtime : & Runtime , buf : Buffer ) -> VortexResult < ArrayRef > {
132- async fn async_read ( buf : Buffer ) -> VortexResult < Array > {
132+ async fn async_read ( buf : Buffer ) -> VortexResult < ArrayData > {
133133 let builder: VortexReadBuilder < _ > = VortexReadBuilder :: new (
134134 buf,
135135 LayoutDeserializer :: new (
@@ -140,7 +140,7 @@ fn vortex_decompress_read(runtime: &Runtime, buf: Buffer) -> VortexResult<ArrayR
140140
141141 let stream = builder. build ( ) . await ?;
142142 let dtype = stream. dtype ( ) . clone ( ) ;
143- let vecs: Vec < Array > = stream. try_collect ( ) . await ?;
143+ let vecs: Vec < ArrayData > = stream. try_collect ( ) . await ?;
144144
145145 ChunkedArray :: try_new ( vecs, dtype) . map ( |e| e. into ( ) )
146146 }
@@ -154,7 +154,7 @@ fn vortex_decompress_read(runtime: &Runtime, buf: Buffer) -> VortexResult<ArrayR
154154fn vortex_compressed_written_size (
155155 runtime : & Runtime ,
156156 compressor : & SamplingCompressor < ' _ > ,
157- array : & Array ,
157+ array : & ArrayData ,
158158) -> VortexResult < u64 > {
159159 vortex_compress_write ( runtime, compressor, array, & mut Vec :: new ( ) )
160160}
@@ -168,7 +168,7 @@ fn benchmark_compress<F, U>(
168168 bench_name : & str ,
169169) where
170170 F : Fn ( ) -> U ,
171- U : AsRef < Array > ,
171+ U : AsRef < ArrayData > ,
172172{
173173 // if no logging is enabled, enable it
174174 if !LOG_INITIALIZED . swap ( true , Ordering :: SeqCst ) {
0 commit comments