@@ -101,7 +101,7 @@ impl<'a, B: BackendForChannel<MC>, MC: MerkleChannel> CommitmentSchemeProver<'a,
101101 pub fn build_weights_hash_map (
102102 & self ,
103103 sampled_points : & TreeVec < ColumnVec < Vec < CirclePoint < SecureField > > > > ,
104- max_log_size : u32 ,
104+ lifting_log_size : u32 ,
105105 ) -> WeightsHashMap < B >
106106 where
107107 Col < B , SecureField > : Send + Sync ,
@@ -123,17 +123,17 @@ impl<'a, B: BackendForChannel<MC>, MC: MerkleChannel> CommitmentSchemeProver<'a,
123123 let log_size = poly. evals . domain . log_size ( ) ;
124124 // For each sample point, compute the weights needed to evaluate the polynomial at
125125 // the folded sample point.
126- // TODO(Leo): the computation `point.repeated_double(max_log_size - log_size)` is
127- // likely repeated a bunch of times in a typical flat air. Consider moving it
128- // outside the loop.
126+ // TODO(Leo): the computation `point.repeated_double(lifting_log_size - log_size)`
127+ // is likely repeated a bunch of times in a typical flat air.
128+ // Consider moving it outside the loop.
129129 #[ cfg( not( feature = "parallel" ) ) ]
130130 points. iter ( ) . for_each ( |& point| {
131- compute_weights ( ( log_size, point. repeated_double ( max_log_size - log_size) ) )
131+ compute_weights ( ( log_size, point. repeated_double ( lifting_log_size - log_size) ) )
132132 } ) ;
133133
134134 #[ cfg( feature = "parallel" ) ]
135135 points. par_iter ( ) . for_each ( |& point| {
136- compute_weights ( ( log_size, point. repeated_double ( max_log_size - log_size) ) )
136+ compute_weights ( ( log_size, point. repeated_double ( lifting_log_size - log_size) ) )
137137 } ) ;
138138 } ) ;
139139
@@ -153,11 +153,11 @@ impl<'a, B: BackendForChannel<MC>, MC: MerkleChannel> CommitmentSchemeProver<'a,
153153 )
154154 . entered ( ) ;
155155
156- let max_log_size = self . trees . last ( ) . unwrap ( ) . commitment . layers . len ( ) as u32 - 1 ;
156+ let lifting_log_size = self . trees . last ( ) . unwrap ( ) . commitment . layers . len ( ) as u32 - 1 ;
157157 let weights_hash_map = if self . store_polynomials_coefficients {
158158 None
159159 } else {
160- Some ( self . build_weights_hash_map ( & sampled_points, max_log_size ) )
160+ Some ( self . build_weights_hash_map ( & sampled_points, lifting_log_size ) )
161161 } ;
162162 let samples: TreeVec < Vec < Vec < PointSample > > > = self
163163 . polynomials ( )
@@ -168,7 +168,7 @@ impl<'a, B: BackendForChannel<MC>, MC: MerkleChannel> CommitmentSchemeProver<'a,
168168 . map ( |& point| PointSample {
169169 point,
170170 value : poly. eval_at_point (
171- point. repeated_double ( max_log_size - poly. evals . domain . log_size ( ) ) ,
171+ point. repeated_double ( lifting_log_size - poly. evals . domain . log_size ( ) ) ,
172172 weights_hash_map. as_ref ( ) ,
173173 ) ,
174174 } )
@@ -209,7 +209,7 @@ impl<'a, B: BackendForChannel<MC>, MC: MerkleChannel> CommitmentSchemeProver<'a,
209209 // Build the query position tree.
210210 let preprocessed_query_positions = prepare_preprocessed_query_positions (
211211 & query_positions,
212- max_log_size ,
212+ lifting_log_size ,
213213 self . trees [ 0 ] . commitment . layers . len ( ) as u32 - 1 ,
214214 ) ;
215215 let query_positions_tree = TreeVec :: new (
0 commit comments