@@ -173,15 +173,15 @@ PYBIND11_MODULE(_llm_runner, m) {
173173 float temperature,
174174 int32_t num_bos,
175175 int32_t num_eos) {
176- GenerationConfig cfg;
177- cfg.echo = echo;
178- cfg.max_new_tokens = max_new_tokens;
179- cfg.warming = warming;
180- cfg.seq_len = seq_len;
181- cfg.temperature = temperature;
182- cfg.num_bos = num_bos;
183- cfg.num_eos = num_eos;
184- return cfg;
176+ GenerationConfig cfg;
177+ cfg.echo = echo;
178+ cfg.max_new_tokens = max_new_tokens;
179+ cfg.warming = warming;
180+ cfg.seq_len = seq_len;
181+ cfg.temperature = temperature;
182+ cfg.num_bos = num_bos;
183+ cfg.num_eos = num_eos;
184+ return cfg;
185185 }),
186186 py::arg (" echo" ) = true ,
187187 py::arg (" max_new_tokens" ) = -1 ,
@@ -204,12 +204,12 @@ PYBIND11_MODULE(_llm_runner, m) {
204204 py::arg (" num_prompt_tokens" ),
205205 " Resolve the maximum number of new tokens to generate based on constraints" )
206206 .def (" __repr__" , [](const GenerationConfig& config) {
207- return " <GenerationConfig max_new_tokens=" +
208- std::to_string (config.max_new_tokens ) +
209- " seq_len=" + std::to_string (config.seq_len ) +
210- " temperature=" + std::to_string (config.temperature ) +
211- " echo=" + (config.echo ? " True" : " False" ) +
212- " warming=" + (config.warming ? " True" : " False" ) + " >" ;
207+ return " <GenerationConfig max_new_tokens=" +
208+ std::to_string (config.max_new_tokens ) +
209+ " seq_len=" + std::to_string (config.seq_len ) +
210+ " temperature=" + std::to_string (config.temperature ) +
211+ " echo=" + (config.echo ? " True" : " False" ) +
212+ " warming=" + (config.warming ? " True" : " False" ) + " >" ;
213213 });
214214
215215 // Bind Stats
0 commit comments