@@ -33,7 +33,6 @@ def initialize(request_client:)
3333 # Use either transcript_ids or input_text as input into LeMUR.
3434 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
3535 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
36- # Defaults to "default".
3736 # @param max_output_size [Integer] Max output size in tokens, up to 4000
3837 # @param temperature [Float] The temperature to use for the model.
3938 # Higher values result in answers that are more creative, lower values are more
@@ -84,7 +83,6 @@ def task(prompt:, transcript_ids: nil, input_text: nil, context: nil, final_mode
8483 # Use either transcript_ids or input_text as input into LeMUR.
8584 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
8685 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
87- # Defaults to "default".
8886 # @param max_output_size [Integer] Max output size in tokens, up to 4000
8987 # @param temperature [Float] The temperature to use for the model.
9088 # Higher values result in answers that are more creative, lower values are more
@@ -136,7 +134,6 @@ def summary(transcript_ids: nil, input_text: nil, context: nil, final_model: nil
136134 # Use either transcript_ids or input_text as input into LeMUR.
137135 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
138136 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
139- # Defaults to "default".
140137 # @param max_output_size [Integer] Max output size in tokens, up to 4000
141138 # @param temperature [Float] The temperature to use for the model.
142139 # Higher values result in answers that are more creative, lower values are more
@@ -187,7 +184,6 @@ def question_answer(questions:, transcript_ids: nil, input_text: nil, context: n
187184 # Use either transcript_ids or input_text as input into LeMUR.
188185 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
189186 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
190- # Defaults to "default".
191187 # @param max_output_size [Integer] Max output size in tokens, up to 4000
192188 # @param temperature [Float] The temperature to use for the model.
193189 # Higher values result in answers that are more creative, lower values are more
@@ -294,7 +290,6 @@ def initialize(request_client:)
294290 # Use either transcript_ids or input_text as input into LeMUR.
295291 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
296292 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
297- # Defaults to "default".
298293 # @param max_output_size [Integer] Max output size in tokens, up to 4000
299294 # @param temperature [Float] The temperature to use for the model.
300295 # Higher values result in answers that are more creative, lower values are more
@@ -347,7 +342,6 @@ def task(prompt:, transcript_ids: nil, input_text: nil, context: nil, final_mode
347342 # Use either transcript_ids or input_text as input into LeMUR.
348343 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
349344 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
350- # Defaults to "default".
351345 # @param max_output_size [Integer] Max output size in tokens, up to 4000
352346 # @param temperature [Float] The temperature to use for the model.
353347 # Higher values result in answers that are more creative, lower values are more
@@ -401,7 +395,6 @@ def summary(transcript_ids: nil, input_text: nil, context: nil, final_model: nil
401395 # Use either transcript_ids or input_text as input into LeMUR.
402396 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
403397 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
404- # Defaults to "default".
405398 # @param max_output_size [Integer] Max output size in tokens, up to 4000
406399 # @param temperature [Float] The temperature to use for the model.
407400 # Higher values result in answers that are more creative, lower values are more
@@ -454,7 +447,6 @@ def question_answer(questions:, transcript_ids: nil, input_text: nil, context: n
454447 # Use either transcript_ids or input_text as input into LeMUR.
455448 # @param context [String, Hash{String => Object}] Context to provide the model. This can be a string or a free-form JSON value.
456449 # @param final_model [AssemblyAI::Lemur::LemurModel] The model that is used for the final prompt after compression is performed.
457- # Defaults to "default".
458450 # @param max_output_size [Integer] Max output size in tokens, up to 4000
459451 # @param temperature [Float] The temperature to use for the model.
460452 # Higher values result in answers that are more creative, lower values are more
0 commit comments