Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 7 additions & 7 deletions opencompass/models/openai_api.py
Original file line number Diff line number Diff line change
Expand Up @@ -484,12 +484,15 @@ def _preprocess_messages(
if input_len > max_seq_len:
raise ValueError(
f'Input length ({input_len}) exceeds max_seq_len '
f'({max_seq_len}) and mode is set to "none". Please '
f'either change the mode or reduce the input length.')
f'({max_seq_len}) and mode is set to "none". Please '
f'either change the mode or increase the max_seq_len.')

# Trim input if needed
def bin_trim_wrapper(text):
return self._bin_trim(text, max_seq_len - 100 - max_out_len, mode)
trim_length = max_seq_len - 100
if max_out_len is not None:
trim_length -= max_out_len
return self._bin_trim(text, trim_length, mode)

if isinstance(input, str) and mode != 'none':
input = bin_trim_wrapper(input)
Expand All @@ -512,7 +515,7 @@ def bin_trim_wrapper(text):
messages.append(msg)

# Adjust max_out_len
try:
if max_out_len is not None:
original_max_out_len = max_out_len
max_out_len = min(
max_out_len,
Expand All @@ -529,9 +532,6 @@ def bin_trim_wrapper(text):
f'max_out_len was truncated from {original_max_out_len} '
f'to {max_out_len} due to input length')

except KeyError:
max_out_len = max_out_len

return messages, max_out_len


Expand Down
Loading