Skip to content

There is no guidance on how to provision more max output tokens #154

@Yuanqi-babe

Description

@Yuanqi-babe

Describe the bug
We're hitting output token limit, the response was cut and incomplete.

Please complete the following information:

  • Which API you used: [e.g. /chat/completions]
  • Which model you used: [e.g. anthropic.claude-3-sonnet-20240229-v1:0]

To Reproduce
Steps to reproduce the behavior. If possible, please share an example request.

Expected behavior
I need guidance as to how to increment the max output tokens.

Screenshots
If applicable, add screenshots to help explain your problem (please DO NOT include sensitive information).

Additional context
Add any other context about the problem here.

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions