Skip to content

Challenge on the training details of Phoenix  #50

@jacklanda

Description

@jacklanda

One word for all

As reported in the technical report, the bs: 256 is seemly a large-scale batch size with 2048 max sequence length. I wonder what hardware environment used for fine-tuning Phoenix-7B? In detailed, how many A100-80GB used?

图片

Thanks!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions