Skip to content

Support for AWQ quantization (W4A16) for GPTNeoXForCausalLM architecture #1987

@Priya95715

Description

@Priya95715

Hi,

Does llm-compressor supports AWQ (W4A16) quantization for GPTNeoXForCausalLM architecture?

Thanks

Metadata

Metadata

Assignees

No one assigned

    Labels

    awqFor any issue / PR related to AWQ support

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions