Skip to content
This repository was archived by the owner on Nov 8, 2022. It is now read-only.

question: [Q8Bert experiment Setting] #219

@daumkh402

Description

@daumkh402

Hello, I read the Q8Bert paper and have tried to reproduce the experiment results.
But, on some GLUE tasks ( e.g cola, mrpc ), the differences between the fp32 results and quantized ones are much larger than the differences reported in the paper.
I tried sweeping initial learning rate but still the result was still far from the reported results.

image

So, I want to ask you if the experiment on Q8bert was done with default parameters set inside nlp-architect code as below.

image

If not, could you tell me the experiment setting.

Metadata

Metadata

Assignees

Labels

questionFurther information is requested

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions