We are currently using the OpenAI gpt-oss-120b parameters model. It offers good inference and a very fast throughput. However, more performant models or better performing models tend to converge to a higher quality annotation faster, making both the faster throughput and the cost even less than the gpt-oss-120b.
Since fine-tuning open-source models is not currently a priority, we want to explore how other models perform.