You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We conducted various experiments on the Spider development dataset using different pre-trained and fine-tuned language model (LLM) architectures. Below are the highest results achieved with their respective LLMs. The highest result, <strong>70.57%</strong>, was obtained with the DB-Chat SQL model, compared to <strong>69% </strong>achieved by the Granite 20B code instruction-based model.
23
+
24
+
<imgsrc= "image/benchmark_result.jpg">
25
+
26
+
#### Outcomes:
27
+
1. Enhanced outcomes are evident with the query correction service, as demonstrated by its post-processing accuracy.
28
+
2. Smaller fine-tuned models outperforms some larger ones pretrained models
0 commit comments