-
Notifications
You must be signed in to change notification settings - Fork 41
Add submissions for BERT_IdentifyIA_HF and RoBERTa_IdentifyIA_HF detectors #64
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Hi @liamdugan — I’ve followed your instructions and removed all results.json files from my submission. Only the predictions.json and metadata.json files remain unchanged. The PR is now ready for evaluation so the bot can generate the official results.json automatically. Thanks for your guidance! 🚀
…bmissions/BERT_IdentifyIA_HF/metadata.json
…/submissions/BERT_IdentifyIA_HF/predictions.json
…/submissions/RoBERTa_IdentifyIA_HF/metadata.json
…ard/submissions/RoBERTa_IdentifyIA_HF/predictions.json
|
Eval run succeeded! Link to run: link Here are the results of the submission(s): RoBERTa_IdentifyIA_HFRelease date: 2025-02-01 I've committed detailed results of this detector's performance on the test set to this PR. Warning Failed to find threshold values that achieve False Positive Rate(s): (['1%']) on all domains. This submission will not appear in the main leaderboard for those FPR values; it will only be visible within the splits in which the target FPR was achieved. BERT_IdentifyIA_HFRelease date: 2025-02-01 I've committed detailed results of this detector's performance on the test set to this PR. Warning Failed to find threshold values that achieve False Positive Rate(s): (['5%', '1%']) on all domains. This submission will not appear in the main leaderboard for those FPR values; it will only be visible within the splits in which the target FPR was achieved. If all looks well, a maintainer will come by soon to merge this PR and your entry/entries will appear on the leaderboard. If you need to make any changes, feel free to push new commits to this PR. Thanks for submitting to RAID! |
|
Hi @Gwyn9 just to add a bit more context for this evaluation result. It seems like there does not exist a threshold for which your classifiers get 99% accuracy on human-written text across all domains. I suggest investigating the I'm happy to answer any more questions if you have them. |
|
Hi @liamdugan 👋, Models:
Changes:
Submission files:
These submissions follow the same format as previous ones (metadata.json + predictions.json), now corresponding to the V2 detectors on Hugging Face. Thanks again for maintaining the RAID benchmark, looking forward to seeing how these updated models perform compared to the previous V1 results. |
|
It looks like this eval run failed. Please check the workflow logs to see what went wrong, then push a new commit to your PR to rerun the eval. |
|
Hi @liamdugan 👋, I've pushed an update to fix the UTF-8 BOM issue that caused the previous evaluation to fail during the hydrate.py decoding step. The affected files were the metadata.json and predictions.json for both:
They have now been re-saved using UTF-8 (no BOM) encoding and recommitted. Thanks again for your help and for maintaining RAID! 🚀 |
|
It looks like this eval run failed. Please check the workflow logs to see what went wrong, then push a new commit to your PR to rerun the eval. |
|
Hello @Gwyn9 it seems like both the |
|
Hi @liamdugan, Restored full predictions.json files for BERT_IdentifyIA_HF_V2 and RoBERTa_IdentifyIA_HF_V2. Thank you!! |
|
Eval run succeeded! Link to run: link Here are the results of the submission(s): RoBERTa_IdentifyIA_HF_V2Release date: 2025-02-01 I've committed detailed results of this detector's performance on the test set to this PR. On the RAID dataset as a whole (aggregated across all generation models, domains, decoding strategies, repetition penalties, and adversarial attacks), it achieved an AUROC of 79.26 and a TPR of 40.50% at FPR=5% and 19.97% at FPR=1%. BERT_IdentifyIA_HF_V2Release date: 2025-02-01 I've committed detailed results of this detector's performance on the test set to this PR. On the RAID dataset as a whole (aggregated across all generation models, domains, decoding strategies, repetition penalties, and adversarial attacks), it achieved an AUROC of 85.65 and a TPR of 63.22% at FPR=5% and 48.17% at FPR=1%. If all looks well, a maintainer will come by soon to merge this PR and your entry/entries will appear on the leaderboard. If you need to make any changes, feel free to push new commits to this PR. Thanks for submitting to RAID! |
|
Hey @Gwyn9 would you like me to merge this into RAID? |
Hi @liamdugan👋
This PR includes two new submissions for evaluation:
Both detectors were fine-tuned on data set of kaggle (https://www.kaggle.com/datasets/shanegerami/ai-vs-human-text) and include predictions.json and metadata.json files following the leaderboard structure.
Apologies for the earlier submission issues; the folder structure has been fixed according to your guidance.
Thanks again for your help!