Skip to content

Commit 9328bd2

Browse files
Merge pull request #53 from harshunr/patch-2
Updating publication.bib
2 parents 9c7c0ab + 5c9402f commit 9328bd2

File tree

1 file changed

+11
-0
lines changed

1 file changed

+11
-0
lines changed

publications.bib

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,14 @@
1+
@inproceedings{
2+
harshita-2025-theAchemyOfThought,
3+
title={The Alchemy of Thought: Understanding In-Context Learning Through Supervised Classification},
4+
author={Narnoli, Harshita and Surdeanu, Mihai},
5+
booktitle={Findings of the International Joint Conference on Natural Language Processing {\&} Asia-Pacific Chapter of the Association for Computational Linguistics 2025},
6+
year={2025},
7+
address = "Mumbai, India",
8+
url={https://arxiv.org/abs/2601.01290},
9+
abstract="In-context learning (ICL) has become a prominent paradigm to rapidly customize LLMs to new tasks without fine-tuning. However, despite the empirical evidence of its usefulness, we still do not truly understand how ICL works. In this paper, we compare the behavior of in-context learning with supervised classifiers trained on ICL demonstrations to investigate three research questions: (1) Do LLMs with ICL behave similarly to classifiers trained on the same examples? (2) If so, which classifiers are closer, those based on gradient descent (GD) or those based on k-nearest neighbors (kNN)? (3) When they do not behave similarly, what conditions are associated with differences in behavior? Using text classification as a use case, with six datasets and three LLMs, we observe that LLMs behave similarly to these classifiers when the relevance of demonstrations is high. On average, ICL is closer to kNN than logistic regression, giving empirical evidence that the attention mechanism behaves more similarly to kNN than GD. However, when demonstration relevance is low, LLMs perform better than these classifiers, likely because LLMs can back off to their parametric memory, a luxury these classifiers do not have."
10+
}
11+
112
@inproceedings{rahimi-etal-2025-relation,
213
title = "Relation-Aware Prompting Makes Large Language Models Effective Zero-shot Relation Extractors",
314
author = "Rahimi, Mahdi and

0 commit comments

Comments
 (0)