Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion data/xml/2022.findings.xml
Original file line number Diff line number Diff line change
Expand Up @@ -10360,7 +10360,7 @@ Faster and Smaller Speech Translation without Quality Compromise</title>
<title>A Multi-Modal Knowledge Graph for Classical <fixed-case>C</fixed-case>hinese Poetry</title>
<author><first>Yuqing</first><last>Li</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author><first>Yuxin</first><last>Zhang</last><affiliation>Renmin University of China</affiliation></author>
<author><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author id="bin-wu"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author><first>Ji-Rong</first><last>Wen</last><affiliation>Renmin University of China</affiliation></author>
<author><first>Ruihua</first><last>Song</last><affiliation>Renmin University of China</affiliation></author>
<author><first>Ting</first><last>Bai</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
Expand Down
2 changes: 1 addition & 1 deletion data/xml/2024.ccl.xml
Original file line number Diff line number Diff line change
Expand Up @@ -513,7 +513,7 @@
<author><first>Jing</first><last>Zhang</last><variant script="hani"><first>京</first><last>张</last></variant></author>
<author><first>Jiangming</first><last>Shu</last><variant script="hani"><first>江明</first><last>舒</last></variant></author>
<author><first>Yuxiang</first><last>Zhang</last><variant script="hani"><first>宇翔</first><last>张</last></variant></author>
<author><first>Bin</first><last>Wu</last><variant script="hani"><first>斌</first><last>吴</last></variant></author>
<author id="bin-wu"><first>Bin</first><last>Wu</last><variant script="hani"><first>斌</first><last>吴</last></variant></author>
<author><first>Wei</first><last>Wang</last><variant script="hani"><first>巍</first><last>王</last></variant></author>
<author><first>Jian</first><last>Yu</last><variant script="hani"><first>剑</first><last>于</last></variant></author>
<author><first>Jitao</first><last>Sang</last><variant script="hani"><first>基韬</first><last>桑</last></variant></author>
Expand Down
4 changes: 2 additions & 2 deletions data/xml/2024.findings.xml
Original file line number Diff line number Diff line change
Expand Up @@ -20597,7 +20597,7 @@
<author><first>Yangfu</first><last>Zhu</last></author>
<author><first>Yuqing</first><last>Li</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0009-0005-8224-300X"><first>Di</first><last>Liu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0002-7112-126X"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0002-7112-126X" id="bin-wu"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<pages>1600-1617</pages>
<abstract>Given the importance of ancient Chinese in capturing the essence of rich historical and cultural heritage, the rapid advancements in Large Language Models (LLMs) necessitate benchmarks that can effectively evaluate their understanding of ancient contexts. To meet this need, we present AC-EVAL, an innovative benchmark designed to assess the advanced knowledge and reasoning capabilities of LLMs within the context of ancient Chinese. AC-EVAL is structured across three levels of difficulty reflecting different facets of language comprehension: general historical knowledge, short text understanding, and long text comprehension. The benchmark comprises 13 tasks, spanning historical facts, geography, social customs, art, philosophy, classical poetry and prose, providing a comprehensive assessment framework. Our extensive evaluation of top-performing LLMs, tailored for both English and Chinese, reveals a substantial potential for enhancing ancient text comprehension. By highlighting the strengths and weaknesses of LLMs, AC-EVAL aims to promote their development and application forward in the realms of ancient Chinese language education and scholarly research.</abstract>
<url hash="28ecec81">2024.findings-emnlp.87</url>
Expand Down Expand Up @@ -21847,7 +21847,7 @@
<author><first>Shuai</first><last>Zhong</last></author>
<author><first>Xinming</first><last>Chen</last></author>
<author><first>Jinsheng</first><last>Qi</last></author>
<author orcid="0000-0002-7112-126X"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0002-7112-126X" id="bin-wu"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<pages>3121-3133</pages>
<abstract>Video Question Answering (VideoQA) tasks require not only correct answers but also visual evidence. The “localize-then-answer” strategy, while enhancing accuracy and interpretability, faces challenges due to the lack of temporal localization labels in VideoQA datasets. Existing methods often train the models’ localization capabilities indirectly using QA labels, leading to inaccurate localization. Moreover, our experiments show that despite high accuracy, current models depend too heavily on language shortcuts or spurious correlations with irrelevant visual context. To address these issues, we propose a Question-Guided and Answer-Calibrated TRansformer (QGAC-TR), which guides and calibrates localization using question and option texts without localization labels. Furthermore, we design two self-supervised learning tasks to further enhance the model’s refined localization capabilities. Extensive experiments on three public datasets focused on temporal and causal reasoning show that our model not only achieves accuracy comparable to large-scale pretrained models but also leads in localization aspects. Code will be available on GitHub.</abstract>
<url hash="fec3535e">2024.findings-emnlp.176</url>
Expand Down
4 changes: 2 additions & 2 deletions data/xml/2025.acl.xml
Original file line number Diff line number Diff line change
Expand Up @@ -2581,7 +2581,7 @@
<author><first>Yuting</first><last>Wei</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author><first>Qi</first><last>Meng</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0009-0003-5039-9937"><first>Yuanxing</first><last>Xu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0002-7112-126X"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0002-7112-126X" id="bin-wu"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<pages>3537-3550</pages>
<abstract>Traditional methods for processing classical Chinese typically segment language understanding into discrete tasks, which overlook crucial background information and reduce user engagement. Large language models (LLMs) provide integrated solutions, yet they entail high computational costs and risks of generating inaccurate historical information. To tackle these challenges, we propose a novel framework, TEACH (conTrastive knowlEdge Adaptive distillation with enhanCed Historical interpretability), which focuses on classical Chinese understanding by integrating word sense disambiguation with sentence translation. This integration leverages a confidence-annotated knowledge base and a step-by-step Chain-of-Thought prompting mechanism to minimize hallucinations and improve semantic analysis. Moreover, TEACH employs contrastive distillation learning to efficiently transfer capabilities from larger models to smaller ones (e.g., Qwen2-1.5B), addressing overly liberal translations. Additionally, we introduce an innovative generation evaluation metric using iterative word alignment, enhancing LLM performance assessments by distinguishing additional information and addressing excessive translation issues. Experiments conducted on real-world datasets validate TEACH’s efficacy in classical Chinese educational scenarios.</abstract>
<url hash="569b4f1a">2025.acl-long.178</url>
Expand Down Expand Up @@ -16063,7 +16063,7 @@
<paper id="1100">
<title>Boosting <fixed-case>LLM</fixed-case>’s Molecular Structure Elucidation with Knowledge Enhanced Tree Search Reasoning</title>
<author orcid="0000-0002-0253-1476"><first>Xiang</first><last>Zhuang</last></author>
<author orcid="0000-0002-8677-2321"><first>Bin</first><last>Wu</last></author>
<author orcid="0000-0002-8677-2321" id="bin-wu-ucl"><first>Bin</first><last>Wu</last></author>
<author><first>Jiyu</first><last>Cui</last></author>
<author><first>Kehua</first><last>Feng</last></author>
<author><first>Xiaotong</first><last>Li</last><affiliation>Zhejiang University</affiliation></author>
Expand Down
2 changes: 1 addition & 1 deletion data/xml/2025.emnlp.xml
Original file line number Diff line number Diff line change
Expand Up @@ -3363,7 +3363,7 @@
<author><first>Zheng</first><last>Wang</last></author>
<author orcid="0009-0002-0330-3982"><first>Yuxuan</first><last>Zhang</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0009-0000-8649-6008"><first>Bo</first><last>Wang</last></author>
<author orcid="0000-0002-7112-126X"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0002-7112-126X" id="bin-wu"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<pages>4501-4520</pages>
<abstract>Recent years have witnessed remarkable advances in Large Language Models (LLMs). However, in the task of social relation recognition, Large Language Models (LLMs) encounter significant challenges due to their reliance on sequential training data, which inherently restricts their capacity to effectively model complex graph-structured relationships. To address this limitation, we propose a novel low-coupling method synergizing multimodal temporal Knowledge Graphs and Large Language Models (mtKG-LLM) for social relation reasoning. Specifically, we extract multimodal information from the videos and model the social networks as spatial Knowledge Graphs (KGs) for each scene. Temporal KGs are constructed based on spatial KGs and updated along the timeline for long-term reasoning. Subsequently, we retrieve multi-scale information from the graph-structured knowledge for LLMs to recognize the underlying social relation. Extensive experiments demonstrate that our method has achieved state-of-the-art performance in social relation recognition. Furthermore, our framework exhibits effectiveness in bridging the gap between KGs and LLMs. Our code will be released after acceptance.</abstract>
<url hash="caa05ace">2025.emnlp-main.224</url>
Expand Down
4 changes: 2 additions & 2 deletions data/xml/2025.findings.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22488,7 +22488,7 @@
</paper>
<paper id="1149">
<title>A Joint Optimization Framework for Enhancing Efficiency of Tool Utilization in <fixed-case>LLM</fixed-case> Agents</title>
<author orcid="0000-0002-8677-2321"><first>Bin</first><last>Wu</last></author>
<author orcid="0000-0002-8677-2321" id="bin-wu-ucl"><first>Bin</first><last>Wu</last></author>
<author orcid="0000-0003-0516-3688"><first>Edgar</first><last>Meij</last><affiliation>Bloomberg</affiliation></author>
<author><first>Emine</first><last>Yilmaz</last></author>
<pages>22361-22373</pages>
Expand Down Expand Up @@ -29933,7 +29933,7 @@
<author orcid="0009-0002-0330-3982"><first>Yuxuan</first><last>Zhang</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0003-2880-9517"><first>Yangfu</first><last>Zhu</last></author>
<author orcid="0009-0006-7918-3349"><first>Haorui</first><last>Wang</last></author>
<author orcid="0000-0002-7112-126X"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<author orcid="0000-0002-7112-126X" id="bin-wu"><first>Bin</first><last>Wu</last><affiliation>Beijing University of Posts and Telecommunications</affiliation></author>
<pages>5174-5184</pages>
<abstract>Social relationship recognition, as one of the fundamental tasks in video understanding, contributes to the construction and application of multi-modal knowledge graph. Previous works have mainly focused on two aspects: generating character graphs and multi-modal fusion. However, they often overlook the impact of cultural differences on relationship recognition. Specifically, relationship recognition models are susceptible to being misled by training data from a specific cultural context. This can result in the learning of culture-specific spurious correlations, ultimately restricting the ability to recognize social relationships in different cultures. Therefore, we employ a customized causal graph to analyze the confounding effects of culture in the relationship recognition task. We propose a Cultural Causal Intervention (CCI) model that mitigates the influence of culture as a confounding factor in the visual and textual modalities. Importantly, we also construct a novel video social relation recognition (CVSR) dataset to facilitate discussion and research on cultural factors in video tasks. Extensive experiments conducted on several datasets demonstrate that the proposed model surpasses state-of-the-art methods.</abstract>
<url hash="c2802f3e">2025.findings-emnlp.277</url>
Expand Down
2 changes: 1 addition & 1 deletion data/xml/2025.naacl.xml
Original file line number Diff line number Diff line change
Expand Up @@ -3234,7 +3234,7 @@
<title>Entropy-Based Decoding for Retrieval-Augmented Large Language Models</title>
<author><first>Zexuan</first><last>Qiu</last><affiliation>The Chinese University of Hong Kong</affiliation></author>
<author><first>Zijing</first><last>Ou</last><affiliation>Imperial College London</affiliation></author>
<author orcid="0000-0002-8677-2321"><first>Bin</first><last>Wu</last></author>
<author orcid="0000-0002-8677-2321" id="bin-wu-ucl"><first>Bin</first><last>Wu</last></author>
<author><first>Jingjing</first><last>Li</last></author>
<author><first>Aiwei</first><last>Liu</last><affiliation>Tsinghua University</affiliation></author>
<author orcid="0000-0001-8106-6447"><first>Irwin</first><last>King</last></author>
Expand Down
8 changes: 8 additions & 0 deletions data/yaml/name_variants.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -1158,6 +1158,14 @@
- canonical: {first: Susan E., last: Brennan}
variants:
- {first: Susan, last: Brennan}
- canonical: {first: Bin, last: Wu}
id: bin-wu-ucl
orcid: 0000-0002-8677-2321
institution: University College London
comment: UCL
- canonical: {first: Bin, last: Wu}
id: bin-wu
comment: May refer to several people
- canonical: {first: Xavier, last: Briffault}
id: xavier-briffault
- canonical: {first: Ted, last: Briscoe}
Expand Down