Sitemap
A list of all the posts and pages found on the site. For you robots out there, there is an XML version available for digesting as well.
Pages
Posts
publications
Lying Through One’s Teeth: A Study on Verbal Leakage Cues Permalink
Published in *Conference on Empirical Methods in Natural Language Processing* (**EMNLP**), 2021
Multi-VQG: Generating Engaging Questions for Multiple Images Permalink
Published in *Conference on Empirical Methods in Natural Language Processing* (**EMNLP**), 2022
Analyzing the Relationship Between Difference and Ratio-Based Fairness Metrics Permalink
Published in *ACM Conference on Fairness, Accountability, and Transparency* (**FAccT**), 2024
COCOLOFA: A Dataset of News Comments with Common Logical Fallacies Written by LLM-Assisted Crowds Permalink
Published in *Conference on Empirical Methods in Natural Language Processing* (**EMNLP**), 2024
Steer LLM Latents for Hallucination Detection Permalink
Published in *International Conference on Machine Learning* (**ICML**), 2025
Position: Challenges and Future Directions of Data-Centric AI Alignment Permalink
Published in *International Conference on Machine Learning* (**ICML**) Position Track, 2025
HalluEntity: Benchmarking and Understanding Entity-Level Hallucination Detection Permalink
Published in *Transactions on Machine Learning Research* (**TMLR**) **J2C Certification**, 2025
Clean First, Align Later: Benchmarking Preference Data Cleaning for Reliable LLM Alignment Permalink
Published in *Advances in Neural Information Processing Systems* (**NeurIPS**) Datasets and Benchmarks Track, 2025
MetaMind: Modeling Human Social Thoughts with Metacognitive Multi-Agent Systems Permalink
Published in *Advances in Neural Information Processing Systems* (**NeurIPS**) **Spotlight**, 2025
Cognition-of-Thought Elicits Social-Aligned Reasoning in Large Language Models Permalink
Published in *Socially Responsible and Trustworthy Foundation Models at NeurIPS 2025*, 2025
Towards Reducible Uncertainty Modeling for Reliable Large Language Model Agents Permalink
Published in *Under Review*, 2026
Simulating and Understanding Deceptive Behaviors in Long-Horizon Interactions Permalink
Published in *The Fourteenth International Conference on Learning Representations*, 2026
LUMINA: Detecting Hallucinations in RAG System with Context–Knowledge Signals Permalink
Published in *The Fourteenth International Conference on Learning Representations*, 2026
How Retrieved Context Shapes Internal Representations in RAG Permalink
Published in *Under Review*, 2026
