publications

(*) denotes equal contribution

2026

  1. arXiv
    Your LLM Agents are Temporally Blind: The Misalignment Between Tool Use Decisions and Human Time Perception
    Yize Cheng*, Arshia Soltani Moakhar*, Chenrui Fan*, Parsa Hosseini, Kazem Faghih, Zahra Sodagar, Wenxiao Wang, and Soheil Feizi
    arXiv preprint, 2026
  2. arXiv
    Failing to Explore: Language Models on Interactive Tasks
    Mahdi JafariRaviz*, Keivan Rezaei*, Arshia Soltani Moakhar*, Zahra Sodagar, Yize Cheng, and Soheil Feizi
    arXiv preprint, 2026

2025

  1. NeurIPS
    Adversarial Paraphrasing: A Universal Attack for Humanizing AI-Generated Text
    Yize Cheng*, Vinu Sankar Sadasivan*, Mehrdad Saberi, Shoumik Saha, and Soheil Feizi
    In Proceedings of the 39th Conference on Neural Information Processing Systems (NeurIPS), 2025
  2. EMNLP
    DyePack: Provably Flagging Test Set Contamination in LLMs Using Backdoors
    Yize Cheng*, Wenxiao Wang*, Mazda Moayeri, and Soheil Feizi
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2025
  3. EMNLP
    Tool Preferences in Agentic LLMs are Unreliable
    Kazem Faghih*, Wenxiao Wang*, Yize Cheng*, Siddhant Bharti, Gaurang Sriramanan, Sriram Balasubramanian, Parsa Hosseini, and Soheil Feizi
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2025
  4. arXiv
    Schoenfeld’s Anatomy of Mathematical Reasoning by Language Models
    Ming Li*, Chenrui Fan*, Yize Cheng*, Soheil Feizi, and Tianyi Zhou
    arXiv preprint, 2025

2023

  1. arXiv
    Attacking by Aligning: Clean-Label Backdoor Attacks on Object Detection
    Yize Cheng*, Wenbin Hu*, and Minhao Cheng
    arXiv preprint, 2023