publications

(*) denotes equal contribution

2025

  1. arXiv
    Adversarial Paraphrasing: A Universal Attack for Humanizing AI-Generated Text
    Yize Cheng*, Vinu Sankar Sadasivan*, Mehrdad Saberi, Shoumik Saha, and Soheil Feizi
    arXiv preprint, 2025
  2. EMNLP
    DyePack: Provably Flagging Test Set Contamination in LLMs Using Backdoors
    Yize Cheng*, Wenxiao Wang*, Mazda Moayeri, and Soheil Feizi
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2025
  3. EMNLP
    Gaming Tool Preferences in Agentic LLMs
    Kazem Faghih*, Wenxiao Wang*, Yize Cheng*, Siddhant Bharti, Gaurang Sriramanan, Sriram Balasubramanian, Parsa Hosseini, and Soheil Feizi
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2025

2023

  1. arXiv
    Attacking by Aligning: Clean-Label Backdoor Attacks on Object Detection
    Yize Cheng*, Wenbin Hu*, and Minhao Cheng
    arXiv preprint, 2023