publications

(*) denotes equal contribution

2025

  1. arXiv
    Adversarial Paraphrasing: A Universal Attack for Humanizing AI-Generated Text
    Yize Cheng*, Vinu Sankar Sadasivan*, Mehrdad Saberi, Shoumik Saha, and Soheil Feizi
    arXiv preprint, 2025
  2. arXiv
    DyePack: Provably Flagging Test Set Contamination in LLMs Using Backdoors
    Yize Cheng*, Wenxiao Wang*, Mazda Moayeri, and Soheil Feizi
    arXiv preprint, 2025
  3. arXiv
    Gaming Tool Preferences in Agentic LLMs
    Kazem Faghih*, Wenxiao Wang*, Yize Cheng*, Siddhant Bharti, Gaurang Sriramanan, Sriram Balasubramanian, Parsa Hosseini, and Soheil Feizi
    arXiv preprint, 2025

2023

  1. arXiv
    Attacking by Aligning: Clean-Label Backdoor Attacks on Object Detection
    Yize Cheng*, Wenbin Hu*, and Minhao Cheng
    arXiv preprint, 2023