default search action
Yi Zeng 0005
Person information
- affiliation: Virginia Tech, Blacksburg, VA, USA
- affiliation: Xidian University, State Key Laboratory of Integrated Service Networks, Xi'an, China
Other persons with the same name
- Yi Zeng — disambiguation page
- Yi Zeng 0001 — Chinese Academy of Sciences, Research Center for Brain-inspired Intelligence, Institute of Automation, Beijing, China (and 1 more)
- Yi Zeng 0002 — ScienceBeijing Forestry University, College of Information Science, China
- Yi Zeng 0003 — Southwest University, Faculty of Computer and Information Science, Chongqing, China
- Yi Zeng 0004 — Harbin Institute of Technology, Department of Control Science and Engineering, Harbin, China (and 1 more)
- Yi Zeng 0006 — Tencent, Youtu Lab, Shenzhen, China (and 1 more)
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
Journal Articles
- 2024
- [j4]Han Qiu, Yi Zeng, Qinkai Zheng, Shangwei Guo, Tianwei Zhang, Hewu Li:
An Efficient Preprocessing-Based Approach to Mitigate Advanced Adversarial Attacks. IEEE Trans. Computers 73(3): 645-655 (2024) - 2023
- [j3]Si Chen, Yi Zeng, Won Park, Jiachen T. Wang, Xun Chen, Lingjuan Lyu, Zhuoqing Mao, Ruoxi Jia:
Turning a Curse into a Blessing: Enabling In-Distribution-Data-Free Backdoor Removal via Stabilized Model Inversion. Trans. Mach. Learn. Res. 2023 (2023) - 2020
- [j2]Wenting Wei, Huaxi Gu, Achille Pattavina, Jiru Wang, Yi Zeng:
Optimizing energy and spectrum efficiency of virtual optical network embedding in elastic optical networks. Opt. Switch. Netw. 37: 100568 (2020) - 2019
- [j1]Yi Zeng, Huaxi Gu, Wenting Wei, Yantao Guo:
$Deep-Full-Range$ : A Deep Learning Based Network Encrypted Traffic Classification and Intrusion Detection Framework. IEEE Access 7: 45182-45190 (2019)
Conference and Workshop Papers
- 2024
- [c22]Yi Zeng, Hongpeng Lin, Jingwen Zhang, Diyi Yang, Ruoxi Jia, Weiyan Shi:
How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs. ACL (1) 2024: 14322-14350 - [c21]Xiangyu Qi, Yi Zeng, Tinghao Xie, Pin-Yu Chen, Ruoxi Jia, Prateek Mittal, Peter Henderson:
Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To! ICLR 2024 - [c20]Shayne Longpre, Sayash Kapoor, Kevin Klyman, Ashwin Ramaswami, Rishi Bommasani, Borhane Blili-Hamelin, Yangsibo Huang, Aviya Skowron, Zheng Xin Yong, Suhas Kotha, Yi Zeng, Weiyan Shi, Xianjun Yang, Reid Southen, Alexander Robey, Patrick Chao, Diyi Yang, Ruoxi Jia, Daniel Kang, Sandy Pentland, Arvind Narayanan, Percy Liang, Peter Henderson:
Position: A Safe Harbor for AI Evaluation and Red Teaming. ICML 2024 - [c19]Zhuowen Yuan, Zidi Xiong, Yi Zeng, Ning Yu, Ruoxi Jia, Dawn Song, Bo Li:
RigorLLM: Resilient Guardrails for Large Language Models against Undesired Content. ICML 2024 - 2023
- [c18]Yi Zeng, Minzhou Pan, Hoang Anh Just, Lingjuan Lyu, Meikang Qiu, Ruoxi Jia:
Narcissus: A Practical Clean-Label Backdoor Attack with Limited Information. CCS 2023: 771-785 - [c17]Hoang Anh Just, Feiyang Kang, Tianhao Wang, Yi Zeng, Myeongseob Ko, Ming Jin, Ruoxi Jia:
LAVA: Data Valuation without Pre-Specified Learning Algorithms. ICLR 2023 - [c16]Yi Zeng, Zhouxing Shi, Ming Jin, Feiyang Kang, Lingjuan Lyu, Cho-Jui Hsieh, Ruoxi Jia:
Towards Robustness Certification Against Universal Perturbations. ICLR 2023 - [c15]Junyuan Hong, Yi Zeng, Shuyang Yu, Lingjuan Lyu, Ruoxi Jia, Jiayu Zhou:
Revisiting Data-Free Knowledge Distillation with Poisoned Teachers. ICML 2023: 13199-13212 - [c14]Zhenting Wang, Chen Chen, Yi Zeng, Lingjuan Lyu, Shiqing Ma:
Where Did I Come From? Origin Attribution of AI-Generated Images. NeurIPS 2023 - [c13]Yi Zeng, Minzhou Pan, Himanshu Jahagirdar, Ming Jin, Lingjuan Lyu, Ruoxi Jia:
Meta-Sift: How to Sift Out a Clean Subset in the Presence of Data Poisoning? USENIX Security Symposium 2023: 1667-1684 - [c12]Minzhou Pan, Yi Zeng, Lingjuan Lyu, Xue Lin, Ruoxi Jia:
ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning Paradigms. USENIX Security Symposium 2023: 2725-2742 - 2022
- [c11]Yi Zeng, Si Chen, Won Park, Zhuoqing Mao, Ming Jin, Ruoxi Jia:
Adversarial Unlearning of Backdoors via Implicit Hypergradient. ICLR 2022 - [c10]Xuanli He, Qiongkai Xu, Yi Zeng, Lingjuan Lyu, Fangzhao Wu, Jiwei Li, Ruoxi Jia:
CATER: Intellectual Property Protection on Text Generation APIs via Conditional Watermarks. NeurIPS 2022 - 2021
- [c9]Han Qiu, Yi Zeng, Shangwei Guo, Tianwei Zhang, Meikang Qiu, Bhavani Thuraisingham:
DeepSweep: An Evaluation Framework for Mitigating DNN Backdoor Attacks using Data Augmentation. AsiaCCS 2021: 363-377 - [c8]Yi Zeng, Won Park, Z. Morley Mao, Ruoxi Jia:
Rethinking the Backdoor Attacks' Triggers: A Frequency Perspective. ICCV 2021: 16453-16461 - [c7]Shangwei Guo, Tianwei Zhang, Han Qiu, Yi Zeng, Tao Xiang, Yang Liu:
Fine-tuning Is Not Enough: A Simple yet Effective Watermark Removal Attack for DNN Models. IJCAI 2021: 3635-3641 - 2020
- [c6]Yi Zeng, Han Qiu, Gérard Memmi, Meikang Qiu:
A Data Augmentation-Based Defense Method Against Adversarial Attacks in Neural Networks. ICA3PP (2) 2020: 274-289 - 2019
- [c5]Yi Zeng, Meikang Qiu, Dan Zhu, Zhihao Xue, Jian Xiong, Meiqin Liu:
DeepVCM: A Deep Learning Based Intrusion Detection Method in VANET. BigDataSecurity/HPSC/IDS 2019: 288-293 - [c4]Yi Zeng, Meikang Qiu, Jingqi Niu, Yanxin Long, Jian Xiong, Meiqin Liu:
V-PSC: A Perturbation-Based Causative Attack Against DL Classifiers' Supply Chain in VANET. CSE/EUC 2019: 86-91 - [c3]Wenting Wei, Huaxi Gu, Achille Pattavina, Jiru Wang, Yi Zeng:
Joint Energy and Spectrum Efficient Virtual Optical Network embedding in EONs. HPSR 2019: 1-6 - [c2]Wencheng Chen, Yi Zeng, Meikang Qiu:
Using Adversarial Examples to Bypass Deep Learning Based URL Detection System. SmartCloud 2019: 128-130 - 2018
- [c1]Yi Zeng, Meikang Qiu, Zhong Ming, Meiqin Liu:
Senior2Local: A Machine Learning Based Intrusion Detection Method for VANETs. SmartCom 2018: 417-426
Informal and Other Publications
- 2024
- [i29]Yi Zeng, Hongpeng Lin, Jingwen Zhang, Diyi Yang, Ruoxi Jia, Weiyan Shi:
How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs. CoRR abs/2401.06373 (2024) - [i28]Shayne Longpre, Sayash Kapoor, Kevin Klyman, Ashwin Ramaswami, Rishi Bommasani, Borhane Blili-Hamelin, Yangsibo Huang, Aviya Skowron, Zheng Xin Yong, Suhas Kotha, Yi Zeng, Weiyan Shi, Xianjun Yang, Reid Southen, Alexander Robey, Patrick Chao, Diyi Yang, Ruoxi Jia, Daniel Kang, Sandy Pentland, Arvind Narayanan, Percy Liang, Peter Henderson:
A Safe Harbor for AI Evaluation and Red Teaming. CoRR abs/2403.04893 (2024) - [i27]Zhuowen Yuan, Zidi Xiong, Yi Zeng, Ning Yu, Ruoxi Jia, Dawn Song, Bo Li:
RigorLLM: Resilient Guardrails for Large Language Models against Undesired Content. CoRR abs/2403.13031 (2024) - [i26]Bertie Vidgen, Adarsh Agrawal, Ahmed M. Ahmed, Victor Akinwande, Namir Al-Nuaimi, Najla Alfaraj, Elie Alhajjar, Lora Aroyo, Trupti Bavalatti, Borhane Blili-Hamelin, Kurt D. Bollacker, Rishi Bomassani, Marisa Ferrara Boston, Siméon Campos, Kal Chakra, Canyu Chen, Cody Coleman, Zacharie Delpierre Coudert, Leon Derczynski, Debojyoti Dutta, Ian Eisenberg, James Ezick, Heather Frase, Brian Fuller, Ram Gandikota, Agasthya Gangavarapu, Ananya Gangavarapu, James Gealy, Rajat Ghosh, James Goel, Usman Gohar, Subhra S. Goswami, Scott A. Hale, Wiebke Hutiri, Joseph Marvin Imperial, Surgan Jandial, Nick Judd, Felix Juefei-Xu, Foutse Khomh, Bhavya Kailkhura, Hannah Rose Kirk, Kevin Klyman, Chris Knotz, Michael Kuchnik, Shachi H. Kumar, Chris Lengerich, Bo Li, Zeyi Liao, Eileen Peters Long, Victor Lu, Yifan Mai, Priyanka Mary Mammen, Kelvin Manyeki, Sean McGregor, Virendra Mehta, Shafee Mohammed, Emanuel Moss, Lama Nachman, Dinesh Jinenhally Naganna, Amin Nikanjam, Besmira Nushi, Luis Oala, Iftach Orr, Alicia Parrish, Cigdem Patlak, William Pietri, Forough Poursabzi-Sangdeh, Eleonora Presani, Fabrizio Puletti, Paul Röttger, Saurav Sahay, Tim Santos, Nino Scherrer, Alice Schoenauer Sebag, Patrick Schramowski, Abolfazl Shahbazi, Vin Sharma, Xudong Shen, Vamsi Sistla, Leonard Tang, Davide Testuggine, Vithursan Thangarasa, Elizabeth Anne Watkins, Rebecca Weiss, Chris Welty, Tyler Wilbers, Adina Williams, Carole-Jean Wu, Poonam Yadav, Xianjun Yang, Yi Zeng, Wenhui Zhang, Fedor Zhdanov, Jiacheng Zhu, Percy Liang, Peter Mattson, Joaquin Vanschoren:
Introducing v0.5 of the AI Safety Benchmark from MLCommons. CoRR abs/2404.12241 (2024) - [i25]Xiangyu Qi, Yangsibo Huang, Yi Zeng, Edoardo Debenedetti, Jonas Geiping, Luxi He, Kaixuan Huang, Udari Madhushani, Vikash Sehwag, Weijia Shi, Boyi Wei, Tinghao Xie, Danqi Chen, Pin-Yu Chen, Jeffrey Ding, Ruoxi Jia, Jiaqi Ma, Arvind Narayanan, Weijie J. Su, Mengdi Wang, Chaowei Xiao, Bo Li, Dawn Song, Peter Henderson, Prateek Mittal:
AI Risk Management Should Incorporate Both Safety and Security. CoRR abs/2405.19524 (2024) - [i24]Minzhou Pan, Yi Zeng, Xue Lin, Ning Yu, Cho-Jui Hsieh, Peter Henderson, Ruoxi Jia:
JIGMARK: A Black-Box Approach for Enhancing Image Watermarks against Diffusion Model Edits. CoRR abs/2406.03720 (2024) - [i23]Yi Zeng, Xuelin Yang, Li Chen, Cristian Canton Ferrer, Ming Jin, Michael I. Jordan, Ruoxi Jia:
Fairness-Aware Meta-Learning via Nash Bargaining. CoRR abs/2406.07029 (2024) - [i22]Tinghao Xie, Xiangyu Qi, Yi Zeng, Yangsibo Huang, Udari Madhushani Sehwag, Kaixuan Huang, Luxi He, Boyi Wei, Dacheng Li, Ying Sheng, Ruoxi Jia, Bo Li, Kai Li, Danqi Chen, Peter Henderson, Prateek Mittal:
SORRY-Bench: Systematically Evaluating Large Language Model Safety Refusal Behaviors. CoRR abs/2406.14598 (2024) - [i21]Yi Zeng, Weiyu Sun, Tran Ngoc Huynh, Dawn Song, Bo Li, Ruoxi Jia:
BEEAR: Embedding-based Adversarial Removal of Safety Backdoors in Instruction-tuned Language Models. CoRR abs/2406.17092 (2024) - [i20]Yi Zeng, Kevin Klyman, Andy Zhou, Yu Yang, Minzhou Pan, Ruoxi Jia, Dawn Song, Percy Liang, Bo Li:
AI Risk Categorization Decoded (AIR 2024): From Government Regulations to Corporate Policies. CoRR abs/2406.17864 (2024) - [i19]Yi Zeng, Yu Yang, Andy Zhou, Jeffrey Ziwei Tan, Yuheng Tu, Yifan Mai, Kevin Klyman, Minzhou Pan, Ruoxi Jia, Dawn Song, Percy Liang, Bo Li:
AIR-Bench 2024: A Safety Benchmark Based on Risk Categories from Regulations and Policies. CoRR abs/2407.17436 (2024) - 2023
- [i18]Minzhou Pan, Yi Zeng, Lingjuan Lyu, Xue Lin, Ruoxi Jia:
ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning Paradigms. CoRR abs/2302.11408 (2023) - [i17]Hoang Anh Just, Feiyang Kang, Jiachen T. Wang, Yi Zeng, Myeongseob Ko, Ming Jin, Ruoxi Jia:
LAVA: Data Valuation without Pre-Specified Learning Algorithms. CoRR abs/2305.00054 (2023) - [i16]Zhenting Wang, Chen Chen, Yi Zeng, Lingjuan Lyu, Shiqing Ma:
Alteration-free and Model-agnostic Origin Attribution of Generated Images. CoRR abs/2305.18439 (2023) - [i15]Junyuan Hong, Yi Zeng, Shuyang Yu, Lingjuan Lyu, Ruoxi Jia, Jiayu Zhou:
Revisiting Data-Free Knowledge Distillation with Poisoned Teachers. CoRR abs/2306.02368 (2023) - [i14]Xiangyu Qi, Yi Zeng, Tinghao Xie, Pin-Yu Chen, Ruoxi Jia, Prateek Mittal, Peter Henderson:
Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To! CoRR abs/2310.03693 (2023) - [i13]Shuyang Yu, Junyuan Hong, Yi Zeng, Fei Wang, Ruoxi Jia, Jiayu Zhou:
Who Leaked the Model? Tracking IP Infringers in Accountable Federated Learning. CoRR abs/2312.03205 (2023) - 2022
- [i12]Yi Zeng, Minzhou Pan, Hoang Anh Just, Lingjuan Lyu, Meikang Qiu, Ruoxi Jia:
Narcissus: A Practical Clean-Label Backdoor Attack with Limited Information. CoRR abs/2204.05255 (2022) - [i11]Xuanli He, Qiongkai Xu, Yi Zeng, Lingjuan Lyu, Fangzhao Wu, Jiwei Li, Ruoxi Jia:
CATER: Intellectual Property Protection on Text Generation APIs via Conditional Watermarks. CoRR abs/2209.08773 (2022) - [i10]Yi Zeng, Minzhou Pan, Himanshu Jahagirdar, Ming Jin, Lingjuan Lyu, Ruoxi Jia:
How to Sift Out a Clean Data Subset in the Presence of Data Poisoning? CoRR abs/2210.06516 (2022) - 2021
- [i9]Yi Zeng, Won Park, Z. Morley Mao, Ruoxi Jia:
Rethinking the Backdoor Attacks' Triggers: A Frequency Perspective. CoRR abs/2104.03413 (2021) - [i8]Tianhao Wang, Yi Zeng, Ming Jin, Ruoxi Jia:
A Unified Framework for Task-Driven Data Quality Management. CoRR abs/2106.05484 (2021) - [i7]Yi Zeng, Si Chen, Won Park, Z. Morley Mao, Ming Jin, Ruoxi Jia:
Adversarial Unlearning of Backdoors via Implicit Hypergradient. CoRR abs/2110.03735 (2021) - 2020
- [i6]Han Qiu, Yi Zeng, Qinkai Zheng, Tianwei Zhang, Meikang Qiu, Gérard Memmi:
Mitigating Advanced Adversarial Attacks with More Advanced Gradient Obfuscation Techniques. CoRR abs/2005.13712 (2020) - [i5]Yi Zeng, Han Qiu, Gérard Memmi, Meikang Qiu:
A Data Augmentation-based Defense Method Against Adversarial Attacks in Neural Networks. CoRR abs/2007.15290 (2020) - [i4]Shangwei Guo, Tianwei Zhang, Han Qiu, Yi Zeng, Tao Xiang, Yang Liu:
The Hidden Vulnerability of Watermarking for Deep Neural Networks. CoRR abs/2009.08697 (2020) - [i3]Han Qiu, Yi Zeng, Tianwei Zhang, Yong Jiang, Meikang Qiu:
FenceBox: A Platform for Defeating Adversarial Examples with Data Augmentation Techniques. CoRR abs/2012.01701 (2020) - [i2]Yi Zeng, Han Qiu, Shangwei Guo, Tianwei Zhang, Meikang Qiu, Bhavani Thuraisingham:
DeepSweep: An Evaluation Framework for Mitigating DNN Backdoor Attacks using Data Augmentation. CoRR abs/2012.07006 (2020) - 2019
- [i1]Yi Zeng, Zihao Qi, Wencheng Chen, Yanzhe Huang, Xingxin Zheng, Han Qiu:
TEST: an End-to-End Network Traffic Examination and Identification Framework Based on Spatio-Temporal Features Extraction. CoRR abs/1908.10271 (2019)
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-10-22 20:13 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint