ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.08005
  4. Cited By
Beyond the Safeguards: Exploring the Security Risks of ChatGPT

Beyond the Safeguards: Exploring the Security Risks of ChatGPT

13 May 2023
Erik Derner
Kristina Batistic
    SILM
ArXiv (abs)PDFHTML

Papers citing "Beyond the Safeguards: Exploring the Security Risks of ChatGPT"

37 / 37 papers shown
LLM in the Middle: A Systematic Review of Threats and Mitigations to Real-World LLM-based Systems
LLM in the Middle: A Systematic Review of Threats and Mitigations to Real-World LLM-based Systems
Vitor Hugo Galhardo Moia
Igor Jochem Sanz
Gabriel Antonio Fontes Rebello
Rodrigo Duarte de Meneses
Briland Hitaj
Ulf Lindqvist
237
0
0
12 Sep 2025
Two Birds with One Stone: Multi-Task Detection and Attribution of LLM-Generated Text
Two Birds with One Stone: Multi-Task Detection and Attribution of LLM-Generated Text
Zixin Rao
Youssef Mohamed
Shang Liu
Zeyan Liu
DeLMO
176
0
0
19 Aug 2025
Securing Educational LLMs: A Generalised Taxonomy of Attacks on LLMs and DREAD Risk Assessment
Securing Educational LLMs: A Generalised Taxonomy of Attacks on LLMs and DREAD Risk Assessment
Farzana Zahid
Anjalika Sewwandi
Lee Brandon
Vimal Kumar
Roopak Sinha
171
1
0
12 Aug 2025
AI Ethics and Social Norms: Exploring ChatGPT's Capabilities From What to How
AI Ethics and Social Norms: Exploring ChatGPT's Capabilities From What to HowProceedings of the ACM on Human-Computer Interaction (PACMHCI), 2025
Omid Veisi
Sasan Bahrami
Roman Englert
Claudia Müller
386
0
0
25 Apr 2025
SOK: Exploring Hallucinations and Security Risks in AI-Assisted Software Development with Insights for LLM Deployment
SOK: Exploring Hallucinations and Security Risks in AI-Assisted Software Development with Insights for LLM Deployment
Ariful Haque
Sunzida Siddique
M. Rahman
Ahmed Rafi Hasan
Laxmi Rani Das
Marufa Kamal
Tasnim Masura
Kishor Datta Gupta
304
6
0
31 Jan 2025
AI Safety in Generative AI Large Language Models: A Survey
AI Safety in Generative AI Large Language Models: A Survey
Jaymari Chua
Yun Yvonna Li
Shiyi Yang
Chen Wang
Lina Yao
LM&MA
349
36
0
06 Jul 2024
The Art of Saying No: Contextual Noncompliance in Language Models
The Art of Saying No: Contextual Noncompliance in Language Models
Faeze Brahman
Sachin Kumar
Vidhisha Balachandran
Pradeep Dasigi
Valentina Pyatkin
...
Jack Hessel
Yulia Tsvetkov
Noah A. Smith
Yejin Choi
Hannaneh Hajishirzi
288
57
0
02 Jul 2024
A Complete Survey on LLM-based AI Chatbots
A Complete Survey on LLM-based AI Chatbots
Sumit Kumar Dam
Choong Seon Hong
Yu Qiao
Chaoning Zhang
279
124
0
17 Jun 2024
Is On-Device AI Broken and Exploitable? Assessing the Trust and Ethics in Small Language Models
Is On-Device AI Broken and Exploitable? Assessing the Trust and Ethics in Small Language Models
Kalyan Nakka
Jimmy Dani
Nitesh Saxena
423
3
0
08 Jun 2024
Measure-Observe-Remeasure: An Interactive Paradigm for
  Differentially-Private Exploratory Analysis
Measure-Observe-Remeasure: An Interactive Paradigm for Differentially-Private Exploratory Analysis
Priyanka Nanayakkara
Hyeok Kim
Yifan Wu
Ali Sarvghad
Narges Mahyar
G. Miklau
Jessica Hullman
236
39
0
04 Jun 2024
Towards Trustworthy AI: A Review of Ethical and Robust Large Language
  Models
Towards Trustworthy AI: A Review of Ethical and Robust Large Language Models
Meftahul Ferdaus
Mahdi Abdelguerfi
Elias Ioup
Kendall N. Niles
Ken Pathak
Steve Sloan
398
26
0
01 Jun 2024
FreezeAsGuard: Mitigating Illegal Adaptation of Diffusion Models via
  Selective Tensor Freezing
FreezeAsGuard: Mitigating Illegal Adaptation of Diffusion Models via Selective Tensor Freezing
Kai Huang
Wei Gao
227
3
0
24 May 2024
Tagengo: A Multilingual Chat Dataset
Tagengo: A Multilingual Chat Dataset
P. Devine
143
4
0
21 May 2024
Risks of Practicing Large Language Models in Smart Grid: Threat Modeling and Validation
Risks of Practicing Large Language Models in Smart Grid: Threat Modeling and Validation
Jiangnan Li
Yingyuan Yang
Jinyuan Stella Sun
351
8
0
10 May 2024
Large Language Models for Cyber Security: A Systematic Literature Review
Large Language Models for Cyber Security: A Systematic Literature Review
HanXiang Xu
Shenao Wang
Ningke Li
Kaidi Wang
Yanjie Zhao
Kai Chen
Ting Yu
Yang Liu
Haoyu Wang
587
106
0
08 May 2024
SmartMem: Layout Transformation Elimination and Adaptation for Efficient
  DNN Execution on Mobile
SmartMem: Layout Transformation Elimination and Adaptation for Efficient DNN Execution on Mobile
Wei Niu
Md. Musfiqur Rahman Sanim
Zhihao Shu
Jiexiong Guan
Xipeng Shen
Miao Yin
Gagan Agrawal
Bin Ren
186
11
0
21 Apr 2024
Risk and Response in Large Language Models: Evaluating Key Threat
  Categories
Risk and Response in Large Language Models: Evaluating Key Threat Categories
Bahareh Harandizadeh
A. Salinas
Fred Morstatter
222
7
0
22 Mar 2024
On Protecting the Data Privacy of Large Language Models (LLMs): A Survey
On Protecting the Data Privacy of Large Language Models (LLMs): A SurveyInternational Conference on Mathematics and Computing (ICMC), 2024
Biwei Yan
Kun Li
Minghui Xu
Yueyan Dong
Yue Zhang
Zhaochun Ren
Xiuzhen Cheng
AILawPILM
407
158
0
08 Mar 2024
Exploring the Potential of Large Language Models for Improving Digital Forensic Investigation Efficiency
Exploring the Potential of Large Language Models for Improving Digital Forensic Investigation Efficiency
Akila Wickramasekara
Frank Breitinger
Mark Scanlon
495
24
0
29 Feb 2024
Making Them Ask and Answer: Jailbreaking Large Language Models in Few
  Queries via Disguise and Reconstruction
Making Them Ask and Answer: Jailbreaking Large Language Models in Few Queries via Disguise and Reconstruction
Tong Liu
Yingjie Zhang
Zhe Zhao
Yinpeng Dong
Guozhu Meng
Kai Chen
AAML
247
105
0
28 Feb 2024
Farsight: Fostering Responsible AI Awareness During AI Application
  Prototyping
Farsight: Fostering Responsible AI Awareness During AI Application Prototyping
Zijie J. Wang
Chinmay Kulkarni
Lauren Wilcox
Michael Terry
Michael A. Madaio
317
71
0
23 Feb 2024
Mapping the Ethics of Generative AI: A Comprehensive Scoping Review
Mapping the Ethics of Generative AI: A Comprehensive Scoping Review
Thilo Hagendorff
253
85
0
13 Feb 2024
Whispers in the Machine: Confidentiality in Agentic Systems
Whispers in the Machine: Confidentiality in Agentic Systems
Jonathan Evertz
Merlin Chlosta
Lea Schonherr
Thorsten Eisenhofer
333
23
0
10 Feb 2024
Improving Dialog Safety using Socially Aware Contrastive Learning
Improving Dialog Safety using Socially Aware Contrastive Learning
Souvik Das
Rohini Srihari
219
1
0
01 Feb 2024
The Ethics of Interaction: Mitigating Security Threats in LLMs
The Ethics of Interaction: Mitigating Security Threats in LLMs
Ashutosh Kumar
Shiv Vignesh Murty
Sagarika Singh
Swathy Ragupathy
281
48
0
22 Jan 2024
A Survey on Large Language Model (LLM) Security and Privacy: The Good,
  the Bad, and the Ugly
A Survey on Large Language Model (LLM) Security and Privacy: The Good, the Bad, and the UglyHigh-Confidence Computing (HC), 2023
Yifan Yao
Jinhao Duan
Kaidi Xu
Yuanfang Cai
Eric Sun
Yue Zhang
PILMELM
582
920
0
04 Dec 2023
From Chatbots to PhishBots? -- Preventing Phishing scams created using
  ChatGPT, Google Bard and Claude
From Chatbots to PhishBots? -- Preventing Phishing scams created using ChatGPT, Google Bard and Claude
Sayak Saha Roy
Poojitha Thota
Krishna Vamsi Naragam
Shirin Nilizadeh
SILM
329
23
0
29 Oct 2023
Ask Again, Then Fail: Large Language Models' Vacillations in Judgment
Ask Again, Then Fail: Large Language Models' Vacillations in JudgmentAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Qiming Xie
Zengzhi Wang
Yi Feng
Rui Xia
AAMLHILM
648
12
0
03 Oct 2023
Can LLM-Generated Misinformation Be Detected?
Can LLM-Generated Misinformation Be Detected?International Conference on Learning Representations (ICLR), 2023
Canyu Chen
Kai Shu
DeLMO
782
239
0
25 Sep 2023
Efficient Avoidance of Vulnerabilities in Auto-completed Smart Contract
  Code Using Vulnerability-constrained Decoding
Efficient Avoidance of Vulnerabilities in Auto-completed Smart Contract Code Using Vulnerability-constrained DecodingIEEE International Symposium on Software Reliability Engineering (ISSRE), 2023
André Storhaug
Jingyue Li
Tianyuan Hu
AAML
175
26
0
18 Sep 2023
Distilled GPT for Source Code Summarization
Distilled GPT for Source Code SummarizationInternational Conference on Automated Software Engineering (ASE), 2023
Chia-Yi Su
Collin McMillan
259
54
0
28 Aug 2023
GPTEval: A Survey on Assessments of ChatGPT and GPT-4
GPTEval: A Survey on Assessments of ChatGPT and GPT-4International Conference on Language Resources and Evaluation (LREC), 2023
Rui Mao
Guanyi Chen
Xulang Zhang
Frank Guerin
Xiaoshi Zhong
ELMLM&MA
185
147
0
24 Aug 2023
Using Large Language Models for Cybersecurity Capture-The-Flag
  Challenges and Certification Questions
Using Large Language Models for Cybersecurity Capture-The-Flag Challenges and Certification Questions
W. Tann
Yuancheng Liu
Jun Heng Sim
C. Seah
E. Chang
ELM
209
43
0
21 Aug 2023
RatGPT: Turning online LLMs into Proxies for Malware Attacks
RatGPT: Turning online LLMs into Proxies for Malware Attacks
Mika Beckerich
L. Plein
Sergio Coronado
SILM
135
38
0
17 Aug 2023
Learning to Prompt in the Classroom to Understand AI Limits: A pilot
  study
Learning to Prompt in the Classroom to Understand AI Limits: A pilot studyInternational Conference of the Italian Association for Artificial Intelligence (AIxIA), 2023
Emily Theophilou
Cansu Koyuturk
Mona Yavari
Sathya Bursic
Gregor Donabauer
...
Davinia Hernández Leo
Martin Ruskov
D. Taibi
A. Gabbiadini
D. Ognibene
220
50
0
04 Jul 2023
On the Detectability of ChatGPT Content: Benchmarking, Methodology, and
  Evaluation through the Lens of Academic Writing
On the Detectability of ChatGPT Content: Benchmarking, Methodology, and Evaluation through the Lens of Academic WritingConference on Computer and Communications Security (CCS), 2023
Zeyan Liu
Zijun Yao
Fengjun Li
Bo Luo
DeLMO
222
39
0
07 Jun 2023
From Text to MITRE Techniques: Exploring the Malicious Use of Large
  Language Models for Generating Cyber Attack Payloads
From Text to MITRE Techniques: Exploring the Malicious Use of Large Language Models for Generating Cyber Attack Payloads
P. Charan
Hrushikesh Chunduri
P. Anand
S. Shukla
178
60
0
24 May 2023
1