Safeguarding large language models: a survey

In the burgeoning field of Large Language Models (LLMs), developing a robust safety mechanism, colloquially known as “safeguards” or “guardrails”, has become imperative to ensure the ethical use of LLMs within prescribed boundaries. This article provides a systematic literature review on the current...

Full description

Saved in:
Bibliographic Details
Main Authors: Yi Dong, Ronghui Mu, Yanghao Zhang, Siqi Sun, Tianle Zhang, Changshun Wu, Gaojie Jin, Yi Qi, Jinwei Hu, Jie Meng, Saddek Bensalem, Xiaowei Huang
Format: Default Article
Published: 2025
Subjects:
Online Access:https://hdl.handle.net/2134/30460130.v1
Tags: Add Tag
No Tags, Be the first to tag this record!