Abstract
The increase in popularity of Large Language Models (LLMs) has given rise to concerns about their vulnerabilities and security risks that make them open to exploitation and misuse. Their widespread adoption prompts the need for further exploration and mitigation of these potential dangers. Here, we adopt a comprehensive framework for examining LLM vulnerabilities centered on three fundamental criteria: reliability, resiliency, and responsibility. These criteria serve as the foundation for ensuring LLMs are reliable in real-world applications, resilient in the face of potential threats, and are responsibly deployed to uphold ethical and societal standards. By utilizing this framework, we aim to provide a structured and systematic approach for understanding, developing, and deploying trustworthy LLMs.
Original language | English |
---|---|
Article number | 252 |
Journal | SN Computer Science |
Volume | 6 |
Issue number | 3 |
DOIs | |
State | Published - Mar 2025 |
Funding
This manuscript has been authored by UT-Battelle, LLC under Contract No. DE-AC05-00OR22725 with the U.S. Department of Energy. The United States Government retains and the publisher, by accepting the article for publication, acknowledges that the United States Government retains a non-exclusive, paid-up, irrevocable, world-wide license to publish or reproduce the published form of this manuscript, or allow others to do so, for United States Government purposes. The Department of Energy will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan(http://energy.gov/downloads/doe-public-access-plan).
Keywords
- Large language models
- Safety risks
- Trustworthiness