Loading…

Open-Ethical AI: Advancements in Open-Source Human-Centric Neural Language Models

This survey summarizes the most recent methods for building and assessing helpful, honest, and harmless neural language models, considering small, medium, and large-size models. Pointers to open-source resources that help to align pre-trained models are given, including methods that use parameter-ef...

Full description

Saved in:
Bibliographic Details
Published in:ACM computing surveys 2025-04, Vol.57 (4), p.1-47
Main Authors: Sicari, Sabrina, Cevallos M., Jesus F., Rizzardi, Alessandra, Coen-Porisini, Alberto
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This survey summarizes the most recent methods for building and assessing helpful, honest, and harmless neural language models, considering small, medium, and large-size models. Pointers to open-source resources that help to align pre-trained models are given, including methods that use parameter-efficient techniques, specialized prompting frameworks, adapter modules, case-specific knowledge injection, and adversarially robust training techniques. Special care is given to evidencing recent progress on value alignment, commonsense reasoning, factuality enhancement, and abstract reasoning of language models. Most reviewed works in this survey publicly shared their code and related data and were accepted in world-leading Machine Learning venues. This work aims to help researchers and practitioners accelerate their entrance into the field of human-centric neural language models, which might be a cornerstone of the contemporary and near-future industrial and societal revolution.
ISSN:0360-0300
1557-7341
DOI:10.1145/3703454