Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Ethical Analysis of Large Language Models: A Comparison of GPT, DeepSeek, Claude, Gemini, LLaMA, and Qwen
0
Zitationen
2
Autoren
2025
Jahr
Abstract
An ethical evaluation of Large Language Models (LLM) requires good background knowledge regarding their technical bases; only then is it possible to see how their architectures and training methodologies lead to ethical concerns and biases. The present review evaluates qualitatively the reports produced for GPT, DeepSeek, Claude, Gemini, LLaMA, and Qwen on different aspects of their technical specifications in terms of how design choices shape ethical considerations like bias, fairness, and transparency. We examine how model architecture, data curation strategies, and ethical bias mitigation and propagation is being handled by different large language models. Through our findings, it appears that proprietary models adopt advanced solutions to manage bias but without being transparent. On the other hand, open models are amenable to an independent audit; however, the datasets that are available for public scrutiny usually harbor some biases that are inherent to the datasets themselves. Thus, in our discussions, we emphasize the need for standardized evaluation frameworks to ensure that ethical considerations are parallel to advances in AI in the future.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.339 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.211 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.614 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.478 Zit.