OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 26.03.2026, 20:56

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Using AI to write scholarly publications

2023·240 Zitationen·Accountability in ResearchOpen Access
Volltext beim Verlag öffnen

240

Zitationen

3

Autoren

2023

Jahr

Abstract

Artificial intelligence (AI) natural language processing (NLP) systems, such as OpenAI's generative pre-trained transformer (GPT) model (https://openai. com) or Meta's Galactica (https://galactica.org/)may soon be widely used in many forms of writing, including scientific and scholarly publications (Heaven 2022). 1 While computer programs (such as Microsoft WORD and Grammarly) have incorporated automated text-editing features (such as checking for spelling and grammar) for many years, these programs are not designed to create content.However, new and emerging NLP systems are, which raises important issues for research ethics and research integrity. 2 NLP is a way of enabling computers to interact with human language.A key step in NLP, known as tokenization, involves converting unstructured text into structured text suitable for computation.For example, the sentence "The cat sat on the mat" can be structured by tagging its parts: "the [article] cat [noun] sat [verb, past tense] on [preposition] the [article] mat [noun]."Once the parts of the text have been tagged, they can be processed by means of algorithms designed to produce appropriate responses to text (i.e., language generation).Rudimentary NLP-systems, such as the first generation of chatbots that assisted customers on websites, operated according to thousands of human-written rules for processing and generating text.Recent advances in computational speed and capacity and the development of machine-learning (ML) algorithms, such as neural networks, have led to tremendous breakthroughs in NLP (Mitchell 2020).Today's NLP systems use ML to produce and refine statistical models (with billions of parameters) for processing and generating natural language.NLP systems are trained on huge databases (45 terabytes or more) of text available on the internet or other sources.Initial training (or supervised learning) involves giving the system the text and then "rewarding" it for giving correct outputs, as determined by human trainers. 3Over time, NLP systems will reduce their percentage of erroneous outputs and will learn from the data (Mitchell 2020).While NLP systems continue to learn as they receive and process data beyond their initial training data, they do not "know" the meaning or truth-

Ähnliche Arbeiten