Since generative artificial intelligence (AI) tools such as OpenAI's ChatGPT became widely available, researchers have used them in the writing process. The consensus of the academic publishing community is that such usage must be declared in the published article. Academ-AI documents examples of suspected undeclared AI usage in the academic literature, discernible primarily due to the appearance in research papers of idiosyncratic verbiage characteristic of large language model (LLM)-based chatbots. This analysis of the first 500 examples collected reveals that the problem is widespread, penetrating the journals and conference proceedings of highly respected publishers. Undeclared AI seems to appear in journals with higher citation metrics and higher article processing charges (APCs), precisely those outlets that should theoretically have the resources and expertise to avoid such oversights. An extremely small minority of cases are corrected post publication, and the corrections are often insufficient to rectify the problem. The 500 examples analyzed here likely represent a small fraction of the undeclared AI present in the academic literature, much of which may be undetectable. Publishers must enforce their policies against undeclared AI usage in cases that are detectable; this is the best defense currently available to the academic publishing community against the proliferation of undisclosed AI.
翻译:自OpenAI的ChatGPT等生成式人工智能工具广泛普及以来,研究者已将其应用于写作过程。学术出版界的共识是,此类使用必须在发表文章中予以声明。Academ-AI记录了学术文献中疑似未声明使用AI的案例,其识别主要依据研究论文中出现的、具有基于大语言模型的聊天机器人特有表达风格的异常措辞。对首批收集的500个案例的分析表明,该问题普遍存在,已渗透至多家权威出版商的期刊和会议论文集。未声明AI的使用现象在引用指标更高、文章处理费更昂贵的期刊中尤为突出,而这些出版渠道理论上本应具备避免此类疏漏的资源与专业知识。极少数案例在发表后得到更正,且更正措施往往不足以彻底解决问题。本文分析的500个案例很可能仅是学术文献中未声明AI使用的冰山一角,其中大部分或许无法被检测。出版商必须针对可检测的案例严格执行禁止未声明AI使用的政策;这是当前学术出版界应对未披露AI泛滥问题最有效的防御手段。