Natalia Kucirkova, a professor 颈苍听Norway, recently wrote movingly 颈苍听探花视频 about the language discrimination experienced by聽scholars who use English as a聽second language. She described the stress caused by聽insensitive referee comments and the time and money spent preparing articles for journal submission. In聽the right context, she argued, AI聽鈥渂ots鈥 could level the publication playing field.
They could. Sadly, in 2024, AI聽systems are actually being used to聽exploit non-anglophone scholars by聽stealing their intellectual property.
Many academic publishers collaborate with large, private editing firms to provide 鈥渁uthor services鈥, which include English language editing. The arrival of AI has triggered a frantic race to the bottom among such firms, which immediately spotted a way to monetise two resources they had in abundance: research papers uploaded in digital formats and well-trained editors. Client papers could be used to train specialised AI聽large language models (LLMs) to recognise and correct the characteristic mistakes made by non-anglophone authors from all parts of the world. Editors could help the system learn by proofreading the automatically generated text and providing feedback for optimisation.
One company bought a small AI聽firm off the shelf; others hired AI聽engineers. Since 2020, most have built LLMs and are now selling stand-alone AI聽editing tools 鈥渢rained on millions of research manuscripts [鈥 enhanced by professionals at [company name]鈥, to quote from one promotional blurb.
探花视频
The best way to understand LLMs is to think about predictive-text systems. Twenty years ago, a language model was just a dictionary that knew how to complete one word at a time. As models became more complex and powerful, they were able to predict the next word or next several words. The latest generation of large language models, like the ones that drive ChatGPT and Copilot for Microsoft聽365, can 鈥減redict鈥 hundreds of words.
Like all LLMs, editing-company systems encode everything, not just editorial corrections. As soon as a researcher uploads a manuscript, their intellectual property 鈥 original ideas, innovative variations on established theories, newly coined terms 鈥撀爄s appropriated by the company and will be used, likely in perpetuity, to 鈥減redict鈥 and generate text in similar papers edited by the service (or anyone using company-provided editing tools).
探花视频
Yet few scholars have noticed this fundamental transformation of academic editing. Publishers avoid mentioning the firms they outsource work聽to. Editing companies boast about AI聽advances when marketing new tools, but not when advertising editing services. Researchers are encouraged to believe that their papers will be edited entirely by humans. Instead, they are edited by human editors working with (and increasingly marginalised by) AI聽systems.
Every journal, publisher and editing company guarantees research confidentiality. Their data protection and privacy policies never mention聽AI. This is misleading but not illegal; current legislation protecting the confidentiality of personal data does聽not regulate or prohibit the use of anonymised academic work.
To stave off future lawsuits, most editing firms provide for AI聽training in their small-print terms of service, where authors unwittingly give them permission to keep their work in perpetuity, share it with affiliates, and use it to improve, develop and deliver current and future products, services and algorithms.
But other prominent victims of AI聽exploitation are starting to push back. In December, for using 鈥渕illions of articles published by 罢丑别听罢颈尘别蝉 [鈥 to train automated chatbots that now compete with the news outlet as a source of reliable information鈥. In June, the National Institutes of Health to analyse or critique grant applications or R&D contract proposals because there was no聽鈥済uarantee of where data are being sent, saved, viewed, or used in the future鈥.
探花视频
As points out, the 鈥渆thical and moral鈥 issues around the largely profit-driven AI聽development race 鈥渁re complex, and the legal ramifications are not limited to the infringement of copyright鈥檚 economic rights, but may include infringement of an author鈥檚 moral rights of attribution and integrity and right to object to false attribution; infringement of data protection laws; invasions of privacy; and acts of passing聽off鈥.
We call on publishers and editing companies to embrace transparency and the fundamental academic principle of informed consent. Editing-service providers should disclose the AI-based systems and tools they use on client work. They should explain clearly how LLMs work and offer scholars a choice, for example by compensating authors for loss of rights by pricing hybrid human/AI聽editing as a cheaper alternative to fully confidential human editing.
To protect themselves from lawsuits and their authors from exploitation, publishers who offer branded author services should 鈥 at a minimum 鈥 name the editing companies they outsource work to so that researchers can make an informed choice.
New laws and regulations around AI training are surely on their way. For now, scholars must protect their own intellectual property by learning the basics of聽AI, reading the small print and interrogating editing services 鈥 even those provided by trusted firms and publishers.
探花视频
Alan Blackwell is professor of interdisciplinary design in the department of computer science and technology, University of Cambridge, and co-director of Cambridge Global Challenges;聽his new book, Moral Codes: Designing Alternatives to AI, will be published by MIT聽Press in 2024. Zoe Swenson-Wright is a freelance academic editor.
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰鈥檚 university and college rankings analysis
Already registered or a current subscriber?








