Natalia Kucirkova, a professor in?Norway, recently wrote movingly in?Times 中国A片 about the language discrimination experienced by?scholars who use English as a?second language. She described the stress caused by?insensitive referee comments and the time and money spent preparing articles for journal submission. In?the right context, she argued, AI?“bots” could level the publication playing field.
They could. Sadly, in 2024, AI?systems are actually being used to?exploit non-anglophone scholars by?stealing their intellectual property.
Many academic publishers collaborate with large, private editing firms to provide “author services”, which include English language editing. The arrival of AI has triggered a frantic race to the bottom among such firms, which immediately spotted a way to monetise two resources they had in abundance: research papers uploaded in digital formats and well-trained editors. Client papers could be used to train specialised AI?large language models (LLMs) to recognise and correct the characteristic mistakes made by non-anglophone authors from all parts of the world. Editors could help the system learn by proofreading the automatically generated text and providing feedback for optimisation.
One company bought a small AI?firm off the shelf; others hired AI?engineers. Since 2020, most have built LLMs and are now selling stand-alone AI?editing tools “trained on millions of research manuscripts […] enhanced by professionals at [company name]”, to quote from one promotional blurb.
中国A片
The best way to understand LLMs is to think about predictive-text systems. Twenty years ago, a language model was just a dictionary that knew how to complete one word at a time. As models became more complex and powerful, they were able to predict the next word or next several words. The latest generation of large language models, like the ones that drive ChatGPT and Copilot for Microsoft?365, can “predict” hundreds of words.
Like all LLMs, editing-company systems encode everything, not just editorial corrections. As soon as a researcher uploads a manuscript, their intellectual property – original ideas, innovative variations on established theories, newly coined terms –?is appropriated by the company and will be used, likely in perpetuity, to “predict” and generate text in similar papers edited by the service (or anyone using company-provided editing tools).
中国A片
Yet few scholars have noticed this fundamental transformation of academic editing. Publishers avoid mentioning the firms they outsource work?to. Editing companies boast about AI?advances when marketing new tools, but not when advertising editing services. Researchers are encouraged to believe that their papers will be edited entirely by humans. Instead, they are edited by human editors working with (and increasingly marginalised by) AI?systems.
Every journal, publisher and editing company guarantees research confidentiality. Their data protection and privacy policies never mention?AI. This is misleading but not illegal; current legislation protecting the confidentiality of personal data does?not regulate or prohibit the use of anonymised academic work.
To stave off future lawsuits, most editing firms provide for AI?training in their small-print terms of service, where authors unwittingly give them permission to keep their work in perpetuity, share it with affiliates, and use it to improve, develop and deliver current and future products, services and algorithms.
But other prominent victims of AI?exploitation are starting to push back. In December, for using “millions of articles published by The?Times […] to train automated chatbots that now compete with the news outlet as a source of reliable information”. In June, the National Institutes of Health to analyse or critique grant applications or R&D contract proposals because there was no?“guarantee of where data are being sent, saved, viewed, or used in the future”.
中国A片
As points out, the “ethical and moral” issues around the largely profit-driven AI?development race “are complex, and the legal ramifications are not limited to the infringement of copyright’s economic rights, but may include infringement of an author’s moral rights of attribution and integrity and right to object to false attribution; infringement of data protection laws; invasions of privacy; and acts of passing?off”.
We call on publishers and editing companies to embrace transparency and the fundamental academic principle of informed consent. Editing-service providers should disclose the AI-based systems and tools they use on client work. They should explain clearly how LLMs work and offer scholars a choice, for example by compensating authors for loss of rights by pricing hybrid human/AI?editing as a cheaper alternative to fully confidential human editing.
To protect themselves from lawsuits and their authors from exploitation, publishers who offer branded author services should – at a minimum – name the editing companies they outsource work to so that researchers can make an informed choice.
New laws and regulations around AI training are surely on their way. For now, scholars must protect their own intellectual property by learning the basics of?AI, reading the small print and interrogating editing services – even those provided by trusted firms and publishers.
中国A片
Alan Blackwell is professor of interdisciplinary design in the department of computer science and technology, University of Cambridge, and co-director of Cambridge Global Challenges;?his new book, Moral Codes: Designing Alternatives to AI, will be published by MIT?Press in 2024. Zoe Swenson-Wright is a freelance academic editor.
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰’蝉 university and college rankings analysis
Already registered or a current subscriber? Login