Communication

Will Artificial Intelligence Foster Plagiarism?

November 8, 2021 2570
Not all high-tech help in writing has to be unwelcome … (Photo: Mirko Tobias Schaefer/CC BY 2.0/via Wikimedia Commons)

The dramatic rise of online learning during the COVID-19 pandemic has spot-lit concerns about the role of technology in exam surveillance — and also in student cheating.

Some universities have reported more cheating during the pandemic, and such concerns are unfolding in a climate where technologies that allow for the automation of writing continue to improve.

Over the past two years, the ability of artificial intelligence to generate writing has leapt forward significantly, particularly with the development of what’s known as the language generator GPT-3. With this, companies such as Google, Microsoft and NVIDIA can now produce “human-like” text.

AI-generated writing has raised the stakes of how universities and schools will gauge what constitutes academic misconduct, such as plagiarism. As scholars with an interest in academic integrity and the intersections of work, society and educators’ labour, we believe that educators and parents should be, at the very least, paying close attention to these significant developments.

The Conversation logo
This article by Michael Mindzak and Sarah Elaine Eaton originally appeared on The Conversation, a Social Science Space partner site, under the title “Artificial intelligence is getting better at writing, and universities should worry about plagiarism”

AI and academic writing

The use of technology in academic writing is already widespread. For example, many universities already use text-based plagiarism detectors like Turnitin, while students might use Grammarly, a cloud-based writing assistant. Examples of writing support include automatic text generation, extraction, prediction, mining, form-filling, paraphrasing, translation and transcription.

Advancements in AI technology have led to new tools, products and services being offered to writers to improve content and efficiency. As these improve, soon entire articles or essays might be generated and written entirely by artificial intelligence. In schools, the implications of such developments will undoubtedly shape the future of learning, writing and teaching.

Misconduct concerns already widespread

Research has revealed that concerns over academic misconduct are already widespread across institutions higher education in Canada and internationally.

In Canada, there is little data regarding the rates of misconduct. Research published in 2006 based on data from mostly undergraduate students at 11 higher education institutions found 53 per cent reported having engaged in one or more instances of serious cheating on written work, which was defined as copying material without footnoting, copying material almost word for word, submitting work done by someone else, fabricating or falsifying a bibliography, submitting a paper they either bought or got from someone else for free.

Academic misconduct is in all likelihood under-reported across Canadian higher education institutions.

There are different types of violations of academic integrity, including plagiarism, contract cheating (where students hire other people to write their papers) and exam cheating, among others.

Unfortunately, with technology, students can use their ingenuity and entrepreneurialism to cheat. These concerns are also applicable to faculty members, academics and writers in other fields, bringing new concerns surrounding academic integrity and AI such as:

  • If a piece of writing was 49 percent written by AI, with the remaining 51 percent written by a human, is this considered original work?
  • What if an essay was 100 per cent written by AI, but a student did some of the coding themselves?
  • What qualifies as “AI assistance” as opposed to “academic cheating”?
  • Do the same rules apply to students as they would to academics and researchers?

We are asking these questions in our own research, and we know that in the face of all this, educators will be required to consider how writing can be effectively assessed or evaluated as these technologies improve.

Augmenting or diminishing integrity?

At the moment, little guidance, policy or oversight is available regarding technology, AI and academic integrity for teachers and educational leaders.

Over the past year, COVID-19 has pushed more students towards online learning — a sphere where teachers may become less familiar with their own students and thus, potentially, their writing.

While it remains impossible to predict the future of these technologies and their implications in education, we can attempt to discern some of the larger trends and trajectories that will impact teaching, learning and research.

Technology and automation in education

A key concern moving forward is the apparent movement towards the increased automation of education where educational technology companies offer commodities such as writing tools as proposed solutions for the various “problems” within education.

An example of this is automated assessment of student work, such as automated grading of student writing. Numerous commercial products already exist for automated grading, though the ethics of these technologies are yet to be fully explored by scholars and educators.

Overall, the traditional landscape surrounding academic integrity and authorship is being rapidly reshaped by technological developments. Such technological developments also spark concerns about a shift of professional control away from educators and ever-increasing new expectations of digital literacy in precarious working environments.

These complexities, concerns and questions will require further thought and discussion. Educational stakeholders at all levels will be required to respond and rethink definitions as well as values surrounding plagiarism, originality, academic ethics and academic labor in the very near future.


The authors would like to sincerely thank Ryan Morrison, from George Brown College, who provided significant expertise, advice and assistance with the development of this article.

Related Articles

Celebrating 20 Years of an Afrocentric Small Scholarly Press
Communication
May 7, 2024

Celebrating 20 Years of an Afrocentric Small Scholarly Press

Read Now
Striving for Linguistic Diversity in Scientific Research
Communication
May 1, 2024

Striving for Linguistic Diversity in Scientific Research

Read Now
Third Edition of ‘The Evidence’: How Can We Overcome Sexism in AI?
Bookshelf
April 29, 2024

Third Edition of ‘The Evidence’: How Can We Overcome Sexism in AI?

Read Now
The Long Arm of Criminality
Opinion
April 29, 2024

The Long Arm of Criminality

Read Now
To Better Forecast AI, We Need to Learn Where Its Money Is Pointing

To Better Forecast AI, We Need to Learn Where Its Money Is Pointing

By carefully interrogating the system of economic incentives underlying innovations and how technologies are monetized in practice, we can generate a better understanding of the risks, both economic and technological, nurtured by a market’s structure.

Read Now
Second Edition of ‘The Evidence’ Examines Women and Climate Change

Second Edition of ‘The Evidence’ Examines Women and Climate Change

The second issue of The Evidence explores the intersection of gender inequality and the global climate crisis. Author Josephine Lethbridge recounts the […]

Read Now
Free Online Course Reveals The Art of ChatGPT Interactions

Free Online Course Reveals The Art of ChatGPT Interactions

You’ve likely heard the hype around artificial intelligence, or AI, but do you find ChatGPT genuinely useful in your professional life? A free course offered by Sage Campus could change all th

Read Now
5 1 vote
Article Rating
Subscribe
Notify of
guest

This site uses Akismet to reduce spam. Learn how your comment data is processed.

0 Comments
Inline Feedbacks
View all comments