Improving Plagiarism Detection in Coding Assignments by Dynamic Removal of Common Ground

Christian Domin, Henning Pohl, Markus Krause

Research output: Contribution to book/anthology/report/conference proceedingArticle in proceedingResearchpeer-review

15 Citations (Scopus)

Abstract

Plagiarism in online learning environments has a detrimental effect on the trust of online courses and their viability. Automatic plagiarism detection systems do exist yet the specific situation in online courses restricts their use. To allow for easy automated grading, online assignments usually are less open and instead require students to fill in small gaps. Therefore solutions tend to be very similar, yet are then not necessarily plagiarized. In this paper we propose a new approach to detect code re-use that increases the prediction accuracy by dynamically removing parts in assignments which are part of almost every assignment—the so called common ground. Our approach shows significantly better F-measure and Cohen's Kappa results than other state of the art algorithms such as Moss or JPlag. The proposed method is also language agnostic to the point that training and test data sets can be taken from different programming languages.
Original languageUndefined/Unknown
Title of host publicationProceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems - CHI EA '16
Number of pages7
Publication date2016
Pages1173-1179
DOIs
Publication statusPublished - 2016
Externally publishedYes
Event2016 CHI Conference on Human Factors in Computing Systems - San Jose, United States
Duration: 7 May 201612 May 2016
http://chi2016.acm.org/wp/

Conference

Conference2016 CHI Conference on Human Factors in Computing Systems
Country/TerritoryUnited States
CitySan Jose
Period07/05/201612/05/2016
Internet address

Cite this