Maintaining Academic Integrity in Programming: Locality-Sensitive Hashing and Recommendations

Not many efficient similarity detectors are employed in practice to maintain academic integrity. Perhaps it is because they lack intuitive reports for investigation, they only have a command line interface, and/or they are not publicly accessible. This paper presents SSTRANGE, an efficient similarit...

Full description

Saved in:
Bibliographic Details
Main Author: Oscar Karnalim (Author)
Format: Book
Published: MDPI AG, 2023-01-01T00:00:00Z.
Subjects:
Online Access:Connect to this object online.
Tags: Add Tag
No Tags, Be the first to tag this record!

MARC

LEADER 00000 am a22000003u 4500
001 doaj_59e7e246f0f240e69477523b64bcf4a9
042 |a dc 
100 1 0 |a Oscar Karnalim  |e author 
245 0 0 |a Maintaining Academic Integrity in Programming: Locality-Sensitive Hashing and Recommendations 
260 |b MDPI AG,   |c 2023-01-01T00:00:00Z. 
500 |a 10.3390/educsci13010054 
500 |a 2227-7102 
520 |a Not many efficient similarity detectors are employed in practice to maintain academic integrity. Perhaps it is because they lack intuitive reports for investigation, they only have a command line interface, and/or they are not publicly accessible. This paper presents SSTRANGE, an efficient similarity detector with locality-sensitive hashing (MinHash and Super-Bit). The tool features intuitive reports for investigation and a graphical user interface. Further, it is accessible on GitHub. SSTRANGE was evaluated on the SOCO dataset under two performance metrics: f-score and processing time. The evaluation shows that both MinHash and Super-Bit are more efficient than their predecessors (Cosine and Jaccard with 60% less processing time) and a common similarity measurement (running Karp-Rabin greedy string tiling with 99% less processing time). Further, the effectiveness trade-off is still reasonable (no more than 24%). Higher effectiveness can be obtained by tuning the number of clusters and stages. To encourage the use of automated similarity detectors, we provide ten recommendations for instructors interested in employing such detectors for the first time. These include consideration of assessment design, irregular patterns of similarity, multiple similarity measurements, and effectiveness-efficiency trade-off. The recommendations are based on our 2.5-year experience employing similarity detectors (SSTRANGE's predecessors) in 13 course offerings with various assessment designs. 
546 |a EN 
690 |a programming 
690 |a plagiarism 
690 |a collusion 
690 |a similarity detection 
690 |a recommendations 
690 |a higher education 
690 |a Education 
690 |a L 
655 7 |a article  |2 local 
786 0 |n Education Sciences, Vol 13, Iss 1, p 54 (2023) 
787 0 |n https://www.mdpi.com/2227-7102/13/1/54 
787 0 |n https://doaj.org/toc/2227-7102 
856 4 1 |u https://doaj.org/article/59e7e246f0f240e69477523b64bcf4a9  |z Connect to this object online.