Plagiarism in the hiring process compromises the integrity of evaluations, making it difficult to assess a candidate’s abilities accurately. Plagiarism detection tools help maintain fairness, creating a level playing field that leads to better hiring decisions.

Types of Code Plagiarism

We encounter three types of code plagiarism: 

Types of Plagiarism Detection Models 

Standard plagiarism detection

Standard plagiarism detection is enabled by default for all tests. It uses Moss (Measure of Software Similarity) to compare a candidate’s code against others in our database. This method effectively identifies plagiarism from external sources or copied test cases (as mentioned above under types of code plagiarism). However, it may result in false positives, depending on factors such as code length and similarity threshold.

Advanced AI plagiarism detection

To address all three types of code plagiarism, including AI-aided malpractice, we have developed an advanced Machine Learning (ML) model. This model analyzes multiple factors, such as:

The ML model predicts suspicious behavior and classifies candidate attempts in a test as either "High" or "Medium." This detection is especially useful for identifying the use of AI tools like ChatGPT when their usage is restricted.

The flagged attempts are highlighted in candidate reports, allowing hiring teams to review the reasons and make informed decisions. Human oversight remains crucial, as the ML model currently has an overall precision of 85%, meaning flagged sessions are correct 85% of the time.

Note: The AI Plagiarism Detection feature currently supports only Coding questions.

Limitations

AI plagiarism detection is limited when evaluating questions requiring minimal effort to achieve a full score or when solutions consist of very few lines of code. These scenarios often lack enough data points for accurate detection. However, the system will still flag the response for potential plagiarism if significant external content is copy-pasted.

Note: Given that it’s a Machine Learning model, we comply with the audit requirements mentioned in the AEDT / NYC AI law.  

Enabling Advanced Plagiarism Detection for a Test

This feature can be enabled from Test Settings. 

Steps

7c6722ef-474f-4575-86f8-7df538645a7a.gif

Candidate Consent

When AI plagiarism detection is enabled, candidates must consent before starting the test.

7bf3e75d-b7ec-434b-9965-65858e62a65e (1) (1).png

Signals Used for Detection

A non-exhaustive list of factors considered:

Viewing Flagged Plagiarism Attempts

When a candidate's test attempt is flagged for plagiarism, it is categorized as "High" or "Medium" based on the confidence level of the model’s prediction. Follow the steps below to view these flagged attempts.

Steps

  1.  Navigate to the Candidate List and choose the candidate whose test attempt has been flagged.

  2. You can view the suspicious activity under the Performance tab in the Candidate Summary report.

  3. To review more details on the plagiarism signals, click on the question that has been flagged.

  4. In the question details, the Keystroke Codeplayer lets you quickly understand why a candidate was flagged for plagiarism. You can navigate to the relevant portions of their coding journey where potential suspicious activity may have occurred.

CodeSimilarity.gif

Understanding the Reasons for Flagging

There are two main categories under which a candidate’s attempt can be flagged:

Code Similarity: The candidate’s code is similar to one or more candidates' submissions, indicating potential plagiarism.

Coding Patterns: Unusual behavior was detected through the candidate’s keystroke patterns during the test. This may involve:

Sub-Patterns in Coding Behavior: The following sub-patterns are commonly detected when coding patterns raise concerns about the candidate’s behavior:

Continuous Model Improvement

HackerRank’s dedicated ML team continuously improves the plagiarism detection model by learning from emerging patterns of code plagiarism and misclassifications. When a statistically significant improvement is made, new updates to the model are deployed automatically.

FAQs

Will coding in an external IDE instead of the HackerRank platform trigger a potential cheating flag?

If candidates use Projects, which support an offline workflow (coding in their own IDE and pushing the code to the HackerRank platform), no cheating flag is triggered.

However, if candidates solve a standard coding question in an external IDE and copy and paste their code into the HackerRank platform, the system is likely to flag this behavior as suspicious activity.

Related Articles